Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Article . 2023
License: CC BY
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Article . 2023
License: CC BY
Data sources: Datacite
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Article . 2023
License: CC BY
Data sources: ZENODO
versions View all 2 versions
addClaim

LIP READING AND PREDICTION SYSTEM BASED ON DEEP LEARNING

Authors: Gishma K.M; Annmaria K.B; Ramna Parvan V.N; Anagha Suresh; Athira Shaji;

LIP READING AND PREDICTION SYSTEM BASED ON DEEP LEARNING

Abstract

— Speech perception is characterized as a multimodal process, which means it elicits several meanings. Understanding a message can be aided by, and in some cases even made necessary by, lip reading, which overlays visual cues on top of auditory signals. Lip-reading is a crucial field with many uses, including biometrics, speech recognition in noisy environments, silent dictation, and enhanced hearing aids. It is a challenging research project in the area of computer vision, whose major goal is to watch the movement of human lips in a video and recognize the textual content that goes with it. Yet, due to the constraints of lip changes and the depth of linguistic information, the complexity of lip identification has increased, which has slowed the growth of study themes in lip language. Nowadays, deep learning has advanced in several sectors, giving us the confidence to perform the task of lip recognition. Lip learning based on deep learning often entails extracting features and comprehending images using a network model, as opposed to classical lip recognition that recognizes lip characteristics. The design of the network framework for data gathering, processing, and data recognition for lip reading is the main topic of this discussion. In this research, we created a reliable and accurate method for lip reading. We first isolate the mouth region and segment it, after which we extract various aspects from the lip image, such as the Hog, Surf, and Haar features. Lastly, we use Gated Recurrent Units to train our deep learning model (GRU).

Related Organizations
Keywords

Haar,Hog and Surf features,GRU based deep learning Architecture

  • BIP!
    Impact byBIP!
    selected citations
    These citations are derived from selected sources.
    This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    0
    popularity
    This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
    Average
    influence
    This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    Average
    impulse
    This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
    Average
    OpenAIRE UsageCounts
    Usage byUsageCounts
    visibility views 8
    download downloads 9
  • 8
    views
    9
    downloads
    Powered byOpenAIRE UsageCounts
Powered by OpenAIRE graph
Found an issue? Give us feedback
visibility
download
selected citations
These citations are derived from selected sources.
This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Citations provided by BIP!
popularity
This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
BIP!Popularity provided by BIP!
influence
This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Influence provided by BIP!
impulse
This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
BIP!Impulse provided by BIP!
views
OpenAIRE UsageCountsViews provided by UsageCounts
downloads
OpenAIRE UsageCountsDownloads provided by UsageCounts
0
Average
Average
Average
8
9
Green