Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
19 views7 pages

ANN Paper

The document presents 'LipNet', an advanced automated lip-reading system that utilizes deep learning techniques to interpret spoken language through visual analysis of lip movements in video sequences. It aims to enhance speech recognition in environments where audio methods are ineffective, such as noisy settings or for individuals with hearing impairments. The system integrates Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Generative Adversarial Networks (GANs) to improve accuracy and robustness in sentence-level lip reading.

Uploaded by

Mohit Lunia
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
19 views7 pages

ANN Paper

The document presents 'LipNet', an advanced automated lip-reading system that utilizes deep learning techniques to interpret spoken language through visual analysis of lip movements in video sequences. It aims to enhance speech recognition in environments where audio methods are ineffective, such as noisy settings or for individuals with hearing impairments. The system integrates Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Generative Adversarial Networks (GANs) to improve accuracy and robustness in sentence-level lip reading.

Uploaded by

Mohit Lunia
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 7

LipNet : End to End Sentence Level Lipreading

Mohit Lunia Dr Anupama Kumar


Dept. of AI&ML Dept of AI&ML
R V College of Engineering Ashika V R V College of Engineering
Bengaluru, India Dept. of AI&ML Bengaluru, India
[email protected] R V College of Engineering
Bengaluru, India [email protected]
[email protected]
as multi-leveled equations, graphics, and tables are not
prescribed, although the various table text styles are
Abstract— This project introduces the development of a provided. The formatter will need to create these
sophisticated automated lip-reading system aimed at
components, incorporating the applicable criteria that
accurately interpreting spoken language through visual
analysis, specifically by examining lip movements captured in follow.
video sequences. The system is designed to function effectively
II. LITERATURE REVIEW
in environments where traditional audio-based speech
recognition methods are impractical or unavailable, such as
noisy industrial settings or for individuals with hearing
impairments. Leveraging cutting-edge deep learning Zhou, X., & Wu, H. explored the application of deep
techniques, the system strives to offer a dependable and learning to lip reading with their work on visual speech
scalable solution for visual speech recognition. The system recognition. Their paper addresses the complexities of
architecture encompasses several essential stages: video input translating visual lip movements into readable text,
preprocessing, feature extraction, Region of Interest (ROI) emphasizing the necessity of a robust deep learning
detection, neural network training, and speech decoding.
framework to manage speaker variability and contextual
Initially, video data is acquired and preprocessed to enhance
differences. [1] The study reviews techniques such as
quality, mitigating the impact of variations in lighting,
background noise, and other environmental factors on system convolutional neural networks (CNNs) for extracting spatial
performance. Convolutional Neural Networks (CNNs) are used features and recurrent neural networks (RNNs) for capturing
to extract spatial features from the video, focusing on the lip the temporal dynamics of lip movements. The authors
region and adjacent facial features crucial for accurate lip- underscore the critical role of spatial-temporal modeling in
reading. The system then isolates the ROI, concentrating on the achieving accurate sentence-level lip reading predictions,
mouth area to refine the analysis precision. To capture the emphasizing the need for precise temporal alignment
temporal dynamics of speech, the system integrates Recurrent between visual input and corresponding speech output to
Neural Networks (RNNs) with Long Short-Term Memory improve the accuracy of visual speech recognition.
(LSTM) units, which effectively track the sequence of lip
movements over time. Additionally, the inclusion of Shillingford, B., Assael, Y., Hoffmann, M., Paine, T.,
Transformers and Temporal Convolutional Networks (TCNs) Hughes, C., and others, in their pioneering research on
enhances the system's capability to process sequential data and LipNet, demonstrate an end-to-end deep learning approach
maintain contextual information throughout the speech. The for sentence-level lip reading. [2] Their model employs a
neural network is trained on extensive datasets to learn the
combination of CNNs for feature extraction and Long Short-
complex relationships between lip movements and
corresponding speech, ultimately producing a text output that
Term Memory (LSTM) networks to model temporal
accurately reflects the spoken language. sequences, resulting in a significant reduction in word error
rates compared to traditional methods. This work
Keywords—Lip reading, convolutional neural network, deep emphasizes the benefits of sequence-to-sequence modeling
learning, visual speech recognition, recurrent neural network. in handling continuous visual input for accurate speech
prediction from lip movements, addressing challenges like
I. INTRODUCTION inter-speaker variability and different lip motion styles.
The project titled "LipNet: End-to-End Sentence-Level Lip
Chung, J. S., Senior, A., Vinyals, O., & Zisserman, A.
Reading" represents a significant advancement in the field
developed a novel lip reading approach that combines deep
of visual speech recognition. LipNet aims to transform how
learning with attention mechanisms. Their model processes
spoken language is interpreted by developing an advanced both visual input from lip movements and the corresponding
deep learning model that reads sentences based solely on the speech audio, improving recognition accuracy even in noisy
visual analysis of lip movements. This innovative approach environments. [3] By leveraging attention, the model
is designed to address the inherent challenges of traditional focuses on key video frames where lip movements are most
lip reading, which has historically required a high level of pronounced, aiding in the interpretation of more complex
skill and experience, and is often fraught with subjectivity sentences. This method overcomes one of the major
and susceptibility to errors, particularly in noisy or complex challenges in lip reading, which is the difficulty in
environments. Lip reading, or visual speech recognition, discerning similar-looking lip movements for different
involves interpreting speech by analyzing the movements of words or phonemes, thereby significantly enhancing
the lips rather than relying on auditory information. This performance in varied environmental conditions.
process has traditionally been labor-intensive and imprecise,
especially in settings where audio-based speech recognition Assael, Y. M., Shillingford, B., Whiteson, S., & De Freitas,
systems falter, such as noisy industrial environments or for N. created LipNet, a system that pushes the boundaries of
automated lip reading. The model uses a combination of
CNNs and LSTMs to process video sequences of spoken
sentences, generating accurate textual predictions. [4]
individuals with hearing impairments. The manual nature of LipNet's performance surpasses that of human lip readers
traditional lip reading often leads to variability in accuracy and other existing models on benchmark datasets like GRID,
due to differences in individual skill levels and highlighting its potential in fields such as security, assistive
environmental factors. LipNet seeks to overcome these technology, and human-computer interaction. Their research
also explores the use of Connectionist Temporal
limitations by leveraging cutting-edge deep learning
Classification (CTC) in training LipNet, ensuring it can
techniques to automate and enhance the lip reading process,
predict speech without the need for manually annotated
making it more reliable, scalable, and accessible. alignment between the input video frames and target speech,
parentheses, following the example. Some components, such
further demonstrating its innovation in the automated lip III. EXISTING AND PROPOSED SYSTEM
reading domain.

Stafylakis, T., & Tzimiropoulos, G. conducted research on Existing lip reading systems primarily depend on
end-to-end models for visual speech recognition, conventional machine learning techniques that involve
introducing a deep residual network combined with a significant manual intervention for feature extraction. These
sequence-to-sequence learning approach. [5] Their work methods include techniques such as Optical Flow, which
stands out for incorporating a lip reading-specific variant of tracks motion across frames, Hidden Markov Models
residual networks, which improves the ability of models to (HMMs) for temporal modeling, and Support Vector
capture subtle spatial cues from lip movements. Their
Machines (SVMs) for classification tasks. While these
experiments indicate that deeper architectures, such as
traditional methods have been instrumental in advancing the
ResNet, can outperform shallower ones when combined
with temporal modeling techniques like bidirectional field, they face significant challenges. They are highly
LSTMs. Additionally, their focus on sentence-level susceptible to environmental factors such as varying lighting
prediction rather than isolated words marks a significant conditions, speaker pose, and inconsistent camera angles.
shift in the development of lip reading systems that can Moreover, these systems struggle with speaker variability
handle more complex real-world applications. and often falter when faced with unseen subjects or complex
sentences, reducing their overall accuracy and robustness.
Martinez, J., Ma, P., & Petridis, S. examined multimodal
approaches to lip reading by integrating both audio and
visual inputs into their models. They argued that relying The proposed system introduces a more advanced deep
solely on visual information can limit performance, learning approach that automates the entire process. Using
particularly in complex environments with challenging Convolutional Neural Networks (CNNs), the system can
lighting conditions or occluded facial features. [6] Their extract detailed spatial features from video frames, focusing
work introduced a hybrid model that fuses audio features on lip shapes and movements. In combination with
(when available) with visual features, using CNNs and Recurrent Neural Networks (RNNs), particularly Long
temporal networks to synchronize the two modalities. The Short-Term Memory (LSTM) networks, the model captures
results showed a marked improvement in word recognition the temporal flow of speech by learning the sequential
rates, particularly for words with similar visual lip dependencies between frames. Unlike traditional methods,
movements, as the audio input helps disambiguate the which tend to focus on individual words or phonemes, the
correct prediction. This multimodal approach presents a path proposed system leverages sentence-level predictions,
forward for more robust lip reading systems that can
enhancing its capability to understand the context of speech.
perform reliably even in less-than-ideal settings.

Petridis, S., Stafylakis, T., & Pantic, M. proposed a novel


approach to lip reading using generative adversarial Additionally, the proposed system employs data
networks (GANs). Their method focuses on generating augmentation techniques, such as rotation, scaling, and noise
synthetic lip movements to augment the training data addition, to improve its robustness against variability in real-
available for deep learning models. [7] This approach world data. Generative Adversarial Networks (GANs) are
addresses one of the key challenges in lip reading research: also integrated into the system to generate synthetic data that
the scarcity of high-quality labeled data. By using GANs to mimics real-world scenarios, allowing the model to
create realistic lip movement sequences, the authors generalize better to unseen conditions. This combination of
improved the model's ability to generalize to unseen CNNs, LSTMs, and GANs not only addresses the
speakers and different speaking styles. Their experiments limitations of traditional systems but also significantly
demonstrate that models trained with synthetic data can improves performance in terms of word error rate (WER)
achieve comparable performance to those trained and character error rate (CER), particularly when evaluated
exclusively on real data, opening up new possibilities for on challenging datasets like GRID and LRW.
data augmentation in the field of lip reading.

.
IV. TECHNICAL ARCHITECTURE
The technical architecture of the proposed lip reading
system represents a significant advancement in the field,
leveraging state-of-the-art deep learning techniques to
achieve precise and reliable sentence-level recognition. This
architecture is designed to address the inherent challenges of
lip reading, such as variability in speaker characteristics,
environmental conditions, and the complexity of visual
speech patterns.

At the foundation of the architecture are Convolutional


Neural Networks (CNNs), which play a crucial role in
processing and analyzing video frames. CNNs are adept at
extracting spatial features from images, making them well-
suited for capturing the detailed nuances of lip movements.
The network consists of multiple convolutional layers that The overall architecture of the proposed system is designed
apply various filters to detect and learn patterns such as to streamline the lip reading process from raw video input to
edges, textures, and shapes. These filters progressively build textual output. By combining CNNs, LSTMs, and GANs in
up from low-level features to high-level representations, an integrated framework, the system addresses the
allowing the model to identify complex lip shapes and limitations of traditional lip reading methods and achieves
expressions crucial for accurate speech interpretation. The high accuracy in sentence-level recognition. The use of end-
use of pooling layers in conjunction with convolutions helps to-end training with CTC further simplifies the model
in reducing the dimensionality of the data while preserving development and improves its performance. This
essential spatial information. This process enables the CNN comprehensive approach ensures that the system is capable
to focus on the most relevant features of the lip movements, of handling the complexities of visual speech recognition
minimizing the impact of background noise and other and provides a robust solution for real-world applications,
distractions. such as assistive technology for the hearing impaired,
security and surveillance, and human-computer interaction.
Following the feature extraction phase, the system
incorporates Recurrent Neural Networks (RNNs),
specifically Long Short-Term Memory (LSTM) networks, to
model the temporal dynamics of speech. Unlike CNNs,
which are designed for spatial feature extraction, LSTMs are
optimized for sequential data. They excel in capturing
dependencies and relationships over time, making them
ideal for understanding the flow of speech across multiple
frames. LSTMs address the challenges of vanishing and
exploding gradients, which are common in traditional
RNNs, by maintaining a memory cell that stores information Fig 1. General Flow of Lip Reading
over long sequences. This capability is crucial for lip
reading, where the sequence of lip movements needs to be
analyzed in the context of both immediate and distant V. TOOLS AND TECHNOLOGIES USED
frames.
Python (v.3.12.2): A widely-used programming language
The integration of LSTMs with CNNs creates a powerful valued for its readability and ease of use. The latest version
combination that allows the system to handle the complexity includes features that are advantageous for developing
of visual speech. The CNNs provide detailed spatial complex applications and performing data analysis, making
representations, while the LSTMs manage the temporal it particularly suitable for machine learning and deep
sequencing of these representations. This synergy enables learning projects.
the system to predict entire sentences rather than isolated
words or phonemes, offering a more comprehensive NumPy (v.1.24.3) : A core library for numerical computing
understanding of the spoken language. in Python, providing support for large arrays and matrices,
along with a variety of mathematical functions. It is essential
To further enhance the performance of the system, the for data manipulation and preparation, which are crucial
architecture employs the Connectionist Temporal steps in machine learning workflows.
Classification (CTC) loss function. CTC is particularly
effective for sequence-to-sequence tasks where the Matplotlib (v.3.7.2): A library for creating static, animated,
alignment between input and output sequences is not and interactive visualizations in Python. It is vital for
predetermined. This loss function allows the model to be plotting and visualizing data, such as training loss curves
trained end-to-end, directly mapping raw video frames to and performance metrics, which aids in monitoring and
textual outputs without requiring manually segmented data. improving model performance.
CTC helps the system to handle varying lengths of input
sequences and generate accurate predictions by learning to OpenCV (v.4.6.0) : An open-source computer vision library
align the sequences dynamically during training. that is extensively used for image and video processing
tasks. It facilitates the extraction, transformation, and
In addition to CNNs, LSTMs, and CTC, the architecture augmentation of visual data, making it a key tool for
incorporates Generative Adversarial Networks (GANs) for preparing inputs for deep learning models.
data augmentation. GANs are used to generate synthetic
training data that mimics real-world conditions, such as TensorFlow (v.2.12.0): A comprehensive library for
variations in lighting, speaker angles, and background noise. building and training deep learning models, including
This synthetic data helps in improving the robustness and Convolutional Neural Networks (CNNs). It provides tools
generalization of the model by exposing it to a wider range for designing complex neural network architectures and
of scenarios. GANs work through a process of adversarial performing efficient computation on both CPUs and GPUs.
training, where a generator network creates synthetic
samples, and a discriminator network evaluates their PyTorch (v.2.0.0): A deep learning framework that offers
authenticity. This approach enhances the model's ability to dynamic computation graphs and robust support for GPU
adapt to diverse conditions and reduces overfitting to acceleration. It is used for developing and training various
specific training data. neural network models, particularly in research and
development contexts.
of the input sequence, improving its ability to capture
Keras-GAN: A specialized library for implementing important details while ignoring less significant information.
Generative Adversarial Networks (GANs). It is utilized for This capability is particularly useful for lip reading, where
generating synthetic data and enhancing the diversity and specific lip movements can be crucial for interpreting speech
robustness of the training dataset. accurately.

Numba (v.0.56.0): A Just-In-Time (JIT) compiler for Generative Adversarial Networks (GANs) are sometimes
Python that translates Python functions to optimized employed to further enhance the model's capabilities. GANs
machine code at runtime. It speeds up numerical can be used to generate synthetic video frames or improve
computations and data processing tasks, crucial for handling the quality of existing frames, addressing challenges such as
large datasets and complex model computations. noise or low resolution in the training data.

Git (v.2.39.0): A version control system that helps manage The training process is a vital phase in model building,
code changes and track different project versions. It is involving several steps to optimize the model’s
essential for collaborative development, code management, performance. Selecting an appropriate loss function, such as
and maintaining project history. Connectionist Temporal Classification (CTC) loss, is
crucial. CTC loss is well-suited for lip reading because it
Jupyter Notebooks (v.6.5.1): An interactive computing allows the model to align predicted sequences with ground
environment that allows for the integration of code, truth sequences without requiring explicit alignment.
visualizations, and narrative text. It is used for Optimization techniques like Stochastic Gradient Descent
experimentation, documentation, and sharing results within (SGD) or Adam are used to adjust the model’s parameters,
the research and development team. minimizing the error between predicted and actual
sequences.

VI. MODEL BUILDING To assess the model’s effectiveness, evaluation metrics such
as Word Error Rate (WER) and Character Error Rate (CER)
The process of building a model for lip reading involves are employed. These metrics quantify the accuracy of the
several critical steps aimed at developing an effective model’s predictions and provide a basis for comparing its
system capable of accurately interpreting spoken sentences performance against other models or baselines. Cross-
based solely on visual inputs. The journey begins with data validation is performed by splitting the data into training and
collection and preparation, which is essential for training a validation sets to ensure that the model generalizes well to
robust model. This involves gathering a diverse dataset of new, unseen data. Error analysis is conducted to identify
video sequences featuring lip movements and their common patterns or specific areas where the model
corresponding text annotations. Preprocessing these videos struggles, guiding further refinements and adjustments.
is crucial; this includes extracting frames from the video
sequences and normalizing their size and color scales to
ensure consistency. To enhance the model’s robustness, data
augmentation techniques such as cropping, rotating, and
scaling are applied, increasing the diversity of the training
set and improving the model's ability to generalize across
different scenarios.
Hyperparameter tuning is another critical aspect of the
Once the data is prepared, the focus shifts to the design of training process. This involves adjusting parameters such as
the model architecture. For lip reading, Convolutional learning rate, batch size, and network depth to optimize
Neural Networks (CNNs) play a pivotal role in extracting performance. Techniques like grid search or random search
spatial features from each frame. CNNs are adept at are used to find the best combination of hyperparameters.
recognizing patterns in images, such as lip shapes and facial
expressions. By stacking multiple convolutional layers, the Once the model achieves satisfactory performance, it moves
model can build a hierarchical representation of the visual to the deployment stage. Integration involves embedding the
data, capturing both low-level textures and high-level model into a user-facing application or system where it can
semantic features. process live video input and deliver real-time lip reading
results. Continuous monitoring of the model’s performance
To handle the temporal aspect of lip reading, Recurrent in the deployment environment ensures that it remains
Neural Networks (RNNs), particularly Long Short-Term accurate and reliable over time.
Memory (LSTM) networks, are utilized. LSTMs are
designed to process sequences and retain information over Overall, building a lip reading model is a multifaceted
extended periods, which is essential for modeling the endeavor that combines data preparation, advanced neural
continuous flow of speech. They help the model capture network architectures, and meticulous training and
dependencies between lip movements across time, which is evaluation to create a system capable of interpreting spoken
critical for accurate sentence-level predictions. language from visual cues with high accuracy.

In addition to CNNs and LSTMs, attention mechanisms are


VII. CONCLUSION
integrated into the model to enhance its performance.
Attention layers allow the model to focus on relevant parts
The development of advanced lip reading systems, such as [5] Kwanyoung Kim, Jong Chul Ye. Noise2Score:
LipNet, represents a significant leap forward in the field of Tweedie’s Approach to Self-Supervised Image Denoising
visual speech recognition. By leveraging state-of-the-art
without Clean Images. Part of Advances in Neural
deep learning techniques, including Convolutional Neural
Networks (CNNs) and Long Short-Term Memory (LSTM) Information Processing Systems 34 (NeurIPS 2021).
networks, these systems can interpret spoken sentences
solely based on visual inputs. This capability has profound [6] Linh Duy Tran, Son Minh Nguyen, Masayuki Arai.
implications for various applications, from assistive GAN-based Noise Model for Denoising Real Images.
technologies for the hearing impaired to security and Proceedings of the Asian Conference on Computer Vision
surveillance in privacy-sensitive environments.
(ACCV), 2020.
The comprehensive approach of LipNet, integrating CNNs
for spatial feature extraction with LSTMs for temporal [7] M. Momen-Tayefeh, M. Momen-Tayefeh, F. Z.
sequence modeling, has demonstrated substantial
Hasheminasab, S. A. G. Ghahramani. SNRGAN: The Semi
improvements over traditional lip reading methods. The use
of attention mechanisms further enhances the model's ability Noise Reduction GAN for Image Denoising. 20th CSI
to focus on relevant visual cues, improving overall accuracy. International Symposium on Artificial Intelligence and
Additionally, innovations such as end-to-end learning and Signal Processing (AISP), Babol, Iran, Islamic Republic of,
sequence-to-sequence modeling streamline the process, 2024, pp. 1-5.
making it more efficient and effective.
The performance evaluation of LipNet highlights its [8] Cha, Sungmin, Taeeon Park, Taesup Moon. Gan2Gan:
superiority over baseline models, achieving lower Word Generative noise learning for blind image denoising with
Error Rates (WER) and Character Error Rates (CER) in
single noisy images. arXiv preprint arXiv:1905.10488
various testing scenarios. This demonstrates its potential to
handle complex lip movements and varying speech patterns (2019).
more effectively than previous approaches. The successful
application of LipNet underscores the potential for deep [9] Dong-Wook Kim, Jae Ryun Chung, Seung-Won Jung.
learning models to address challenges in visual speech GRDN: Grouped Residual Dense Network for Real Image
recognition, paving the way for more reliable and scalable Denoising and GAN-Based Real-World Noise Modeling.
solutions.
Proceedings of the IEEE/CVF Conference on Computer
Vision and Pattern Recognition (CVPR) Workshops, 2019,
As technology continues to evolve, the integration of pp. 0-0.
advanced models like LipNet into practical applications will
likely lead to further advancements in assistive technology, [10] Ziyuan Wang, Lidan Wang, Shukai Duan, Yunfei Li.
human-computer interaction, and security systems. The
An Image Denoising Method Based on Deep Residual
ongoing research and development in this field promise to
enhance communication capabilities and provide new tools GAN. J. Phys.: Conf. Ser. 1550 032127 (2020).
for understanding and interpreting human speech in diverse
and challenging environments. [11] Ziyuan Wang, Lidan Wang, Shukai Duan, Yunfei Li.
An Image Denoising Method Based on Deep Residual
GAN. J. Phys.: Conf. Ser. 1550 032127 (2020).

VIII. REFERENCES [12] Yue Zhong, Lizhuang Liu, Dan Zhao, Hongyang Li. A
generative adversarial network for image denoising.
[1] Fan, L., Zhang, F., Fan, H. Brief Review of Image Multimed Tools Appl 79, 16517–16529 (2020).
Denoising Techniques. Vis. Comput. Ind. Biomed. Art 2, 7
(2019). [13] Zailiang Chen, Ziyang Zeng, Hailan Shen, Xianxian
Zheng, Peishan Dai, Pingbo Ouyang. DN-GAN: Denoising
[2] Chunwei Tian, Lunke Fei, Wenxian Zheng, Yong Xu, generative adversarial networks for speckle noise reduction
Wangmeng Zuo, Chia-Wen Lin. Deep learning on image in optical coherence tomography images. Biomedical Signal
denoising: An overview. Neural Networks, Volume 131, Processing and Control, Volume 55, 2020, 101632.
2020, Pages 251-275.
[14] S. Chen, D. Shi, M. Sadiq, X. Cheng. Image Denoising
[3] Ademola E. Ilesanmi, Taiwo O. Ilesanmi. Methods for With Generative Adversarial Networks and its Application
image denoising using convolutional neural network: a to Cell Image Enhancement. IEEE Access, vol. 8, pp.
review. Complex Intell. Syst. 7, 2179–2198 (2021). 82819-82831, 2020.

[4] Chunwei Tian, Yong Xu. Attention-guided CNN for [15] Dey, R., Bhattacharjee, D., Nasipuri, M., Mandal, J.,
image denoising. Neural Networks Volume 124, April 2020, Banerjee, S. Image Denoising Using Generative Adversarial
Pages 117-129. Network. In Intelligent Computing: Image Processing Based
Applications. Advances in Intelligent Systems and
Computing, vol 1157, 2020. Springer, Singapore.
[16] Ademola E. Ilesanmi, Taiwo O. Ilesanmi. Methods for [19] Tao Huang, Songjiang Li, Xu Jia, Huchuan Lu,
image denoising using convolutional neural networks: a Jianzhuang Liu. Neighbor2Neighbor: Self-Supervised
review. Complex Intell. Syst. 7, 2179–2198 (2021). Denoising From Single Noisy Images. Proceedings of the
IEEE/CVF Conference on Computer Vision and Pattern
[17] Chengjun Li, Yacen Wu, Haijun Lin. ECG Denoising Recognition (CVPR), 2021, pp. 14781-14790.
Method Based on an Improved VMD Algorithm. IEEE
Sensors Journal, vol. 22, no. 23, pp. 22725-22733, 1 Dec. [20] Pranaba K. Mishro, Sanjay Agrawal. A Survey on
2022. State-of-the-Art Denoising Techniques for Brain Magnetic
Resonance Images. IEEE Reviews in Biomedical
[18] Bhawna Goyal, Ayush Dogra. Image denoising review: Engineering ( Volume: 15), 2021.
From classical to state-of-the-art approaches. Information
Fusion Volume 55, March 2020, Pages 220-244.

You might also like