
The eye is an essential sensory organ that allows us to perceive our surroundings at a glance. Losing this sense can result in numerous challenges in daily life. However, society is designed for the majority, which can create even more difficulties for visually impaired individuals. Therefore, empowering them and promoting self-reliance are crucial. To address this need, we propose a new Android application called “The Eye” that utilizes Machine Learning (ML)-based object detection techniques to recognize objects in real-time using a smartphone camera or a camera attached to a stick. The article proposed an improved YOLOv5l algorithm to improve object detection in visual applications. YOLOv5l has a larger model size and captures more complex features and details, leading to enhanced object detection accuracy compared to smaller variants like YOLOv5s and YOLOv5m. The primary enhancement in the improved YOLOv5l algorithm is integrating L1 and L2 regularization techniques. These techniques prevent overfitting and improve generalization by adding a regularization term to the loss function during training. Our approach combines image processing and text-to-speech conversion modules to produce reliable results. The Android text-to-speech module is then used to convert the object recognition results into an audio output. According to the experimental results, the improved YOLOv5l has higher detection accuracy than the original YOLOv5 and can detect small, multiple, and overlapped targets with higher accuracy. This study contributes to the advancement of technology to help visually impaired individuals become more self-sufficient and confident. Doi: 10.28991/ESJ-2023-07-05-011 Full Text: PDF
Artificial neural network, Radiology, Nuclear Medicine and Imaging, Artificial intelligence, Eye Movements, Object detection, Overfitting, Mobile device, Pattern recognition (psychology), Visual Object Tracking and Person Re-identification, Detection and Management of Retinal Diseases, Android (operating system), Health Sciences, Machine learning, T1-995, Regularization (linguistics), visually challenged, Technology (General), H1-99, text to speech, yolo, 006, Visual Attention, Computer science, Salient Object Detection, Social sciences (General), Object Tracking, Operating system, QA71-90 Instruments and machines, Computer Science, Physical Sciences, Multiple Object Tracking, Computational Modeling of Visual Saliency Detection, Medicine, Computer vision, Computer Vision and Pattern Recognition, machine learning.
Artificial neural network, Radiology, Nuclear Medicine and Imaging, Artificial intelligence, Eye Movements, Object detection, Overfitting, Mobile device, Pattern recognition (psychology), Visual Object Tracking and Person Re-identification, Detection and Management of Retinal Diseases, Android (operating system), Health Sciences, Machine learning, T1-995, Regularization (linguistics), visually challenged, Technology (General), H1-99, text to speech, yolo, 006, Visual Attention, Computer science, Salient Object Detection, Social sciences (General), Object Tracking, Operating system, QA71-90 Instruments and machines, Computer Science, Physical Sciences, Multiple Object Tracking, Computational Modeling of Visual Saliency Detection, Medicine, Computer vision, Computer Vision and Pattern Recognition, machine learning.
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 2 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
