Back to Search Start Over

Extracting Keywords from Images Using Deep Learning for the Visually Challenged

Authors :
Jaboob, Said
Chauhan, Munes Singh
Dhanasekaran, Balaji
Natarajan, Senthil Kumar
Source :
International Society for Technology, Education, and Science. 2022.
Publication Year :
2022

Abstract

Assistive technologies can in many ways facilitate the normal day-to-day lives of the disabled. As part of the ongoing research on assistive technologies at UTAS, Oman, that deals with augmenting and finding multimodal aspects of applications for the disabled, this paper aspires to investigate the role of deep learning in the field of image interpretation. Images are one of the most important mediums of conveying information among humans. Visually impaired persons especially with low cognitive abilities face insurmountable difficulties in understanding cues through images. This challenge is met by filtering words from image captions to facilitate understanding of the key notion conveyed by an image. This work utilizes the image captioning technique using deep learning frameworks such as convolution neural networks (CNN) and recurrent neural networks (RNN) to generate captions. These captions are fed to Rake, an NLP library that identifies keywords in the caption. The entire process is automated and uses transfer learning techniques for caption generation from images. This process is then further integrated with our main project, Finger Movement Multimodal Assistive System (FMAS) thereby incorporating text cues for interpreting images for the visually impaired. [For the full proceedings, see ED630948.]

Details

Language :
English
Database :
ERIC
Journal :
International Society for Technology, Education, and Science
Publication Type :
Conference
Accession number :
ED630963
Document Type :
Speeches/Meeting Papers<br />Reports - Descriptive