Please use this identifier to cite or link to this item: http://hdl.handle.net/123456789/4342
Full metadata record
DC FieldValueLanguage
dc.contributor.authorAlnuaim, A.-
dc.contributor.authorZakariah, A.-
dc.contributor.authorHatamleh, W. A.-
dc.contributor.authorTarazi, H.-
dc.contributor.authorTripathi, V.-
dc.contributor.authorAmoatey, E. T.-
dc.date.accessioned2025-02-04T11:42:28Z-
dc.date.available2025-02-04T11:42:28Z-
dc.date.issued2022-
dc.identifier.issn1687-5265-
dc.identifier.urihttp://hdl.handle.net/123456789/4342-
dc.description.abstractSign language is the native language of deaf people, which they use in their daily life, and it facilitates the communication process between deaf people. )e problem faced by deaf people is targeted using sign language technique. Sign language refers to the use of the arms and hands to communicate, particularly among those who are deaf. )is varies depending on the person and the location from which they come. As a result, there is no standardization about the sign language to be used; for example, American, British, Chinese, and Arab sign languages are all distinct. Here, in this study we trained a model, which will be able to classify the Arabic sign language, which consists of 32 Arabic alphabet sign classes. In images, sign language is detected through the pose of the hand. In this study, we proposed a framework, which consists of two CNN models, and each of them is individually trained on the training set. )e final predictions of the two models were ensembled to achieve higher results. )e dataset used in this study is released in 2019 and is called as ArSL2018. It is launched at the Prince Mohammad Bin Fahd University, Al Khobar, Saudi Arabia. )e main contribution in this study is resizing the images to 64 ∗ 64 pixels, converting from grayscale images to three-channel images, and then applying the median filter to the images, which acts as lowpass filtering in order to smooth the images and reduce noise and to make the model more robust to avoid overfitting. )en, the preprocessed image is fed into two different models, which are ResNet50 and MobileNetV2. ResNet50 and MobileNetV2 architectures were implemented together. )e results we achieved on the test set for the whole data are with an accuracy of about 97% after applying many preprocessing techniques and different hyperparameters for each model, and also different data augmentation techniques.en_US
dc.language.isoenen_US
dc.publisherHindawien_US
dc.relation.ispartofseriesVol.2022;-
dc.titleHUMAN-COMPUTER INTERACTION WITH HAND GESTURE RECOGNITION USING RESNET AND MOBILENETen_US
dc.typeArticleen_US
Appears in Collections:School of Engineering

Files in This Item:
File Description SizeFormat 
HUMAN-COMPUTER INTERACTION WITH HAND GESTURE RECOGNITION USING RESNET AND MOBILENET.pdf3.09 MBAdobe PDFView/Open


Items in UDSspace are protected by copyright, with all rights reserved, unless otherwise indicated.