Journal article 248 views
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5
Electronics, Volume: 14, Issue: 20, Start page: 3976
Swansea University Authors:
Thomas Hobbs, Anwar Ali
-
PDF | Version of Record
© 2025 by the authors. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Download (7.81MB)
DOI (Published version): 10.3390/electronics14203976
Abstract
This paper outlines the process of developing a low-cost system for home appliance control via real-time hand gesture classification using Computer Vision and a custom lightweight machine learning model. This system strives to enable those with speech or hearing disabilities to interface with smart...
| Published in: | Electronics |
|---|---|
| ISSN: | 2079-9292 |
| Published: |
MDPI AG
2025
|
| Online Access: |
Check full text
|
| URI: | https://cronfa.swan.ac.uk/Record/cronfa70718 |
| first_indexed |
2025-10-18T08:36:56Z |
|---|---|
| last_indexed |
2025-12-05T18:10:22Z |
| id |
cronfa70718 |
| recordtype |
SURis |
| fullrecord |
<?xml version="1.0"?><rfc1807><datestamp>2025-12-04T10:23:52.1532198</datestamp><bib-version>v2</bib-version><id>70718</id><entry>2025-10-18</entry><title>Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5</title><swanseaauthors><author><sid>2433005846a71867a5aa3363b95314f4</sid><firstname>Thomas</firstname><surname>Hobbs</surname><name>Thomas Hobbs</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>f206105e1de57bebba0fd04fe9870779</sid><ORCID>0000-0001-7366-9002</ORCID><firstname>Anwar</firstname><surname>Ali</surname><name>Anwar Ali</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2025-10-18</date><deptcode>ACEM</deptcode><abstract>This paper outlines the process of developing a low-cost system for home appliance control via real-time hand gesture classification using Computer Vision and a custom lightweight machine learning model. This system strives to enable those with speech or hearing disabilities to interface with smart home devices in real time using hand gestures, such as is possible with voice-activated ‘smart assistants’ currently available. The system runs on a Raspberry Pi 5 to enable future IoT integration and reduce costs. The system also uses the official camera module v2 and 7-inch touchscreen. Frame preprocessing uses MediaPipe to assign hand coordinates, and NumPy tools to normalise them. A machine learning model then predicts the gesture. The model, a feed-forward network consisting of five fully connected layers, was built using Keras 3 and compiled with TensorFlow Lite. Training data utilised the HaGRIDv2 dataset, modified to consist of 15 one-handed gestures from its original of 23 one- and two-handed gestures. When used to train the model, validation metrics of 0.90 accuracy and 0.31 loss were returned. The system can control both analogue and digital hardware via GPIO pins and, when recognising a gesture, averages 20.4 frames per second with no observable delay.</abstract><type>Journal Article</type><journal>Electronics</journal><volume>14</volume><journalNumber>20</journalNumber><paginationStart>3976</paginationStart><paginationEnd/><publisher>MDPI AG</publisher><placeOfPublication/><isbnPrint/><isbnElectronic/><issnPrint/><issnElectronic>2079-9292</issnElectronic><keywords>machine learning; Computer Vision; gesture recognition; accessibility; smart home control; landmark normalisation; TensorFlow Lite; OpenCV</keywords><publishedDay>10</publishedDay><publishedMonth>10</publishedMonth><publishedYear>2025</publishedYear><publishedDate>2025-10-10</publishedDate><doi>10.3390/electronics14203976</doi><url/><notes/><college>COLLEGE NANME</college><department>Aerospace, Civil, Electrical, and Mechanical Engineering</department><CollegeCode>COLLEGE CODE</CollegeCode><DepartmentCode>ACEM</DepartmentCode><institution>Swansea University</institution><apcterm>Other</apcterm><funders>This research received no external funding.</funders><projectreference/><lastEdited>2025-12-04T10:23:52.1532198</lastEdited><Created>2025-10-18T09:34:22.3094189</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Aerospace, Civil, Electrical, General and Mechanical Engineering - Electronic and Electrical Engineering</level></path><authors><author><firstname>Thomas</firstname><surname>Hobbs</surname><order>1</order></author><author><firstname>Anwar</firstname><surname>Ali</surname><orcid>0000-0001-7366-9002</orcid><order>2</order></author></authors><documents><document><filename>70718__35384__32db0305c6e14461935eea6b6e80606d.pdf</filename><originalFilename>70718.pdf</originalFilename><uploaded>2025-10-18T09:36:49.5699900</uploaded><type>Output</type><contentLength>8186681</contentLength><contentType>application/pdf</contentType><version>Version of Record</version><cronfaStatus>true</cronfaStatus><documentNotes>© 2025 by the authors. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license.</documentNotes><copyrightCorrect>true</copyrightCorrect><language>eng</language><licence>https://creativecommons.org/licenses/by/4.0/</licence></document></documents><OutputDurs/></rfc1807> |
| spelling |
2025-12-04T10:23:52.1532198 v2 70718 2025-10-18 Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 2433005846a71867a5aa3363b95314f4 Thomas Hobbs Thomas Hobbs true false f206105e1de57bebba0fd04fe9870779 0000-0001-7366-9002 Anwar Ali Anwar Ali true false 2025-10-18 ACEM This paper outlines the process of developing a low-cost system for home appliance control via real-time hand gesture classification using Computer Vision and a custom lightweight machine learning model. This system strives to enable those with speech or hearing disabilities to interface with smart home devices in real time using hand gestures, such as is possible with voice-activated ‘smart assistants’ currently available. The system runs on a Raspberry Pi 5 to enable future IoT integration and reduce costs. The system also uses the official camera module v2 and 7-inch touchscreen. Frame preprocessing uses MediaPipe to assign hand coordinates, and NumPy tools to normalise them. A machine learning model then predicts the gesture. The model, a feed-forward network consisting of five fully connected layers, was built using Keras 3 and compiled with TensorFlow Lite. Training data utilised the HaGRIDv2 dataset, modified to consist of 15 one-handed gestures from its original of 23 one- and two-handed gestures. When used to train the model, validation metrics of 0.90 accuracy and 0.31 loss were returned. The system can control both analogue and digital hardware via GPIO pins and, when recognising a gesture, averages 20.4 frames per second with no observable delay. Journal Article Electronics 14 20 3976 MDPI AG 2079-9292 machine learning; Computer Vision; gesture recognition; accessibility; smart home control; landmark normalisation; TensorFlow Lite; OpenCV 10 10 2025 2025-10-10 10.3390/electronics14203976 COLLEGE NANME Aerospace, Civil, Electrical, and Mechanical Engineering COLLEGE CODE ACEM Swansea University Other This research received no external funding. 2025-12-04T10:23:52.1532198 2025-10-18T09:34:22.3094189 Faculty of Science and Engineering School of Aerospace, Civil, Electrical, General and Mechanical Engineering - Electronic and Electrical Engineering Thomas Hobbs 1 Anwar Ali 0000-0001-7366-9002 2 70718__35384__32db0305c6e14461935eea6b6e80606d.pdf 70718.pdf 2025-10-18T09:36:49.5699900 Output 8186681 application/pdf Version of Record true © 2025 by the authors. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. true eng https://creativecommons.org/licenses/by/4.0/ |
| title |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| spellingShingle |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 Thomas Hobbs Anwar Ali |
| title_short |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| title_full |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| title_fullStr |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| title_full_unstemmed |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| title_sort |
Smart Home Control Using Real-Time Hand Gesture Recognition and Artificial Intelligence on Raspberry Pi 5 |
| author_id_str_mv |
2433005846a71867a5aa3363b95314f4 f206105e1de57bebba0fd04fe9870779 |
| author_id_fullname_str_mv |
2433005846a71867a5aa3363b95314f4_***_Thomas Hobbs f206105e1de57bebba0fd04fe9870779_***_Anwar Ali |
| author |
Thomas Hobbs Anwar Ali |
| author2 |
Thomas Hobbs Anwar Ali |
| format |
Journal article |
| container_title |
Electronics |
| container_volume |
14 |
| container_issue |
20 |
| container_start_page |
3976 |
| publishDate |
2025 |
| institution |
Swansea University |
| issn |
2079-9292 |
| doi_str_mv |
10.3390/electronics14203976 |
| publisher |
MDPI AG |
| college_str |
Faculty of Science and Engineering |
| hierarchytype |
|
| hierarchy_top_id |
facultyofscienceandengineering |
| hierarchy_top_title |
Faculty of Science and Engineering |
| hierarchy_parent_id |
facultyofscienceandengineering |
| hierarchy_parent_title |
Faculty of Science and Engineering |
| department_str |
School of Aerospace, Civil, Electrical, General and Mechanical Engineering - Electronic and Electrical Engineering{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Aerospace, Civil, Electrical, General and Mechanical Engineering - Electronic and Electrical Engineering |
| document_store_str |
1 |
| active_str |
0 |
| description |
This paper outlines the process of developing a low-cost system for home appliance control via real-time hand gesture classification using Computer Vision and a custom lightweight machine learning model. This system strives to enable those with speech or hearing disabilities to interface with smart home devices in real time using hand gestures, such as is possible with voice-activated ‘smart assistants’ currently available. The system runs on a Raspberry Pi 5 to enable future IoT integration and reduce costs. The system also uses the official camera module v2 and 7-inch touchscreen. Frame preprocessing uses MediaPipe to assign hand coordinates, and NumPy tools to normalise them. A machine learning model then predicts the gesture. The model, a feed-forward network consisting of five fully connected layers, was built using Keras 3 and compiled with TensorFlow Lite. Training data utilised the HaGRIDv2 dataset, modified to consist of 15 one-handed gestures from its original of 23 one- and two-handed gestures. When used to train the model, validation metrics of 0.90 accuracy and 0.31 loss were returned. The system can control both analogue and digital hardware via GPIO pins and, when recognising a gesture, averages 20.4 frames per second with no observable delay. |
| published_date |
2025-10-10T18:10:22Z |
| _version_ |
1850692833205288960 |
| score |
11.08899 |

