IoT Journal Paper[1]
IoT Journal Paper[1]
ABSTRACT In this work, a vision-based gesture-controlled robotic arm is presented for remote and
intuitive control of physical systems. The system uses a camera and OpenCV-based gesture recognition
algorithms to identify predefined hand gestures in real time. These identified gestures are translated into
specific commands, which are wirelessly sent to a microcontroller. The microcontroller decodes these
signals to drive a robotic arm accordingly. This method obviates the requirement for physical contact, thus
making it particularly well-suited for applications involving hygiene, safety, or distance—like industrial
automation, healthcare, and dangerous environments. The paper discusses the overall system architecture,
implementation approach, gesture classification logic, and experimental results that illustrate the efficiency,
accuracy, and responsiveness of the proposed IoT-integrated control system.
INDEX TERM Gesture Recognition, Robotic Arm Control, Computer Vision, OpenCV, Human-Machine
Interaction (HMI), Hand Tracking, Contactless Control, Microcontroller-based Automation, Wireless
Communication, Vision-Based Control System, Real-Time Gesture Detection, Python-based Interface, IoT
Applications, Servo Motor Control, Low-Cost Automation.
1
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits
gestures, allowing for more natural and intuitive control. arm. The system operates by capturing hand gestures using a
For instance, Park et al. (2019) developed a system where a webcam or camera module, which is processed in real-time
depth camera was used to detect hand gestures and control a using OpenCV and MediaPipe to identify predefined
robotic arm. This system, however, was limited by the high gestures. These gestures are then transmitted to a
cost of the required hardware. microcontroller through Wi-Fi or Bluetooth, which controls
the 4-DOF robotic arm based on the detected gestures.
2.2 Computer Vision for Gesture Recognition
3.1 System Overview
The field of computer vision has witnessed significant The system architecture is divided into the following key
advancements with the introduction of deep learning and modules:
real-time hand tracking technologies. One notable 1. Gesture Recognition (PC/Computer): The first
framework is MediaPipe, developed by Google, which stage of the system is the detection of hand gestures
provides lightweight solutions for real-time hand landmark through a web camera. OpenCV processes the
detection. MediaPipe’s Hand module can track up to 21
video feed in real-time to extract key features such
hand landmarks, enabling accurate gesture recognition (Liu
as the position of the hand and its movements. The
et al., 2020). The use of OpenCV alongside MediaPipe has
MediaPipe Hand Module is used to track the 21
been shown to deliver highly efficient and real-time
landmarks of the hand, which are then mapped to
performance for gesture recognition tasks, making it an
specific gestures.
ideal choice for robotics and automation applications.
Several studies have successfully implemented hand gesture 2. Wireless Communication: Once the gesture is
recognition using OpenCV and MediaPipe. For example, identified, a command is sent wirelessly from the
Kumar and Jadhav (2020) presented a gesture-controlled computer to the microcontroller via Wi-Fi or
robotic arm system that used OpenCV’s Haar Cascade Bluetooth. This allows the control signal to be
Classifiers for detecting hand gestures. However, this transmitted without the need for physical wires,
method often struggled with complex gestures and real-time enabling a truly wireless control system.
performance. In contrast, MediaPipe’s real-time tracking 3. Microcontroller: The microcontroller is responsible
capabilities have proven to be more accurate and efficient, for receiving the control command and processing
especially for dynamic and continuous gestures (Liu et al., it. It interfaces with the servo motors that control
2020). the 4-DOF robotic arm. The microcontroller can
either communicate with the PC via Wi-Fi (MQTT
3.3 IoT Integration in Robotics or HTTP) or Bluetooth (Serial Communication).
The integration of the Internet of Things (IoT) with robotic 4. Robotic Arm: The robotic arm consists of servo
systems has made it possible to remotely control robots and motors controlled by the microcontroller. The
monitor their status via the internet. This trend has been number of degrees of freedom (DOF) of the arm is
accelerated by the proliferation of Wi-Fi and Bluetooth dependent on the number of servo motors. In this
technologies. A significant body of work in IoT-based system, a 4-DOF robotic arm is used, providing
robotic systems has focused on creating networked systems motion along the X, Y, and Z axes as well as
that can operate in distributed environments. In these rotation.
systems, a microcontroller like the ESP32 can serve as the
bridge between the sensor inputs (in this case, the hand 3.2 Gesture Recognition
gesture recognition system) and the robotic hardware. The gesture recognition process is performed in real-time
In their work, Singh et al. (2021) proposed an IoT-enabled using OpenCV and MediaPipe. The steps involved in the
robotic system where an ESP32 was used to control a gesture recognition pipeline are as follows:
robotic arm based on gestures detected by a smartphone 1. Input: The system uses a standard web camera or
application. While this system relied on mobile devices for USB camera to capture video frames. The frames
gesture input, it highlighted the potential of IoT-based are processed in real-time.
robotics to enable flexible and scalable robotic control. 2. Hand Detection: MediaPipe Hand Module is used
Another example is the work of Huang et al. (2022), where to detect and track hand landmarks. The hand
they developed a cloud-based robotic arm system that uses landmarks are key points on the palm and fingers,
IoT protocols such as MQTT to send control commands to totaling 21 points that represent the hand’s position
a robotic arm over Wi-Fi. This system allowed users to and motion.
control a robot from anywhere in the world, emphasizing
the power of IoT for long-range robotic operations.
2
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits
3
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits
Microcontroller
Servo Motors (x4)
Robotic Arm Frame
Where: Camera
θ\thetaθ is the desired servo angle (0° to 180°)
Power Supply
PWM Duty Cycle Calculation:
Jumper Wires
4
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits
V.METHODOLOGY
5
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits
VIII.REFERENCE.
6
IEEE Journal on Exploratory Solid-State Computational Devices and Circuits