The paper presents a comprehensive protocol for simultaneously recording hand-electromyography (EMG) and visual finger tracking during natural finger gesturing. The visual data is designed to serve as the ground truth for the development of accurate EMG-based computational models for finger gesture recognition.
Finger gestures are a critical element in human communication, and as such, finger gesture recognition is widely studied as a human-computer interface for state-of-the-art prosthetics and optimized rehabilitation. Surface electromyography (sEMG), in conjunction with deep learning methods, is considered a promising method in this domain. However, current methods often rely on cumbersome recording setups and the identification of static hand positions, limiting their effectiveness in real-world applications. The protocol we report here presents an advanced approach combining a wearable surface EMG and finger tracking system to capture comprehensive data during dynamic hand movements. The method records muscle activity from soft printed electrode arrays (16 electrodes) placed on the forearm as subjects perform gestures in different hand positions and during movement. Visual instructions prompt subjects to perform specific gestures while EMG and finger positions are recorded. The integration of synchronized EMG recordings and finger tracking data enables comprehensive analysis of muscle activity patterns and corresponding gestures. The reported approach demonstrates the potential of combining EMG and visual tracking technologies as an important resource for developing intuitive and responsive gesture recognition systems with applications in prosthetics, rehabilitation, and interactive technologies. This protocol aims to guide researchers and practitioners, fostering further innovation and application of gesture recognition in dynamic and real-world scenarios.
Hand gesturing is essential in human communication, making the recognition of finger gestures a crucial area of research across fields such as human-computer interaction, advanced prosthetics1,2,3,4, and rehabilitation technologies5,6. As a result, finger gesture recognition has garnered significant attention for its potential to enhance intuitive control systems and assistive devices. Surface electromyography (sEMG) combined with deep learning algorithms is emerging as a highly promising approach for capturing and interpreting these gestures due to its ability to detect the electrical activity of muscles associated with hand movements7,8,9,10,11,12,13,14,15.
However, despite these advances, current approaches face limitations in real-world applications. Most existing systems require complex, cumbersome recording setups with numerous electrodes5,7,9,16,17 and precise positioning3,18, which are often difficult to implement outside of controlled environments. Additionally, these systems tend to focus on static hand positions13,18,19,20,21, limiting their ability to interpret dynamic, fluid gestures that occur in daily activities. The protocol aims to address these limitations by supporting dynamic gesture recognition in more natural conditions. Such methodology would enable more practical and user-friendly applications in areas like prosthetics and rehabilitation, where real-time, natural gesture interpretation is essential.
To address these challenges, developing more accurate and adaptable algorithms requires datasets that reflect natural, everyday conditions3,4. Such datasets must capture a wide range of dynamic movements, various hand positions, and large volumes of data to ensure model robustness. Furthermore, the variability between training and test datasets is crucial, allowing models to generalize across different hand postures, muscle activation patterns, and motions. Incorporating such diversity into the data will enable algorithms to perform gesture recognition more accurately in everyday, real-world applications22.
Overcoming these challenges will be essential for the future development of more practical and widely applicable gesture recognition systems. The study and protocol described here stem from the need to have a portable, user-friendly setup that can capture dynamic hand movements in natural settings. Comprehensive datasets and advanced algorithms are critical to fully unlocking the potential of sEMG and deep learning in human-computer interfaces, neuroprosthetics, and rehabilitation technologies. We expect this protocol to contribute to the field by facilitating comprehensive data collection to further enable the development of algorithm models that generalize across diverse hand positions.
A significant challenge in gesture recognition lies in the sensitivity of sEMG signals to hand positioning. While many studies focus on fixed-hand positions for gesture prediction, real-world applications demand models capable of recognizing finger movements across various hand postures. Recent approaches have addressed this by incorporating computer vision as a ground truth reference, enhancing the accuracy and flexibility of these models15,19. Additionally, hybrid models that integrate sEMG signals with visual data offer further improvements in recognition accuracy across diverse scenarios23.
In this protocol, we present a synchronized approach to data collection that enhances dynamic gesture recognition by incorporating both EMG and hand-tracking data in real-world-like conditions. Unlike traditional methods that restrict gesture performance to static positions, this protocol includes gestures performed across four distinct positions: hand down, hand up, hand straight, and hand moving. The hand-tracking camera tracks hand movements within a three-dimensional interactive zone, identifying distinct hand elements and capturing dynamic movements with high resolution. A soft electrode array of 16 electrodes placed on the forearm to record muscle activity offers stable and wireless recordings without impeding participant mobility. The synchronized data from these two sources provides a comprehensive foundation for developing advanced gesture recognition algorithms capable of operating in real-world conditions. The approach specifically addresses the limitations of current setups by facilitating free movement and stable signal recording in realistic scenarios. This advancement supports gesture recognition technologies for applications in prosthetics, rehabilitation, and interactive technologies, where intuitive control and flexibility are essential.
Healthy participants (n = 18, aged 18-32 years, both males and females) were recruited for this study, which was approved by the Tel Aviv University Ethics Review Board (Approval No. 0004877-3). The protocol adheres to the board's guidelines for research involving human participants. Informed consent was obtained from all participants in accordance with institutional requirements.
1. Experimenter briefing
Figure 1: Schematic representation of the data collection process. The subject is equipped with a soft electrode array placed on the forearm (3), which captures high-resolution surface electromyography (sEMG) signals during gesture performance. The subject performs 14 different finger gestures presented in random order on a computer display (4). The EMG data is streamed wirelessly to a personal computer (PC) from the data acquisition unit (DAU; 1). Simultaneously, hand kinematic data (HKD) representing finger joint angles is captured using a hand-tracking camera (2). Please click here to view a larger version of this figure.
2. Setting up the data acquisition units
3. Participant preparation
4. Data collection
5. End of experiment and post-experiment data handling
The dataset consists of two time-synchronized components: a 16-channel EMG dataset and data from a hand-tracking camera system. The 16-channel EMG data captures muscle activity by recording electrical signals from different muscles over time. The hand-tracking system provides 16 channels of data corresponding to key points on a skeletal model of the hand. While the model has 21 points, excluding the wrist, this number was reduced to 16 due to motion constraints24. The EMG and visual data were collected by running two separate processes on the same computer during recording to establish synchrony. A timestamp was used to mark the start of each process, allowing data analysis code to align muscle activity and hand movement data at the end of the recording. Timestamp annotations were saved automatically in both EDF and CSV files, marking the exact time when specific finger gestures were instructed and facilitating the alignment during data analysis. The filtered EMG signal (20 Hz 4th-order Butterworth high pass filter) is characterized by a low baseline (grey-shaded areas), which typically falls within the range of 3-9 µV25. This baseline is observed when the subject's hand is stationary and the muscles are at rest. However, if muscle tone is present even in the rest position, a distinct EMG signal can be detected. Mechanical artifacts caused by movement usually manifest in the 10-20 Hz range and should be filtered out accordingly. Significantly elevated baseline values may indicate 50 Hz line interference and should be avoided during the experimental setup stage. In cases where moderate 50 Hz noise persists; a notch filter is applied. Sharp movement artifacts, which are more difficult to remove, often appear as pronounced high-amplitude spikes in the signal (see asterisk in Figure 2A). The amplitude of the EMG signal across the 16-electrode array varies, reflecting the spatial distribution of the muscle activity over the region measured. This variance provides valuable insight into the heterogeneity of muscle contraction during hand gestures.
The hand-tracking camera provides direct information of finger angles (hand kinematic data, HKD), which are expected to correlate closely with the recorded EMG signals. During gestures, finger angles in normal range26, depending on the specific gesture. When the visual path between the hand-tracking camera and the hand is unobstructed, the resulting signal is stable and accurate, as demonstrated in Figure 2. However, in instances where visual contact is lost or when the system experiences technical limitations, the HKD may become erratic, displaying jumps between incorrect values. Such outlier data should be minimized during data collection and discarded in the final analysis to maintain the integrity of the results.
The HKD is intuitive and provides a direct comparison with the actual gestures performed. It exhibits low variability between subjects and across different hand positions. In contrast, the EMG data tends to vary significantly between individuals due to anatomical differences such as hand size and muscle development27. Additionally, variability may be observed between dominant and non-dominant hands. This subject-specific variability can be addressed during offline analysis.
In Figure 2, it is evident that both the EMG and HKD are offset relative to the instructed gesture trigger. This discrepancy arises due to response time and the natural movement execution28. In regression tasks, such variability could contribute to the richness of the data, while in classification tasks, it can be managed using a generalized likelihood ratio approach, as applied in similar scenarios28.
Figure 2:Â Representative sEMG and HKD during finger abduction. Surface electromyography (sEMG) signals and hand kinematic data (HKD) recorded during dynamic finger abduction and rest performed during hand position 1 (hand down, straight, and relaxed) by a single participant. (A) Filtered EMG signals from 16 channels as a function of time. Asterisk (*) denotes a mechanical artifact detected in the EMG recording of Channel 5. (B) HKD, showing the joint angles as a function of time. Joint angles are measured at various joints: trapeziometacarpal (TMC), metacarpophalangeal (MCP), and proximal interphalangeal (PIP). The phases of the experiment (rest and abduction) are indicated along the x-axis. Please click here to view a larger version of this figure.
These representative results demonstrated the utility of the synchronized EMG and HKD data in capturing hand gestures. The alignment of EMG signals with corresponding HKD allows mapping muscle activity to specific finger movements. When constructing a predictive model, researchers can use HKD as ground truth, iteratively verifying and refining EMG-based gesture predictions. This approach highlights the practical applicability of the protocol and suggests the need for further research in more natural settings.
Supplementary Figure 1: Spectrogram windows displayed during the signal verification step. The left panels show raw EMG data, while the right panels show the detected frequency domains. (A) Example of a very noisy EMG signal with strong 50 Hz and 100 Hz interference. (B) Example of the same EMG signal recording after moving the participant further away from electrical devices, resulting in a clean EMG signal with minimal interference. Please click here to download this File.
The protocol presented in this study outlines critical steps, modifications, and troubleshooting strategies aimed at enhancing hand gesture recognition through the combination of sEMG signals and HKD. It addresses key limitations and compares this approach to existing alternatives, highlighting its potential applications in various research domains. One of the most important aspects of the protocol is ensuring the correct positioning and alignment of the hand-tracking camera. Accurate gesture capture is highly dependent on the angle and distance of the camera relative to the participant's hand. Even slight deviations in camera positioning can lead to tracking inaccuracies, reducing the fidelity of the gesture data. This alignment must be carefully adjusted for each participant and hand position to ensure consistent and reliable data collection. Additionally, it is crucial that participants are well-acquainted with the protocol to prevent junk data - where gestures are either incorrectly executed or misaligned with the experimental flow. Ensuring that participants are comfortable and familiar with the gestures and the experimental setup can minimize data noise and improve the quality of the recordings.
A common challenge in this type of study is noise contamination in both sEMG and HKD. sEMG signals are particularly sensitive to factors such as muscle fatigue, motion artifacts, and environmental noise like electromagnetic interference. Pre-processing techniques, such as band-pass filtering, are essential for reducing noise and improving signal clarity. Proper electrode placement and instructing participants to maintain relaxed muscles during rest phases can further mitigate motion artifacts. Despite these precautions, some variability in sEMG signals is inevitable due to individual differences in anatomy, hand strength, and muscle activation patterns. This variability can be addressed through flexible algorithms capable of normalizing these differences across subjects and conditions.
A key factor in achieving high-quality sEMG signals is initial signal verification. Traditional protocols using gel electrodes require skin preparation, such as exfoliating or cleaning with alcohol, to improve signal clarity. However, in a previous study we showed that with dry electrodes, skin preparation may not significantly impact signal quality25. In this protocol, skin cleaning is optional and thus simplifies the process. Another skin-related issue affecting signal quality is excessive and thick arm hair. In such cases, we suggest either shaving the area or excluding the subject from the study.
One of the critical challenges in using sEMG for gesture recognition is its sensitivity to hand positioning. Even when performing the same gesture, variations in hand orientation can lead to different EMG signal patterns. To address this issue, machine learning models that can accommodate variability in hand positions are essential22. These models must be trained with data from multiple hand postures to improve robustness and generalizability. Synchronization of visual and sEMG data is another important consideration. Consistent timing of gestures is critical to avoid discrepancies between the gesture execution and the data recording. This protocol uses visual countdowns and auditory cues to help ensure accurate timing and recalibration steps are employed when necessary to correct any misalignment during data collection.
Despite its strengths, this protocol has several limitations. One major constraint is the limited field of view of the hand-tracking camera, which requires the participant's hands to remain within the camera's detection range. This restricts the analysis to a small set of movements. For outside the lab experiments a more complex video imaging will be required or the use of smart gloves. Participant fatigue also poses a challenge during longer sessions, potentially affecting gesture accuracy and muscle activation, which can degrade the quality of the sEMG data. To mitigate these effects, it may be necessary to limit the session length or introduce breaks to minimize fatigue. Additionally, powerline interference can introduce noise into the sEMG signals, particularly when the participants are close to the PC for data capture. A wireless version of the system could reduce such interference by allowing participants to be farther from the computer.
A significant methodological limitation of EMG-based finger gesture detection stems from the high inter-subject variability in sEMG signals, which requires the development of custom models for each participant. This subject-specific approach, while more accurate, limits the protocol's scalability and requires additional calibration and training time for each new user. EMG and HKD data streams show minor temporal synchronization differences due to dual process recording. These timing discrepancies have a minimal impact on the static gesture analysis since the maintained poses are temporally stable. The sustained nature of static gestures provides adequate time for both EMG and kinematic features to stabilize, unlike dynamic gestures, which require more precise synchronization.
A key advantage of this method is its flexibility in capturing gestures. Unlike other systems that require rigid setups and strict gesture parameters, this protocol accommodates dynamic and flexible hand positions19. This flexibility is especially useful in studies aimed at analyzing a broad range of motions, making it more adaptable to real-world applications. Furthermore, this protocol is cost-effective compared to more advanced motion capture and sEMG systems, which often involve complex setups29. By integrating a hand-tracking camera with semi-automated sEMG algorithms, this method provides a viable alternative for gesture recognition studies without compromising data quality. Additionally, the system's potential for real-time data processing opens possibilities for immediate feedback in applications such as neuroprosthetics and rehabilitation, where real-time responsiveness is essential. This protocol has significant implications for several fields, particularly neuroprosthetics. Accurate prediction of hand gestures from sEMG signals is crucial for controlling prosthetic limbs, and the flexibility in hand positioning offered by this method makes it an ideal candidate for real-time prosthetic devices. In rehabilitation, this protocol could be employed to monitor and enhance motor recovery in patients with hand or finger impairments. By analyzing muscle activation patterns during gesture performance, this system could be used to tailor rehabilitation exercises to individual needs, offering a personalized approach to motor recovery. For human-computer interaction (HCI), this method enables more natural gesture-based control systems, improving the intuitiveness and efficacy of user interfaces. Lastly, the protocol could be applied to ergonomic studies to assess how different hand positions and gestures influence muscle activity and fatigue, potentially leading to advancements in workplace design and user ergonomics.
To ensure consistent contraction strength across participants, future studies could implement a glove with force-sensitive resistors to measure force directly. This would allow for standardized effort across subjects, improving the reliability of EMG data. Additionally, integrating this force measurement as a label in joint kinematics would provide a more detailed representation of the muscle's internal state, potentially enriching the analysis of muscle function and movement patterns. This approach would not only enhance data consistency but also offer deeper insights into the relationship between muscle contraction and joint motion.
In conclusion, this protocol provides a novel and flexible approach to hand gesture recognition with broad applications across neuroprosthetics, rehabilitation, HCI, and ergonomics. Although the system has limitations, its flexibility, cost-effectiveness, and potential for real-time use represent substantial advancements over existing methods. These strengths make it a promising tool for further development and innovation in gesture recognition technologies.
Yael Hanein declares a financial interest in X-trodes Ltd, which commercialized the screen-printed electrode technology used in this paper. The other authors have no other relevant financial involvement with any organization or entity with a financial interest in or financial conflict with the subject matter or materials discussed in the manuscript apart from those disclosed.
This project was partially funded with a grant from the ERC (OuterRetina) and ISF. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. We thank David Buzaglo, Cheni Hermon, Liron Ben Ari and Adi Ben Ari for their assistance with designing the original version of the protocol.
Name | Company | Catalog Number | Comments |
Adjustable Selfie Stick | Used to position and angle the hand-tracking camera in the desired orientation for optimal data capture during the experiment. | ||
Alcohol pad | To clean the area for electrode placement. | ||
Data acquisition unit (DAU) | X-trodes Ltd. | Â XTR-BT V1.3 | Realtime compatible DAU 2.0.17 or 4.0.1 |
Finger Gestures Recognition library | https://github.com/NeuroEngLabTAU/Fingers_Gestures_Recognition.git | ||
Leap Motion Controller 2 | Ultraleap | 129-SP4-00005-03 | Hand-tracking camera |
Long Type-C to Type-C cable | Connection of the hand-tracking camera to the PC. | ||
PC Monitors | One for guidelines, one for viewing the hand-tracking camera data | ||
Personal Computer (PC) | Windows | Windows 10+; Processors: Inteli7 processor. BT receiver. | |
Python code | A script enabling seamless data streaming and recording up to 500 S/s when DAU is connected to PC via Bluetooth | ||
Ultraleap Camera Python API | Ultraleap | Python API and instructions from Ultraleap’s GitHub repository (https://github.com/ultraleap/leapc-python-bindings) used to collect data from the Ultraleap unit during the experiment | |
Ultraleap Hyperion | Ultraleap | Tracking software | |
XTR EXG16 | X-trodes Ltd. | XTELC0003405RM | 16-channel wearable dry electrode array patch for EMG, ECG monitoring |
X-trodes PC App | X-trodes Ltd. | 1.1.35.0 | An application for connecting X-trodes DAU to PC via BT |
Request permission to reuse the text or figures of this JoVE article
Request PermissionThis article has been published
Video Coming Soon
Copyright © 2025 MyJoVE Corporation. All rights reserved