Abstract
Control of reach-to-grasp movements for deft and robust interactions with objects requires rapid sensorimotor updating that enables online adjustments to changing external goals (e.g., perturbations or instability of objects we interact with). Rarely do we appreciate the remarkable coordination in reach-to-grasp, until control becomes impaired by neurological injuries such as stroke, neurodegenerative diseases, or even aging. Modeling online control of human reach-to-grasp movements is a challenging problem but fundamental to several domains, including behavioral and computational neuroscience, neurorehabilitation, neural prostheses, and robotics. Currently, there are no publicly available datasets that include online adjustment of reach-to-grasp movements to object perturbations. This work aims to advance modeling efforts of reach-to-grasp movements by making publicly available a large kinematic and EMG dataset of online adjustment of reach-to-grasp movements to instantaneous perturbations of object size and distance performed in immersive haptic-free virtual environment (hf-VE). The presented dataset is composed of a large number of perturbation types (10 for both object size and distance) applied at three different latencies after the start of the movement.
Subject terms: Motor control, Sensory processing
Measurement(s) | kinematics • reach-to-grasp movements |
Technology Type(s) | motion capture system • virtual reality • electromyography |
Factor Type(s) | movement time [ms] • peak transport velocity [cm/s] • time to peak transport velocity [ms] • peak transport acceleration [cm/s2] • time to peak transport acceleration [ms] • peak transport deceleration [cm/s2] • time to peak transport deceleration [ms] • peak aperture [cm] • peak aperture velocity [cm/s] • time to peak aperture velocity [ms] • peak aperture deceleration [cm/s2] • time to peak aperture deceleration [ms] • opening time [ms] • closure time [ms] • opening distance [cm] • closure distance [cm] • transport velocity at CO [cm/s] • transport acceleration at CO [cm/s2] • peak closure velocity [cm/s] • peak closure deceleration [cm/s2] |
Sample Characteristic - Organism | Homo sapiens |
Machine-accessible metadata file describing the reported data: 10.6084/m9.figshare.16786258
Background & Summary
Rarely do we appreciate the remarkable coordination involved in our routine reach-to-grasp movements, until control becomes impaired by neurological injuries such as stroke, neurodegenerative diseases, or even aging. Despite seemingly effortless execution, a simple reach-to-grasp movement involves complex, multi-level collective interaction of the brain, spinal cord, and the peripheral system which is tuned and coordinated through sensory experience. Efficient and flexible behavior in everyday context requires rapid online adjustments of reach-to-grasp to sudden changes in the position or affordances of the target object1–10. Accurate physical modeling of reach-to-grasp movement could advance applications in neurorehabilitation11–13, neural prostheses14–16, and robotics17,18. Critically, none of the numerous approaches to modeling of the coordination between reach and grasp components have been able to accurately replicate human behavior19–21. To advance our understanding of how reach-to-grasp movements are orchestrated and updated, the scientific community needs to turn to more sophisticated forms of characterizing this complex motor behavior. Large publicly available datasets of hand movements of grasping 3D objects22–26 recorded using video/Kinect/infrared motion capture have immensely benefited modeling efforts in grasp classification. However, these datasets are often collected with the explicit purpose of training robotic grasping and are not optimized for modeling of human manual behavior. Furthermore, there is no publicly available dataset that includes online adjustment of reach-to-grasp movements to perturbations, freely accessible to researchers from multiple fields for modeling and analytical means. In the absence of such a dataset, the existing models of reach-grasp coordination19–21,27–32—which typically rely on data collected under a limited set of task manipulations—have remained untested. A reach-to-grasp dataset that offers synchronized kinematic and electromyography (EMG) data for a broader set of conditions, including coordinated reach and grasp responses to perturbations of the task goal, would greatly aid future efforts directed toward modeling of reach-to-grasp movements.
The purpose of this report is to make publicly available rich dataset of kinematic and EMG collected during hand and arm movements as participants reached to grasp objects in an immersive haptic-free virtual environment (henceforth, hf-VE) that includes a large variety of object size and distance perturbations that required rapid online adjustments of movement to compensate for instantaneous perturbations of the goal. Moreover, in the interest of being able to study the extent to which there may be temporal dependencies related to the perturbation timing, the perturbations occurred at three different latencies after movement onset. This dataset has been collected using the state-of-the-art experimental setup developed over several years in the Movement Neuroscience Laboratory at Northeastern University, which includes the seamless integration of an immersive haptic-free virtual reality system, an active marker motion capture system, a wireless multichannel electromyography (EMG) recording system, and an immersive unity 3D-programmed hf-VE programmed in C# and Python. The hardware and software renderings were synchronized to obtain a ~13.3 ms feedback loop between participants’ hand movements and their virtual rendering corresponding to 75 Hz sampling of kinematic data (all kinematic data is provided after resampling to 100 Hz). The kinematic data primarily pertain to the transport and aperture aspects of reach-to-grasp movements. Whereas ‘transport’ refers to the motion of the hand towards the target object, aperture refers to the distance between the tips of the thumb and index finger that forms the enclosure around the target object. The dataset is organized as a Matlab (Mathworks Inc., Natick, MA) data structure (.mat) with kinematic and EMG data. The dataset’s novelty lies in a large number of conditions, including Perturbation Type (object size perturbation, object distance perturbation), Perturbation Timing (100 ms, 200 ms, and 300 ms after movement onset), and the combination of synchronized kinematic and EMG acquisition.
To our knowledge, no dataset exists for online adjustment of coordinated reach-to-grasp movements to perturbations of the task goal. Although some reach and grasp kinematics datasets (e.g., performing different reaches and grasps) are available, as well as forearm EMG datasets (e.g., performing hand gestures or freehand movements), they have several limitations for use in modeling online control of reach-to-grasp movements:
A focus on isolated grasp: The available datasets focus solely on the grasp component with little to no mention of the reach component, or coordinated reach-to-grasp movements33–39.
Limited sample size: Some of the existing datasets provide data only from a small number of participants (e.g., just one to four participants35,40, as opposed to a total of 20 participants in the present dataset, ten participants each for object size and distance perturbations), limiting their generalizability and the ability of modeling efforts to make generalizable inferential predictions.
No synchronization between kinematic and EMG data: The available datasets offer either kinematics38,39 or EMG41,42 data but do not offer synchronized kinematic and EMG data.
This dataset overcomes the above-mentioned limitations. It is our hope that the data will be useful for modeling coordinated reach-to-grasp movements for both the basic and applied aspects of research. The present dataset consists of a Matlab/GNU Octave data structure (in.mat) with kinematic and EMG data (including maximal voluntary contraction or MVC for each muscle from which EMG was recorded). A separate.csv file contains sex, age, anthropometry data and laterality for all participants.
Methods
Participants and ethical requirements
Ten adults (eight men and two women; mean ± 1s.d. age: 22.5 ± 6.0 years, right-handed) participated in the size-perturbation study, and ten adults (eight men and two women; mean ± 1s.d. age: 25.3 ± 6.4 years, right-handed) participated in the distance-perturbation study. The participants were free of any muscular, orthopedic, or neurological health concerns. The participant pool comprised undergraduate and graduate students at Northeastern University. The participants were offered $10 per hour for participation. Each participant provided verbal and written consent approved by the Institutional Review Board (IRB) at Northeastern University. Some participants had previously participated in reach-to-grasp studies in our hf-VE, however, none of the participants reported extensive experience in virtual reality (e.g., gaming, simulations, etc.). To ensure adequate familiarization with the reach-to-grasp task in a virtual environment, all participants completed a training block of 120 reach-to-grasp trials [24 trials per object size: small (w × h × d = 3.5 × 8 × 2.5 cm), small-medium (4.5 × 8 × 2.5 cm), medium (5.5 × 8 × 2.5 cm), medium-large (6.5 × 8 × 2.5 cm), and large (7.5 × 8 × 2.5 cm) object placed at 30 cm; or 24 trials per object distance: medium object placed at near (20 cm), near-middle (25 cm), middle (30 cm), middle-far (35 cm), and far (40 cm) distances]. If participants felt comfortable after 60 trials, the training block was terminated, and the experimental trials began; otherwise, the participants completed all 120 training trials.
Reach-to-grasp task, virtual environment, and kinematic/kinetic measurement
Participants reached to grasp virtual objects of different sizes and placed at different distances from the starting position in an immersive hf-VE developed in UNITY (ver. 5.6.1f1, 64-bit, Unity Technologies, San Francisco, CA) and delivered via an Oculus head-mounted display (HMD; Rift DK2, Oculus Inc., Menlo Park, CA; Fig. 1) using HANDoVR (Movement Neuroscience Laboratory, Northeastern University, Boston, MA). An eight-camera motion tracking system (sampling rate: 75 Hz; PPT Studio NTM, WorldViz Inc., Santa Barbara, CA) recorded the 3D motion of IRED markers attached to the participants’ wrist (at the center of the segment running between the ulnar and radial styloid processes), and the tips of the thumb and index finger. A pair of IRED markers were attached to the HMD to co-register the participant’s head motion to the virtual environment. Participants viewed the thumb and index fingertips as two 3D spheres (green in color, 0.8 cm diameter) in the hf-VE, reflecting the 3D position of the respective IRED marker. The schedule of trials, virtual renderings of objects, and timing/triggering of perturbations were controlled using custom software developed in C# and Python.
EMG recordings
EMG activity (in µV) was recorded from the following ten muscles of each participant’s shoulder, arm and hand on the dominant right side. EMG was acquired from the first dorsal interosseous (FDI), flexor digitorum superficialis (FDS), extensor digitorum communis (EDC), extensor indicis (EI), abductor pollicis brevis (APB), extensor pollicis brevis (EPB), biceps brachii (BB), triceps brachii (TB), anterior deltoid (AD), and posterior deltoid (PD).
EMG was recorded using a Delsys TrignoTM wireless EMG system (sampling rate: 1 kHz; Delsys Inc., Natick, MA). Surface EMG sensor bars were attached perpendicular to the muscle fibers over the muscle belly. Excess hair was shaved, and the skin prepped/cleaned with isopropyl alcohol pads before attaching the sensors to reduce skin impedance. Proper positioning of EMG sensors was ensured by physically palpating the muscle during sustained isometric contraction and visual confirmation of the EMG signal. EMG activity during MVC was saved and is included in the dataset. Kinematic and EMG data were synchronized using a 5V digital output (10 ms) sent from Unity and recorded as an analog signal synchronously with EMG.
Synchronization between EMG data and kinematics data
Details of the synchronization between EMG data and kinematics data are as follows: EMG data were collected using custom software in Matlab to communicate with a Multifunctional I/O Device (NI6255; National Instruments Inc., Austin, TX). Analog data from the Delsys wireless EMG system were streamed to the NI6255 with a known constant 47 ms delay. Kinematic data were collected using C# and Unity-based HANDoVR, described above. HANDoVR software communicated with a National Instrument Multifunctional I/O Device (NI6211). Upon detecting switch release, HANDoVR triggered a 5 V digital (10 ms) output from the NI6211. This digital output was connected to an analog input channel on the NI6255 and recorded into Matlab. EMG and Kinematic data sets were aligned via the start switch trigger recorded digitally in HANDoVR, and the analog reading of the digital output sent from HANDoVR to Matlab. Misalignment of the kinematic data and EMG was constrained to the sampling period of the kinematic data (~11.1 ms), as we were unable to estimate when in the inter-sample period the digital output was sent with respect to when the motion capture sensors were read. Hardware delays were tested to be less than the sampling period of the EMG recording (1 ms).
Schedule of trials and visual perturbations of object size and distance
Participants were tested in a single experimental session that lasted up to 180 min. Before data collection, participants were allowed to practice the reach-to-grasp task (unperturbed) until they felt comfortable with the task. Then the experiment began, consisting of a total of 960 reach-to-grasp trials, each trial lasting 3.5 s (480 no-perturbation and 480 perturbation trials). The trials were conducted over four sessions of 240 trials each (120 no-perturbation and 120 perturbation trials). The order of no-perturbation and perturbation trials were randomized differently in each session. We ensured that each type of no-perturbation and perturbation trial was evenly distributed across the four sessions. A 5-min break was given after each block and whenever the participant expressed a need to do so.
Size perturbation:
Table 1 tabulates the breakdown of 960 trials.
Table 1.
Condition name | No perturbation | Object size S [cm] | Object distance D [cm] | # Trials | |
---|---|---|---|---|---|
Small | S | 3.5 | 30 | 96 | |
Small-Medium | SM | 4.5 | 30 | 96 | |
Medium | M | 5.5 | 30 | 96 | |
Medium-Large | ML | 6.5 | 30 | 96 | |
Large | L | 7.5 | 30 | 96 | |
Condition name | Perturbation | Object size S [cm] | Object distance D [cm] | Perturbation timing [ms] | # Trials |
Small→Small-Medium | S→SM | 3.5→4.5 | 30 | 100, 200, 300 | 16 |
Small→Medium | S→M | 3.5→4.5 | 30 | 100, 200, 300 | 16 |
Small→Medium-Large | S→ML | 3.5→6.5 | 30 | 100, 200, 300 | 16 |
Small→Large | S→L | 3.5→7.5 | 30 | 100, 200, 300 | 16 |
Small-Medium→Medium | SM→M | 4.5→5.5 | 30 | 100, 200, 300 | 16 |
Small-Medium→Medium-Large | SM→ML | 4.5→6.5 | 30 | 100, 200, 300 | 16 |
Small-Medium→Large | SM→L | 4.5→7.5 | 30 | 100, 200, 300 | 16 |
Medium→Medium-Large | M→ML | 5.5→6.5 | 30 | 100, 200, 300 | 16 |
Medium→Large | M→L | 5.5→7.5 | 30 | 100, 200, 300 | 16 |
Medium-Large→Large | ML→ L | 6.5→7.5 | 30 | 100, 200, 300 | 16 |
The 480 no-perturbation (control) trials were evenly distributed among objects of five different sizes (96 trials per object): small (w × h × d = 3.5 × 8 × 2.5 cm), small-medium (4.5 × 8 × 2.5 cm), medium (5.5 × 8 × 2.5 cm), medium-large (6.5 × 8 × 2.5 cm), and large (7.5 × 8 × 2.5 cm) placed at the same distance of 30 cm from the starting position of the participant’s thumb and index finger. The 96 trials for each object size were evenly distributed across the four blocks (24 trials per block).
The 480 perturbation trials were evenly distributed among ten possible combinations of object size changes such that the object’s width increased from the object’s initial size to a larger size (48 trials per perturbation type). The perturbation types included: small (S) to small-medium (SM), small to medium (M), small to medium-large (ML), small to large (L), small-medium to medium, small-medium to medium-large, small-medium to large, medium to medium-large, medium to large, and medium-large to large (all perturbations from smaller to larger objects). Each perturbation type was applied at three different latencies: 100 ms after movement onset, 200 ms after movement onset, and 300 ms after movement onset, resulting in 16 trials for each perturbation type applied at each of the three timings. The 16 trials for each perturbation type and timing were evenly distributed across the four blocks (four trials per block).
Distance perturbation:
Table 2 tabulates the breakdown of 960 trials.
Table 2.
Condition name | No perturbation | Object size S [cm] | Object distance D [cm] | # Trials | |
---|---|---|---|---|---|
Near | N | 5.5 | 20 | 96 | |
Near-Middle | NM | 5.5 | 25 | 96 | |
Middle | M | 5.5 | 30 | 96 | |
Middle-Far | MF | 5.5 | 35 | 96 | |
Far | F | 5.5 | 40 | 96 | |
Condition name | Perturbation | Object size S [cm] | Object distance D [cm] | Perturbation timing [ms] | # Trials |
Near→Near-Middle | N→NM | 5.5 | 20→25 | 100, 200, 300 | 16 |
Near→Middle | N→M | 5.5 | 20→30 | 100, 200, 300 | 16 |
Near→Middle-Far | N→MF | 5.5 | 20→35 | 100, 200, 300 | 16 |
Near→Far | N→F | 5.5 | 20→40 | 100, 200, 300 | 16 |
Near-Middle→Middle | NM→M | 5.5 | 25→30 | 100, 200, 300 | 16 |
Near-Middle→Middle-Far | NM→MF | 5.5 | 25→35 | 100, 200, 300 | 16 |
Near-Middle→Far | NM→F | 5.5 | 25→40 | 100, 200, 300 | 16 |
Middle→Middle-Far | M→MF | 5.5 | 30→35 | 100, 200, 300 | 16 |
Middle→Far | M→F | 5.5 | 30→40 | 100, 200, 300 | 16 |
Middle-Far→Far | MF→F | 5.5 | 35→40 | 100, 200, 300 | 16 |
The 480 no-perturbation (control) trials were evenly distributed among objects (all 5.5 × 8 × 2.5 cm) placed at five different distances (96 trials per object): near (20 cm), near-middle (25 cm), middle (30 cm), middle-far (35 cm), and far (40 cm). The 96 trials for each object distance were evenly distributed across the four blocks (i.e., 24 trials per block).
The 480 perturbation trials were evenly distributed among ten possible combinations of object distance changes such that the object’s distance increased from the object’s initial location to a farther location (48 trials per perturbation type). The perturbation types included: near (N) to near-middle (NM), near to middle (M), near to middle-far (MF), near to far (F), near-middle to middle, near-middle to middle-far, near-middle to far, middle to middle-far, middle to far, and middle-far to far (all permutations from closer to farther distances). Each perturbation type was applied at three different latencies: 100 ms after movement onset, 200 ms after movement onset, and 300 ms after movement onset, resulting in 16 trials for each perturbation type applied at each of the three timings. The 16 trials for each perturbation type and timing were evenly distributed across the four blocks (four trials per block). The reach-to-grasp animation of representative conditions (control, size and distance perturbations with 100 and 300 ms latencies) is available on the Figshare43.
Procedures and instructions to participants
Each participant was seated in a chair with their right hand placed on a table in front of them (Fig. 1). At the start position, the thumb and index finger straddled a 1.5 cm wide wooden peg located 12 cm in front and 24 cm to the right of the sternum, with the thumb depressing a start switch. Lifting the thumb off the switch marked movement onset. A digital transistor–transistor logic (TTL) connected to the start switch was used to synchronize kinematic and EMG recordings. In each trial, the following events occurred: (1) Participants depressed the start switch to begin the trial. (2) The object appeared in hf-VE, oriented at a 75° angle along the vertical axis to minimize excessive wrist extension during reach-to-grasp. (3) After 1 s, an auditory cue—a beep—signaled the participants to reach for, grasp, and lift the object with 1.2 cm combined error margin44. The object was considered to have been grasped when both 3D spheres (reflecting the tips of the thumb and index finger) had come in contact with the lateral surfaces of the virtual object. (4) Once grasp of the virtual object was detected, the object changed color from blue to red and a ‘click’ sound was presented. (5) Participants lifted and raised each object briefly before returning their hand to the starting position, after which the next trial began.
Instructions to the participant were: “Each trial will start once the thumb depresses the start switch (the correct initial position of the hand was demonstrated). Following the beep, reach-to and grasp the narrow sides of the object between the thumb and index finger using a pincer grip (demonstrated). When the object is grasped, it will turn from blue to red and a ‘click’ sound will be presented. Lift the object briefly until the object disappears, and return your hand to the start position. On some trials, the object may change “size” or “position”, requiring you to adjust your movements. A break would be provided after each block of 240 trials but you may rest at any point between trials within a block by not depressing the start switch to begin the next trial. Do you have any questions?”.
Data processing and kinematic feature extraction
The raw data included the x, y, and z marker positions of the wrist, thumb, and index finger positions with associated timestamps (75 Hz; this raw data is not provided in the data records). All position data were analyzed offline using custom Matlab codes. The time-series data for each trial were cropped from movement onset (the moment the switch was released) to movement offset (the moment the collision detection criteria were met) and resampled at 100 Hz using the interp1() function in Matlab. Transport distance (the straight-line distance of the wrist marker from the starting position in the transverse plane) and aperture (the straight-line distance between the thumb and index finger markers in the transverse plane) were computed for each trial. The first and second derivatives of transport displacement and aperture were computed to obtain the velocity and acceleration profiles for kinematic feature extraction. A 6 Hz, fourth-order low-pass Butterworth filter was applied on all time-series. Trails in which participants did not move, were delayed in moving, or had inappropriate movements were excluded from the database (i.e., bad trials). Link to the GitHub repository of custom code used to generate the data is available under Code Availability section.
For each trial, the following kinematic features, units in parentheses (Table 3), were extracted using the filtered time series data (Fig. 2):
Movement time [ms]—duration from movement onset to movement offset.
Peak transport velocity [cm/s]—maximum velocity of the wrist marker.
Time to peak transport velocity [ms]—time from movement onset to maximum velocity of the wrist marker.
Peak transport acceleration [cm/s2]—maximum acceleration of the wrist marker.
Time to peak transport acceleration [ms]—time from movement onset to maximum acceleration of the wrist marker.
Peak transport deceleration [cm/s2]—maximum deceleration of the wrist marker.
Time to peak transport deceleration [ms]—time from movement onset to maximum deceleration of the wrist marker.
Peak aperture [cm]—maximum distance between the fingertip markers. Peak aperture also marked the initiation of closure or Closure Onset (henceforth, CO) and which we refer to as the aperture at CO.
Peak aperture velocity [cm/s]—maximum velocity of the aperture.
Time to peak aperture velocity [ms]—time from movement onset to maximum velocity of the aperture before CO.
Peak aperture deceleration [cm/s2]—maximum deceleration of the aperture before CO.
Time to peak aperture deceleration [ms]—time from movement onset to maximum deceleration of the aperture before CO.
Opening time [ms]—duration from movement onset to peak aperture.
Closure time [ms]—duration from CO to movement offset.
Opening distance [cm]—distance between the wrist’s position at movement onset and the wrist’s position at CO.
Closure distance [cm]—distance between the wrist’s position at CO and the object’s center.
Transport velocity at CO [cm/s]—velocity of the wrist marker at the time of CO.
Transport acceleration at CO [cm/s2]—acceleration of the wrist marker at the time of CO.
Peak closure velocity [cm/s]—minimum velocity of the aperture after CO.
Peak closure deceleration [cm/s2]—maximum deceleration of the aperture after CO.
Table 3.
Feature name in data record | Kinematic feature | Unit |
---|---|---|
MT | Movement time | ms |
Peak_TV | Peak transport velocity | cm/s |
T_Peak_TV | Time to peak transport velocity | ms |
Peak_TA | Peak transport acceleration | cm/s2 |
T_Peak_TA | Time to peak transport acceleration | ms |
Peak_TD | Peak transport deceleration | cm/s2 |
T_Peak_TD | Time to peak transport deceleration | ms |
Peak_A | Peak aperture | cm |
Peak_AV | Peak aperture velocity | cm/s |
T_Peak_AV | Time to peak aperture velocity | ms |
Peak_AD | Peak aperture deceleration | cm/s2 |
T_Peak_AD | Time to peak aperture deceleration | ms |
OT | Opening time | ms |
CT | Closure time | ms |
OD | Opening distance | cm |
CD | Closure distance | cm |
TV_CO | Transport velocity at CO | cm/s |
TA_CO | Transport acceleration at CO | cm/s2 |
Before data collection, the maximal voluntary contraction (MVC) of each muscle was obtained. Muscle activation was recorded during each reach-to-grasp movement for 3.5 s. The data files accompanying this dataset contain MVCs and raw EMG beginning 500 ms before movement onset. EMG data from one participant (P6 - size perturbation only) were not saved correctly due to technical issues.
Data Records
All data is made available using Figshare45. All 10 participants for each type of perturbation have been identified using alphanumeric format P# in the folders “SizePert” and “DistPert” for object size and distance perturbation, respectively. The deidentified participant information (sex, age, body mass, and height) is stored in the Excel file named “Participants”. Kinematic and EMG data have been grouped into subject-specific folders, each folder bearing the participant’s alphanumeric code (e.g., P5 for the fifth participant). Within each participant folder, there are five.mat files: 1) Raw_Data; 2) Resampled data; 2) Kinematic profiles of position, velocity, and acceleration for both the transport and aperture; 3) Kinematic features; 5) MVC and raw EMG. Figure 3 illustrates the Matlab structure in which files 2–5 are saved, with the row and column vectors shown in red (also see Tables 3 and 4).
Raw_Data.mat—this file contains nine arrays:
Trial_Number—1 × 960 cell array with each cell containing the trial number of the respective trial in the same order in which that trial was conducted.
Trial_Status—1 × 960 cell array with each cell containing information about whether that trial was a bad trial, that is, the one in which the participant did not move, was delayed in moving or had inappropriate movement.
Condition_Names—1 × 960 string array of condition names indicating the size and distance of object, perturbation, and the timing of perturbation (see Table 4).
Variable_Names—1 × 10 string array of variable names and measurement units corresponding to each column in the data matrix for each trial in the ‘Raw_Trajectories’ array.
Raw_Trajectories—1 × 960 cell array with nth column corresponding to nth trial. Each cell of this array contains a t × 10 matrix, where t is the number of samples captured by the motion capture system at 75 Hz for the respective trial. The next nine columns of this matrix correspond to the time stamp and the x-, y-, and z- coordinates of the wrist, thumb, and index finger markers (see Variable_Names).
Onset—1 × 960 numeric array with each cell containing the timing of movement onset for the respective trial based on switch release.
Offset—1 × 960 numeric array with each cell containing the timing of movement offset for the respective trial based on collision detection.
Corrected_Onset—1 × 960 numeric array with each cell containing the manually-selected timing of movement onset for the respective trial used to crop data during post-processing. Each trial was visually inspected in the post-processing stage. Movement onset was corrected if the onset was delayed (i.e., exceeded 3% of peak aperture) or untimely marked (i.e., there was no change in aperture for >2 samples). Onset was corrected only for control trials.
Corrected_Offset—1 × 960 numeric array with each cell containing the manually-selected timing of movement offset for the respective trial used to crop data during post-processing. Each trial was visually inspected in the post-processing stage. Movement offset was corrected if the offset did not fall and remain below 3% of transport velocity of peak transport velocity.
Trajectories.mat—this file contains four arrays:
Resampled—35 × 1 cell array with each row corresponding to a different condition (see Condition_Names). Each cell of this array contains a 1 × nTrials cell array (nTrials = number of trials) with each cell containing data for an individual trial in 200 × 10 matrix. The columns of this matrix correspond to the time stamp and the x-, y-, and z- coordinates of the wrist, thumb, and index finger markers (see Variable_Names).
Condition_Names—35 × 1 string array of condition names indicating the size and distance of object, perturbation, and the timing of perturbation (see Table 4).
Variable_Names—1 × 10 string array of variable names and measurement units corresponding to each column in the data matrix for each trial.
Final_Object_Location—35 × 1 cell array with each cell containing the x-, y-, and z- coordinates of the final object position for each of the 35 conditions.
Profiles.mat—this file contains three arrays:
Profiles—3 × 2 cell array with the three rows corresponding to the position, velocity, and acceleration and the two columns corresponding to the transport and aperture. Each cell of this array contains a 35 × 1 cell array with each row corresponding to a different condition (see Condition_Names). Each cell of this array contains a 200 × nTrials matrix (nTrials = number of trials) with each column containing data for an individual trial.
Variable_Names—3 × 2 string array of variable names.
Condition_Names—35 × 1 string array of condition names indicating the size and distance of object, perturbation, and the timing of perturbation (see Table 4).
Features.mat—this file contains three arrays:
Features—18 × 1 cell array with each row corresponding to a different kinematic feature (see Feature_Names). Each cell of this array contains a 35 × 1 cell array with each row corresponding to a different condition (see Condition_Names). Each cell of this array is a 1 × nTrials matrix with the columns containing data for individual trials.
Feature_Names—18 × 1 string array with the names and measurement units of kinematic features (see Table 3).
Condition_Names—35 × 1 string array of condition names indicating the size and distance of object, perturbation, and the timing of perturbation (see Table 4).
EMG.mat—this file contains five arrays:
MVC—4000 × 11 matrix with the columns corresponding to the time stamp and EMG activity in each of the ten recorded muscles (see Variable_Names).
Variable_Names—1 × 11 string array of variable names and measurement units corresponding to each column in the data matrix for each trial.
Raw_EMG—35 × 1 cell array with each row corresponding to a different condition (see Condition_Names). Each cell of this array contains a 1 × nTrials cell array (nTrials = number of trials) with each cell containing data for an individual trial in a 4000 × 11 matrix. The columns of this matrix correspond to the time stamp and EMG activity in the ten recorded muscles (see Variable_Names).
Condition_Names—35 × 1 string array of condition names indicating the size and distance of object, perturbation, and the timing of perturbation (see Table 4).
Movement_Time—18 × 1 cell array with each row corresponding to a different kinematic feature (see Feature_Names). Each cell of this array contains a 35 × 1 cell array with each row corresponding to a different condition (see Condition_Names). Each cell of this array is a 1 × nTrials matrix with the columns containing movement time [ms] for individual trials.
Table 4.
Size perturbation | ||
---|---|---|
Condition name in data record* | Condition† | |
No perturbation | S_M | Small:Middle |
SM_M | Small-Medium:Middle | |
M_M | Medium:Middle | |
ML_M | Medium-Large:Middle | |
L_M | Large:Middle | |
Perturbation | StoSM_M_100/200/300 | Small→Small-Medium (100, 200, & 300 ms) |
StoM_M_100/200/300 | Small→Medium (100, 200, & 300 ms) | |
StoML_M_100/200/300 | Small→Medium-Large (100, 200, & 300 ms) | |
StoL_M_100/200/300 | Small→Large (100, 200, & 300 ms) | |
SMtoM_M_100/200/300 | Small-Medium→Medium (100, 200, & 300 ms) | |
SMtoML_M_100/200/300 | Small-Medium→Medium-Large (100, 200, & 300 ms) | |
SMtoL_M_100/200/300 | Small-Medium→Large (100, 200, & 300 ms) | |
MtoML_M_100/200/300 | Medium→Medium-Large (100, 200, & 300 ms) | |
MtoL_M_100/200/300 | Medium→Large (100, 200, & 300 ms) | |
MLtoL_M_100/200/300 | Medium-Large→Large (100, 200, & 300 ms) | |
Distance perturbation | ||
Condition name in data record* | Condition† | |
No perturbation | M_N | Medium:Near |
M_NM | Medium:Near-Middle | |
M_M | Medium:Middle | |
M_MF | Medium:Middle-Far | |
M_F | Medium:Far | |
Perturbation | M_NtoNM_100/200/300 | Near→Near-Middle (100, 200, & 300 ms) |
M_NtoM_100/200/300 | Near→Middle (100, 200, & 300 ms) | |
M_NtoMF_100/200/300 | Near→Middle-Far (100, 200, & 300 ms) | |
M_NtoF_100/200/300 | Near→Far (100, 200, & 300 ms) | |
N_NMtoM_100/200/300 | Near-Middle→Middle (100, 200, & 300 ms) | |
M_NMtoMF_100/200/300 | Near-Middle→Middle-Far (100, 200, & 300 ms) | |
M_NMtoF_100/200/300 | Near-Middle→Far (100, 200, & 300 ms) | |
M_MtoMF_100/200/300 | Middle→Middle-Far (100, 200, & 300 ms) | |
M_MtoF_100/200/300 | Middle→Far (100, 200, & 300 ms) | |
M_MFtoF_100/200/300 | Middle-Far→Far (100, 200, & 300 ms) |
*Format: Object size_Object distance_Perturbation timing; arrows indicate perturbation.
†Parenthesized values indicate perturbation timin.
Technical Validation
Kinematic data: Effect of perturbations on reach-grasp coordination
Figures 4 and 5 show plots of the mean transport distance and aperture for the control (no perturbation) and a selected set of size- and distance-perturbation conditions, respectively, for a representative participant. Figures 6–9 describe phase plots of mean transport and aperture kinematics for the control and a selected set of size- and distance- perturbation conditions (perturbations applied at 100 ms, 200 ms, and 300 ms after movement onset) for a representative participant. As we presented in our previous work46, phase plots allow to distinguish the three phases of the reach-to-grasp coordination (i) Initiation Phase, which includes the initial acceleration of transport velocity and the first half of the hand opening, which begins with the rapid opening of the thumb and index finger. (ii) Shaping Phase, which begins at maximum transport velocity. It includes the first half of transport deceleration and the second half of the hand opening, which ends when the maximum aperture is achieved, marking the initiation of closure or closure onset, CO. (iii) Closure Phase, which includes the second half of transport deceleration and lasts until the object is grasped. Finally, Fig. 10 shows normalized EMG for each of the 10 muscles in the control and a distance-perturbation condition (perturbation applied at 300 ms after movement onset) for a representative trial. These figures provide a glimpse into the qualitative effects of visual perturbations of object size and distance on reach-grasp coordination.
It has been firmly established that the aperture component of reach-to-grasp movement is influenced by the object’s physical dimensions, while the transport component remains relatively unaffected by changes in object size8,47. In contrast, the transport component of reach-to-grasp movement is influenced by the object’s spatial location (i.e., the distance from the observer) and precision requirements due to object size, while the aperture component remains relatively unaffected by changes in object distance7,48. Hence, visual perturbations of object size evoke online adjustments in grasp aperture, and visual perturbations of object distance evoke online adjustments in transport velocity. Accordingly, to ensure all applied visual perturbations of object size and distance influenced the aperture and transport components in known ways, it was examined whether the size and distance perturbations influenced the peak aperture and peak transport velocity. To this end, movement time, peak aperture, and peak transport velocity were compared between each of the thirty size and distance perturbation conditions and the respective unperturbed condition (e.g., peak aperture for S→SM, S→M, S→ML, S→L each was compared to peak aperture for the S condition). Consistent with the findings from past studies conducted in the real world49 and VE44,46,50, movement time linearly scaled to object size and distance in the control conditions. The smaller the object the longer the reach-to-grasp movement (L: 977 ms, ML: 1019 ms, M: 1037 ms, SM: 1053 ms S: 1134 ms; rm-ANOVA: F4,36 = 24, p < 0.001; Fig. 4), and the farther the object the longer the reach-to-grasp movement (N: 897 ms, NM: 942 ms, M: 984 ms, MF: 1039 ms, F: 1095 ms; F4,36 = 49.7, p < 0.001; Fig. 5). As typically observed of reach-to-grasp movements8,46,47, peak aperture linearly scaled to object size (S: 8.6 cm, SM: 9.2 cm, M: 9.7 cm, ML: 10.2 cm, L: 10.8 cm; F4,36 = 159, p < 0.001; Fig. 6) and showed expected increase with perturbation of object size (Figs. 6 and 7). Similarly, respecting known trends7,48,49, peak transport velocity linearly scaled to object distance (N: 49.7 cm/s, NM: 59.3 cm/s, M: 68.7 cm/s, MF: 77.2 cm/s, F: 85.3 cm/s; F4,36 = 304.4, p < 0.001; Fig. 8) and showed expected increase with perturbation of object distance (Figs. 8 and 9). Importantly, final aperture was always scaled to object size. These trends provide a strong validation of the expected responses to perturbations of object size and distance during reach-to-grasp.
EMG data: Spectral properties of EMG signals
Each recorded EMG signal was validated via analysis of its spectral properties and then compared with known results from the literature. For each muscle for each trial, the power spectral density was calculated using Welch’s method with a Hann window of 1024 samples (i.e., 1024 ms) and 50% overlap. An example is presented on Fig. 11 for the EMG signal obtained during MVC test both for inactive and active muscle. Power was normalized to the maximum power on the respective trial and averaged across all trials and subjects for each muscle. Figures 12 and 13 show the mean normalized power spectral densities for EMG collected in size and distance perturbation conditions, respectively. Signal energy was primarily contained within 0–400 Hz, which is typical for EMG51,52. Power line noise (60 Hz) or its harmonics was observed in a minority of muscles. This artifact was a narrow band and is amenable to standard filtering procedures. In a minority of muscles, a second artifact was observed ~74 Hz. We cannot explain this artifact, but it is also a narrow band, consistent, and amenable to filtering using a band stop filter.
Usage Notes
A major strength of the present dataset is that it provides reach-to-grasp kinematics and EMG data for a larger number of combinations of object size and distance and a large number of perturbations of object size and distance applied at three different times during the movement. Numerous examples of face validity, the degree to which our data appear to measure what was intended to be measured, are readily apparent in our data. For example, peak aperture increased with object size, peak velocity increased with object distance, and perturbed movements generally showed extended movement times compared to the analogous controls. However, the present dataset is also limited in several ways, mostly pertaining to our choice of the object type, grasp type, and kinematic recording. First, we used only one object type (a cuboid), whereas everyday reach-to-grasp movements involve diverse objects, often asymmetrical in shape. Second, the participants reached-to and grasped objects using the pincer grip, which involved only the thumb and index finger, which does not capture the full diversity of grasping movements associated with grasping the same object or objects of different size or shape53. Finally, we attached markers to the wrist, thumb, and index finger, which does not capture the hand’s joint angle movement. These factors might limit the range of potential uses the present dataset, but it should not preclude the modeling of reach-to-grasp movements.
Acknowledgements
This work was supported by NIH grants #R01NS085122 and #2R01HD058301, and NSF grants #CBET-1804550 and #CMMI-M3X-1935337, to Eugene Tunik. We thank Alex Hunton and Samuel Berin for developing the VR platform. We thank Sambina Anthony for help with data collection and preparation of Fig. 1.
Author contributions
M.P.F., M.Y., and G.T. conceived and designed research; M.P.F. and M.Y. performed experiments; M.P.F., M.M., and K.L. curated data; M.M. prepared figures; M.M. drafted manuscript; M.P.F., M.M., K.L., M.Y., and G.T. edited and revised manuscript; M.P.F., M.M., K.L., M.Y., and G.T. approved final version of manuscript.
Code availability
The code used for post-processing of the kinematic data is available at https://github.com/tuniklab/scientific-data.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
These authors contributed equally: Mariusz P. Furmanek, Madhur Mangalam, Mathew Yarossi.
References
- 1.Haggard P, Wing AM. Remote responses to perturbation in human prehension. Neurosci. Lett. 1991;122:103–108. doi: 10.1016/0304-3940(91)90204-7. [DOI] [PubMed] [Google Scholar]
- 2.Haggard P, Wing A. Coordinated responses following mechanical perturbation of the arm during prehension. Exp. Brain Res. 1995;102:483–494. doi: 10.1007/BF00230652. [DOI] [PubMed] [Google Scholar]
- 3.Schettino LF, Adamovich SV, Tunik E. Coordination of pincer grasp and transport after mechanical perturbation of the index finger. J. Neurophysiol. 2017;117:2292–2297. doi: 10.1152/jn.00642.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Desmurget M, Grafton S. Forward modeling allows feedback control for fast reaching movements. Trends Cogn. Sci. 2000;4:423–431. doi: 10.1016/S1364-6613(00)01537-0. [DOI] [PubMed] [Google Scholar]
- 5.Rice NJ, Tunik E, Cross ES, Grafton ST. On-line grasp control is mediated by the contralateral hemisphere. Brain Res. 2007;1175:76–84. doi: 10.1016/j.brainres.2007.08.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Tunik E, Frey SH, Grafton ST. Virtual lesions of the anterior intraparietal area disrupt goal-dependent on-line adjustments of grasp. Nat. Neurosci. 2005;8:505–511. doi: 10.1038/nn1430. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Paulignan Y, MacKenzie C, Marteniuk R, Jeannerod M. Selective perturbation of visual input during prehension movements. 1. The effects of changing object position. Exp. Brain Res. 1991;83:502–512. doi: 10.1007/BF00229827. [DOI] [PubMed] [Google Scholar]
- 8.Paulignan Y, Jeannerod M, MacKenzie C, Marteniuk R. Selective perturbation of visual input during prehension movements. 2. The effects of changing object size. Exp. Brain Res. 1991;87:407–420. doi: 10.1007/BF00231858. [DOI] [PubMed] [Google Scholar]
- 9.Volcic R, Domini F. On-line visual control of grasping movements. Exp. Brain Res. 2016;234:2165–2177. doi: 10.1007/s00221-016-4620-x. [DOI] [PubMed] [Google Scholar]
- 10.Ansuini C, Santello M, Tubaldi F, Massaccesi S, Castiello U. Control of hand shaping in response to object shape perturbation. Exp. Brain Res. 2007;180:85–96. doi: 10.1007/s00221-006-0840-9. [DOI] [PubMed] [Google Scholar]
- 11.Popovic DB, Popovic MB, Sinkjær T. Neurorehabilitation of upper extremities in humans with sensory-motor impairment. Neuromodulation Technol. Neural Interface. 2002;5:54–66. doi: 10.1046/j.1525-1403.2002._2009.x. [DOI] [PubMed] [Google Scholar]
- 12.van Vliet P, Pelton TA, Hollands KL, Carey L, Wing AM. Neuroscience findings on coordination of reaching to grasp an object: Implications for research. Neurorehabil. Neural Repair. 2013;27:622–635. doi: 10.1177/1545968313483578. [DOI] [PubMed] [Google Scholar]
- 13.Tretriluxana J, et al. Feasibility investigation of the accelerated skill acquisition program (ASAP): Insights into reach-to-grasp coordination of individuals with postacute stroke. Top. Stroke Rehabil. 2013;20:151–160. doi: 10.1310/tsr2002-151. [DOI] [PubMed] [Google Scholar]
- 14.Popović MB. Control of neural prostheses for grasping and reaching. Med. Eng. Phys. 2003;25:41–50. doi: 10.1016/S1350-4533(02)00187-X. [DOI] [PubMed] [Google Scholar]
- 15.Balasubramanian K, et al. Changes in cortical network connectivity with long-term brain-machine interface exposure after chronic amputation. Nat. Commun. 2017;8:1796. doi: 10.1038/s41467-017-01909-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Mastinu E, et al. Neural feedback strategies to improve grasping coordination in neuromusculoskeletal prostheses. Sci. Rep. 2020;10:11793. doi: 10.1038/s41598-020-67985-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Laschi C, et al. A bio-inspired predictive sensory-motor coordination scheme for robot reaching and preshaping. Auton. Robots. 2008;25:85–101. doi: 10.1007/s10514-007-9065-4. [DOI] [Google Scholar]
- 18.Levine S, Pastor P, Krizhevsky A, Ibarz J, Quillen D. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Rob. Res. 2017;37:421–436. doi: 10.1177/0278364917710318. [DOI] [Google Scholar]
- 19.Hoff B, Arbib MA. Models of trajectory formation and temporal interaction of reach and grasp. J. Mot. Behav. 1993;25:175–192. doi: 10.1080/00222895.1993.9942048. [DOI] [PubMed] [Google Scholar]
- 20.Ulloa A, Bullock D. A neural network simulating human reach–grasp coordination by continuous updating of vector positioning commands. Neural Networks. 2003;16:1141–1160. doi: 10.1016/S0893-6080(03)00079-0. [DOI] [PubMed] [Google Scholar]
- 21.Rand MK, Shimansky YP, Hossain ABMI, Stelmach GE. Quantitative model of transport-aperture coordination during reach-to-grasp movements. Exp. Brain Res. 2008;188:263–274. doi: 10.1007/s00221-008-1361-5. [DOI] [PubMed] [Google Scholar]
- 22.Bullock IM, Feix T, Dollar AM. The Yale human grasping dataset: Grasp, object, and task data in household and machine shop environments. Int. J. Rob. Res. 2014;34:251–255. doi: 10.1177/0278364914555720. [DOI] [Google Scholar]
- 23.Han M, Günay SY, Schirner G, Padır T, Erdoğmuş D. HANDS: A multimodal dataset for modeling toward human grasp intent inference in prosthetic hands. Intell. Serv. Robot. 2020;13:179–185. doi: 10.1007/s11370-019-00293-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Garcia-Hernando, G., Yuan, S., Baek, S. & Kim, T.-K. First-person hand action benchmark with rgb-d videos and 3d hand pose annotations. in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 409–419 (2018).
- 25.Li, Y. & Pollard, N. S. A shape matching algorithm for synthesizing humanlike enveloping grasps. in 5th IEEE-RAS International Conference on Humanoid Robots 442–449, 10.1109/ICHR.2005.1573607 (2005).
- 26.Brahmbhatt, S., Tang, C., Twigg, C. D., Kemp, C. C. & Hays, J. ContactPose: A dataset of grasps with object contact and hand pose. in European Conference on Computer Vision (eds. Vedaldi, A., Bischof, H., Brox, T. & Frahm, J.-M.) 361–378, 10.1007/978-3-030-58601-0_22 (Springer International Publishing, 2020).
- 27.Hu Y, Osu R, Okada M, Goodale MA, Kawato M. A model of the coupling between grip aperture and hand transport during human prehension. Exp. Brain Res. 2005;167:301–304. doi: 10.1007/s00221-005-0111-1. [DOI] [PubMed] [Google Scholar]
- 28.Rand MK, Shimansky YP, Hossain ABMI, Stelmach GE. Phase dependence of transport–aperture coordination variability reveals control strategy of reach-to-grasp movements. Exp. Brain Res. 2010;207:49–63. doi: 10.1007/s00221-010-2428-7. [DOI] [PubMed] [Google Scholar]
- 29.Rand MK, Shimansky YP. Two-phase strategy of neural control for planar reaching movements: I. XY coordination variability and its relation to end-point variability. Exp. Brain Res. 2013;225:55–73. doi: 10.1007/s00221-012-3348-5. [DOI] [PubMed] [Google Scholar]
- 30.Takemura N, Fukui T, Inui T. A computational model for aperture control in reach-to-grasp movement based on predictive variability. Front. Comput. Neurosci. 2015;9:143. doi: 10.3389/fncom.2015.00143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Verheij R, Brenner E, Smeets JBJ. Grasping kinematics from the perspective of the individual digits: A modelling study. PLoS One. 2012;7:e33150. doi: 10.1371/journal.pone.0033150. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Zhang S, Zhang Z, Zhou N. A new control model for the temporal coordination of arm transport and hand preshape applying to two-dimensional space. Neurocomputing. 2015;168:588–598. doi: 10.1016/j.neucom.2015.05.067. [DOI] [Google Scholar]
- 33.Jarque-Bou NJ, Vergara M, Sancho-Bru JL, Gracia-Ibáñez V, Roda-Sales A. A calibrated database of kinematics and EMG of the forearm and hand during activities of daily living. Sci. Data. 2019;6:270. doi: 10.1038/s41597-019-0285-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Matran-Fernandez A, Rodríguez Martínez IJ, Poli R, Cipriani C, Citi L. SEEDS, simultaneous recordings of high-density EMG and finger joint angles during multiple hand movements. Sci. Data. 2019;6:186. doi: 10.1038/s41597-019-0200-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Gabiccini, M., Stillfried, G., Marino, H. & Bianchi, M. A data-driven kinematic model of the human hand with soft-tissue artifact compensation mechanism for grasp synergy analysis. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems 3738–3745, 10.1109/IROS.2013.6696890 (2013).
- 36.Deimel R, Brock O. A novel type of compliant and underactuated robotic hand for dexterous grasping. Int. J. Rob. Res. 2015;35:161–185. doi: 10.1177/0278364915592961. [DOI] [Google Scholar]
- 37.Katsiaris, P. T., Artemiadis, P. K. & Kyriakopoulos, K. J. Relating postural synergies to low-D muscular activations: Towards bio-inspired control of robotic hands. in 2012 IEEE 12thInternational Conference on Bioinformatics & Bioengineering (BIBE) 245–250, 10.1109/BIBE.2012.6399682 (2012).
- 38.Roda-Sales A, Vergara M, Sancho-Bru JL, Gracia-Ibáñez V, Jarque-Bou NJ. Human hand kinematic data during feeding and cooking tasks. Sci. Data. 2019;6:167. doi: 10.1038/s41597-019-0175-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Jarque-Bou NJ, Atzori M, Müller H. A large calibrated database of hand movements and grasps kinematics. Sci. Data. 2020;7:12. doi: 10.1038/s41597-019-0349-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Bianchi M, Salaris P, Bicchi A. Synergy-based hand pose sensing: Reconstruction enhancement. Int. J. Rob. Res. 2013;32:396–406. doi: 10.1177/0278364912474078. [DOI] [Google Scholar]
- 41.Atzori M, et al. Electromyography data for non-invasive naturally-controlled robotic hand prostheses. Sci. Data. 2014;1:140053. doi: 10.1038/sdata.2014.53. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Luciw MD, Jarocka E, Edin BB. Multi-channel EEG recordings during 3,936 grasp and lift trials with varying weight and friction. Sci. Data. 2014;1:14004. doi: 10.1038/sdata.2014.47. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Furmanek MP, Mangalam M, Yarossi M, Lockwood K, Tunik E. 2021. Reach-to-grasp animation, figshare. [DOI]
- 44.Furmanek MP, et al. Effects of sensory feedback and collider size on reach-to-grasp coordination in haptic-free virtual reality. Front. Virtual Real. 2021;2:1–16. doi: 10.3389/frvir.2021.648529. [DOI] [Google Scholar]
- 45.Furmanek MP, Mangalam M, Yarossi M, Lockwood K, Tunik E. 2021. A kinematic and EMG dataset of online adjustment of reach-to-grasp movements to visual perturbations, figshare. [DOI] [PMC free article] [PubMed]
- 46.Furmanek MP, et al. Coordination of reach-to-grasp in physical and haptic-free virtual environments. J. Neuroeng. Rehabil. 2019;16:78. doi: 10.1186/s12984-019-0525-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Castiello U, Bennett KMB, Stelmach GE. Reach to grasp: The natural response to perturbation of object size. Exp. Brain Res. 1993;94:163–178. doi: 10.1007/bf00230479. [DOI] [PubMed] [Google Scholar]
- 48.Gentilucci M, Chieffi S, Scarpa M, Castiello U. Temporal coupling between transport and grasp components during prehension movements: Effects of visual perturbation. Behav. Brain Res. 1992;47:71–82. doi: 10.1016/s0166-4328(05)80253-0. [DOI] [PubMed] [Google Scholar]
- 49.Chieffi S, Gentilucci M. Coordination between the transport and the grasp components during prehension movements. Exp. Brain Res. 1993;94:471–477. doi: 10.1007/BF00230205. [DOI] [PubMed] [Google Scholar]
- 50.Mangalam M, Yarossi M, Furmanek MP, Tunik E. Control of aperture closure during reach-to-grasp movements in immersive haptic-free virtual reality. Exp. Brain Res. 2021;239:1651–1665. doi: 10.1007/s00221-021-06079-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Clancy EA, Bertolina MV, Merletti R, Farina D. Time- and frequency-domain monitoring of the myoelectric signal during a long-duration, cyclic, force-varying, fatiguing hand-grip task. J. Electromyogr. Kinesiol. 2008;18:789–797. doi: 10.1016/j.jelekin.2007.02.007. [DOI] [PubMed] [Google Scholar]
- 52.Kattla S, Lowery MM. Fatigue related changes in electromyographic coherence between synergistic hand muscles. Exp. Brain Res. 2010;202:89–99. doi: 10.1007/s00221-009-2110-0. [DOI] [PubMed] [Google Scholar]
- 53.Santello M, Flanders M, Soechting JF. Postural hand synergies for tool use. J. Neurosci. 1998;18:10105–10115. doi: 10.1523/JNEUROSCI.18-23-10105.1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Citations
- Furmanek MP, Mangalam M, Yarossi M, Lockwood K, Tunik E. 2021. Reach-to-grasp animation, figshare. [DOI]
- Furmanek MP, Mangalam M, Yarossi M, Lockwood K, Tunik E. 2021. A kinematic and EMG dataset of online adjustment of reach-to-grasp movements to visual perturbations, figshare. [DOI] [PMC free article] [PubMed]
Data Availability Statement
The code used for post-processing of the kinematic data is available at https://github.com/tuniklab/scientific-data.