MIT Engineers Unveil Ultrasound Wristband Designed to Revolutionize Robotic Control and Virtual Reality Dexterity
Engineers at MIT design an AI-powered ultrasound wristband that tracks hand movements internally to control robots and virtual reality with high precision.
By: AXL Media
Published: Mar 26, 2026, 9:21 AM EDT
Source: Information for this report was sourced from Massachusetts Institute of Technology

Bridging the Gap Between Human Intention and Machine Motion
The pursuit of replicating the intricate mechanics of the human hand, which relies on dozens of joints and over a hundred tendons, has long stymied the fields of robotics and digital interface design. According to Xuanhe Zhao, a professor at MIT, the newly designed ultrasound wristband addresses this by capturing the nuanced internal shifts of the wrist. By focusing on the internal physiological triggers of movement rather than external visual cues, the device allows for a more seamless transition between a user's physical gestures and the actions of a remote or virtual surrogate.
Overcoming the Limitations of Traditional Motion Capture
Current methods for tracking hand dexterity often fall short due to environmental or physical constraints, such as cameras being blocked by obstacles or sensor-laden gloves inhibiting natural movement. As noted by the research team, even advanced systems that measure electrical muscle signals frequently struggle with environmental noise and lack the sensitivity required to distinguish subtle finger adjustments. The ultrasound approach bypasses these hurdles by providing high-resolution imaging of the "strings" that pull the fingers, effectively seeing through the skin to monitor the mechanics of motion without interference.
Architecting a Digital Map of Internal Wrist Mechanics
The hardware consists of a specialized ultrasound sticker, comparable in size to a smartwatch, paired with compact electronics that monitor 22 degrees of freedom within the hand. To bridge the gap between raw ultrasound imagery and specific finger positions, the researchers utilized a data-driven approach involving multiple cameras and volunteer participants. By labeling specific regions of the ultrasound images that correlate to motions like thumb extension or index finger flexion, the team created a comprehensive map of how internal wrist changes manifest as external hand positions.
Categories
Topics
Related Coverage
- MIT Engineers Develop Ultrasound Wristband for High-Precision Wireless Control of Dexterous Robotic Hands
- MIT Researchers Unveil EnergAIzer Tool to Predict AI Data Center Power Consumption in Seconds
- Science Centre Singapore Launches Inaugural RoboFest 2026 Featuring Social Robot Pilot and Student Competitions
- US and China Locked in Trillion Dollar Race for Global AI Supremacy