Page 14: Research news on Embodied robotic manipulation

Embodied robotic manipulation investigates robotic and prosthetic limbs that physically interact with the environment using human-like, adaptive control. Work in this area integrates soft robotic structures, tendon-driven and biohybrid actuators, and exoskeletons with rich multimodal sensing, including vision, tactile, and proprioceptive feedback. Machine learning methods such as imitation learning, reinforcement learning, and meta-learning are used to acquire dexterous skills, enable shared and autonomous control, and support intuitive human–robot interaction through haptic interfaces, brain–computer interfaces, and teleoperation systems.

Robotics

Robot see, robot do: System learns after watching how-to videos

Cornell University researchers have developed a new robotic framework powered by artificial intelligence—called RHyME (Retrieval for Hybrid Imitation under Mismatched Execution)—that allows robots to learn tasks by watching ...

page 14 from 16