Dexterity from Smart Lenses: Multi‑Fingered Robot Skills from Everyday Human Videos

Dexterity from Smart Lenses: Multi‑Fingered Robot Skills from Everyday Human Videos

What if robots could pick up dexterous, multi-finger skills just by watching you do everyday chores? That’s the promise of AINA, a new framework that learns robot hand policies from in‑the‑wild human demonstrations captured with Aria Gen 2 smart glasses.

Why it matters: collecting robot data is hard. AINA lets anyone, anywhere record useful training data while going about normal activities—no studio rigs.

  • Smart glasses provide high‑res video, precise 3D head and hand poses, and a wide stereo view for depth.
  • AINA turns this into 3D point‑based policies for multi‑fingered robot hands.
  • No robot data, online corrections, simulation, or reinforcement learning needed.
  • Policies are robust to changing backgrounds and deploy directly on real robots.

In tests across nine everyday manipulation tasks, AINA outperformed prior human‑to‑robot approaches and includes careful ablations of design choices.

See rollouts and details: https://aina-robot.github.io • Paper: https://arxiv.org/abs/2511.16661v1

Paper: https://arxiv.org/abs/2511.16661v1

Register: https://www.AiFeta.com

#Robotics #AI #RobotLearning #ImitationLearning #DexterousManipulation #ComputerVision #Wearables #SmartGlasses #EmbodiedAI #Research

Read more