We just released RUKA, a $1300 humanoid hand that is 3D-printable, strong, precise, and fully open sourced!
The key technical breakthrough here is that we can control joints and fingertips of the robot **without joint encoders**. All we need here is self-supervised data collection and learning.
18.04.2025 18:53 β π 29 π 7 π¬ 1 π 0
When life gives you lemons, you pick them up.
(trained with robotutilitymodels.com)
28.03.2025 04:02 β π 15 π 4 π¬ 1 π 0
The robot behaviors shown below are trained without any teleop, sim2real, genai, or motion planning. Simply show the robot a few examples of doing the task yourself, and our new method, called Point Policy, spits out a robot-compatible policy!
28.02.2025 19:09 β π 20 π 5 π¬ 1 π 1
We just released AnySense, an iPhone app for effortless data acquisition and streaming for robotics. We leverage Appleβs development frameworks to record and stream:
1. RGBD + Pose data
2. Audio from the mic or custom contact microphones
3. Seamless Bluetooth integration for external sensors
26.02.2025 15:14 β π 36 π 10 π¬ 2 π 0
Stretch Community News - February 2025 β Hello Robot
Gazebo Harmonic, Stretch AI, Dynamic memory, and more!
What's new in the Stretch community this month?
βοΈ Gazebo Harmonic
βοΈ Dynamic semantic maps for open-vocabulary tasks
βοΈ Natural-language narration of robot experiences
βοΈ Implicit human-robot communication
And more! Follow the link below for more details:
hello-robot.com/community-up...
06.02.2025 00:15 β π 4 π 1 π¬ 0 π 0
We all want a home robot that can actually help us out. Why can't I ask my robot "where did I leave my water bottle" and get a good answer?
In Graph-EQA, we build a 3d memory as the robot explores, using that memory to make decisions.
saumyasaxena.github.io/grapheqa/
30.12.2024 16:20 β π 44 π 8 π¬ 3 π 1
A look at the future: chatting with my robot via Discord to ask it to find something in my house.
This uses an LLM to understand what the human wants and generate a task plan, then builds an open-vocab 3d scene representation to find and pick up objects
31.12.2024 16:38 β π 20 π 4 π¬ 2 π 1
I'd like to introduce what I've been working at @hellorobot.bsky.social: Stretch AI, a set of open-source tools for language-guided autonomy, exploration, navigation, and learning from demonstration.
Check it out: github.com/hello-robot/...
Thread ->
03.12.2024 16:51 β π 132 π 23 π¬ 6 π 4
New paper! We show that by using keypoint-based image representation, robot policies become robust to different object types and background changes.
We call this method Prescriptive Point Priors for robot Policies or P3-PO in short. Full project is here: point-priors.github.io
10.12.2024 20:32 β π 37 π 7 π¬ 1 π 2
Modern policy architectures are unnecessarily complex. In our #NeurIPS2024 project called BAKU, we focus on what really matters for good policy learning.
BAKU is modular, language-conditioned, compatible with multiple sensor streams & action multi-modality, and importantly fully open-source!
09.12.2024 23:33 β π 30 π 9 π¬ 1 π 2
Since we are nearing the end of the year, I'll revisit some of our work I'm most excited about from the last year and maybe a sneak peek of what we are up to next.
To start of, Robot Utility Models, which enables zero-shot deployment. In the video below, the robot hasnt seen these doors before.
08.12.2024 02:32 β π 36 π 8 π¬ 2 π 3
Research Scientist at Toyota Research Institute | MIT PhD
CIS PhD at Penn | MIT CS + Math '24
sagnikanupam.com
PhD student working on AI reasoning in large multimodal models. I design methods to build better models for math, code, visual reasoning, agents, and robotics.
Stupid #robotics guy at ETHz
Twitter: https://x.com/ChongZitaZhang
Research Website: https://zita-ch.github.io/
ai/robotics | eng #15 @Meta | 2x founder/exit | contrib SO-ARM | advisor @Cline
linkedin.com/in/phil-fung
Undergraduate researcher. Interested in generalization, multi-objective reinforcement learning, and open-endedness | Looking for PhD in RL in 2026
My works: https://scholar.google.com/citations?user=GnHpLE8AAAAJ&hl=en
World Labs. Former research scientist at Google. Ph.D UWCSE.
π San Francisco π keunhong.com
Foundation Models for Generalizable Autonomy.
Assistant Professor in AI Robotics, Georgia Tech
prev Berkeley, Stanford, Toronto, Nvidia
Assistant Professor at Duke University. Robotics and AI.
http://boyuanchen.com/
AI for Robotics at @HuggingFace π€
Focusing on @LeRobotHF
π MSc Computer Science @TU_Muenchen π¬ Student Researcher @tumcvg π€ Make robot intelligent #AI #Rob #ComputerVision
Do cool things @ https://randing2000.github.io/
Professor of Computer Vision and AI at TU Munich, Director of the Munich Center for Machine Learning mcml.ai and of ELLIS Munich ellismunich.ai
cvg.cit.tum.de
PhD student @ DTU - Technical University of Denmark
Robotic Perception | 3D Vision | ML/AI
PhD student, cooperative perception in multi robot systems @Nottingham Trent University
#MachineLearning #Robotics #ROS2 #Gazebo π¨βπ»
Postdoc @csail.mit.edu, Ph.D. from @scai-asu.bsky.social
Working on AI Safety, AI Assessment, Automated Planning, Interpretability, Robotics
Previously: Masters from IITGuwahati, Research Intern at MetaAI
https://pulkitverma.net
Chair, Department of Computational Biomedicine at Cedars-Sinai Medical Center in Los Angeles. Director, Center for Artificial Intelligence Research & Education. Atari enthusiast. Retrocomputing. Maker.
PhD Candidate at Cambridge | ex Meta, Amazon | Studying diversity in multi-agent and multi-robot learning
https://matteobettini.com/
Research Scientist GoogleDeepMind
Ex @UniofOxford, AIatMeta, GoogleAI