Weβre really excited about the speakers we have lined up, including Animesh Garg (Georgia Tech) @animesh-garg.bsky.social , Daniel Ho (1X Technologies), Hao Su (UCSD), Katerina Fragkiadki (CMU), Yilun Du (Harvard) @yilundu.bsky.social, and Russ Tedrake (TRI + MIT).
01.08.2025 18:11 β π 5 π 1 π¬ 1 π 0
Learning to Simulate Robot Worlds
Join the Learning to Simulate Robot Worlds workshop.
π Weβre pleased to announce our workshop at CoRL 2025: Learning to Simulate Robot Worlds! Workshop website: simulatingrobotworlds.github.io
The workshop aims to cover topics like physics-grounded simulation, photorealistic digital twins, AI-controlled simulators, to learned neural world models.
01.08.2025 18:11 β π 7 π 3 π¬ 1 π 1
Adapt3R: Adaptive 3D Scene Representation for Domain Transfer in Imitation Learning
Adapt3R
This was led by Albert Wilcox with support from Mohamed Ghanem, Masoud Moghani, Pierre Barroso, Benjamin Joffe and myself.
.
Learn more:
π Website: pairlab.github.io/Adapt3R
π Paper: arxiv.org/abs/2503.04877
π₯οΈ Code: github.com/pairlab/Adap...
25.07.2025 16:56 β π 0 π 0 π¬ 0 π 0
On our real-world multitask IL benchmark, Adapt3R achieves a strong in-distribution success rate and sees the smallest performance loss by far under a dramatically new viewpoint.
25.07.2025 16:56 β π 0 π 0 π¬ 1 π 0
Here we rotate the scene camera by ΞΈ radians about the EE starting position, and see that Adapt3R consistently achieves the strongest performance. Notably, it maintains >80% success rate on LIBERO, and has the only nonzero MimicGen success rate when ΞΈ β₯ 0.6.
25.07.2025 16:56 β π 0 π 0 π¬ 1 π 0
With an unseen embodiment, Adapt3R consistently outperforms its 2D counterparts and DP3, and outperforms 3D Diffuser Actor in two of our four settings. 2D features lifted into 3D are an effective representation for this scenario and Adapt3R makes good use of them!
25.07.2025 16:55 β π 0 π 0 π¬ 1 π 0
When we evaluate policies trained with Adapt3R on the multitask LIBERO benchmark or the high-precision tasks from the MimicGen paper, we see that they are just as performant as their RGB counterparts.
25.07.2025 16:55 β π 0 π 0 π¬ 1 π 0
Adapt3R unprojects 2D features into a point cloud, transforms them into the end effectorβs coordinate frame, and uses attention pooling to condense them into a single conditioning vector for IL. Notice that Adapt3R attends to the same points before and after the camera change!
25.07.2025 16:55 β π 0 π 0 π¬ 1 π 0
Learning 3D representations is hard without 3D data
π‘ The key idea is to use a 2D foundation model to extract semantic features, and use 3D information to localize those features in a canonical 3D space without extracting any semantic information from the 3D data.
25.07.2025 16:55 β π 0 π 0 π¬ 1 π 0
Albert Wilcox has been working on the using canonical 3D reps instead.
Yet, naive 3D alternatives don't work since most of the data is not easily featurized.
Adapt3R is a 3D backbone that works with your favorite robot learning method and generalizes to unseen embodiments & camera viewpoints!
25.07.2025 16:54 β π 0 π 0 π¬ 1 π 0
Imitation learning frameworks are often with 2D inputs.
but 2D limits generalization even to camera poses.
This has been an ongoing challenge, especially for humanoids since the camera pose is not steady and need not match the training data.
We build Adapt3R to solve this problem!
Read on for more
25.07.2025 16:54 β π 0 π 0 π¬ 1 π 0
@unitreerobotics.bsky.social great job.
Looking forward to more specs on the robot.
25.07.2025 13:29 β π 0 π 0 π¬ 0 π 0
Unitree R1 is a new $6K humanoid with a onboard LLM/VLM!
The price & complexity is not a barrier for ML folks to enter robot learning.
Stable, low-cost developer platforms will accelerate humanoid development with new ideas coming from everywhere.
Unitree did it for quadrupeds & now bipeds!
25.07.2025 13:28 β π 1 π 0 π¬ 1 π 0
This summer, I've picked up my long-form writing again on my blog at praxiscurrents.substack.com.
I'm starting with a deep dive into the importance of data-driven methods in the robotics -- The age of empiricism in Physical AI
Feel free to visit and subscribe for regular updates.
25.06.2025 17:12 β π 2 π 0 π¬ 0 π 0
Huge thanks to the #ICLR2025 Organizing Committee (including many who couldn't make it to the conference) πππ
27.04.2025 03:22 β π 50 π 6 π¬ 1 π 0
Come chat with us at ICLR 2025 today 3-5pm in Hall 3 + Hall 2B
#182 EgoSim: Egocentric Exploration in VirtualWorlds with
Multi-modal Conditioning
egosim.github.io/EgoSim/
Wei Yu
#401 PWM: Policy Learning with Multi-task World Models
imgeorgiev.com/pwm/
Varun Giridhar ncklashansen.bsky.social
25.04.2025 01:25 β π 1 π 0 π¬ 0 π 0
To all the folks arguing studying AI safety is scaremongering
Were you tracking the markets yesterday.
Unintended use cases by powerful yet uninformed users can cause immense irreversible damage nearly immediately!
#AI_SafetyMatters!
05.04.2025 11:06 β π 3 π 0 π¬ 0 π 0
Humanoid Robotics Startup Agility is Raising $400 Million
Agility Robotics, a humanoid robot maker led by former Microsoft executive Peggy Johnson, is raising $400 million at a pre-investment valuation of $1.75 billion, according to a person who has seen the...
A lot of kids who grew up on robotics science fiction are working hard to make it real.
The next few years in robotics π π₯!
Agility raises 400m on 1.75B pre!
www.theinformation.com/articles/hum...
Congrats to Agilityrobotics team!
@cpaxton.bsky.social, Peggy Johnson, Pras V. , Jonathan H.
31.03.2025 17:27 β π 3 π 0 π¬ 0 π 0
Are we there yet? No!
Humanoids will need all the human help they can get
I believe working on this problem might well be one of the coolest things you'll ever do!
I am leading our AI efforts for Humanoids.
I will elaborate in a blogpost in the upcoming weeks.
for now...to the stars and beyond!π
14.03.2025 16:29 β π 4 π 0 π¬ 0 π 0
It's a multifaceted decadal challenge.
There is a palpable excitement with tremendous progress - we are moving faster than ever before.
Apptronik has the ambition & heft to take on one of most impactful problems of this decade!
Thrilled to join Jeff Cardenas & Nick Paine on this journey
14.03.2025 16:29 β π 0 π 0 π¬ 0 π 0
AnyPlace: Learning Generalized Object Placement for Robot Manipulation
Check out our website any-place.github.io & paper www.arxiv.org/abs/2502.04531 for more details!
π£ to all our authors & collaborators: @allanzhao.bsky.social, Miroslav Bogdanovic, Chengyuan Luo, Steven Tohme, Kourosh Darvish, @aspuru.bsky.social, Florian Shkurti, @animesh-garg.bsky.social [5/5]
03.03.2025 20:41 β π 2 π 1 π¬ 0 π 0
AnyPlace: Learning Generalized Object Placement for Robot Manipulation
π Check out more at
any-place.github.io
www.arxiv.org/abs/2502.04531
Yuchi Zhao Miroslav Bogdanovic Chengyuan Luo Steven Tohm Kourosh Darvish @aspuru.bsky.social Florian Skhurti & Animesh Garg
@uoftcompsci.bsky.social
@vectorinstitute.ai @accelerationc.bsky.social @gtresearchnews.bsky.social
24.02.2025 22:11 β π 2 π 1 π¬ 0 π 0
How well does AnyPlace perform?
π Simulation results: Outperforms baselines in
β Success rate
β Coverage of placement modes
β Fine-placement precision
π Real-world results: Our method transfers directly from synthetic to real-world tasks, succeeding where others struggle!
24.02.2025 22:11 β π 1 π 0 π¬ 1 π 0
To generalize across objects & placements, we generate a fully synthetic dataset with:
β
Randomly generated objects in Blender
β
Diverse placement configurations (stacking, insertion, hanging) in IsaacSim
This allows us to train our model without real-world data collection! π
24.02.2025 22:11 β π 1 π 0 π¬ 1 π 0
Our key insight is that by leveraging a Vision-Language Model (VLM) to identify rough placement locations, we focus only on the relevant regions for local placement, which enables us to train the low-level placement-pose-prediction model to capture diverse placements efficiently.
24.02.2025 22:11 β π 1 π 0 π¬ 2 π 0
How can robots reliably place objects in diverse real-world tasks?
π€π Placement is hard! - objects vary in shape & placement modes (such as stacking, hanging, insertion)
AnyPlace predicts placement poses of unseen objects in real-world with ony using synthetic training data!
Read onπ
24.02.2025 22:11 β π 10 π 4 π¬ 1 π 1
The Ultra-Scale Playbook - a Hugging Face Space by nanotron
The ultimate guide to training LLM on large GPU Clusters
sure you can read the cool new LLM paper.
but, you know what is needed to make it work behind the scenes?
Thousands of GPUs humming in perfect harmony.
Thx @thomwolf.bsky.social & @huggingface.bsky.social for writing such a wonderful treatise
huggingface.co/spaces/nanot...
19.02.2025 21:57 β π 14 π 0 π¬ 0 π 0
a cartoon of peter griffin standing in front of a sign that says the beatles
ALT: a cartoon of peter griffin standing in front of a sign that says the beatles
We must remind ourselves that those who hold views contrary to oneβs own are rarely evil or stupid, and may know or understand things that we do not.
news.stanford.edu/stories/2017...
Resist intellectual monoculture ππΌπͺ
16.02.2025 19:26 β π 2 π 0 π¬ 0 π 0
I completely agree, however the synthesis often both requires a science and leads to more interesting questions
I often find myself defending CS systems research because of the outsized contributions they have in moving the field forward
It should not be relegated to industry
16.02.2025 10:48 β π 1 π 0 π¬ 0 π 0
VP of Research, GenAI @ Meta (Multimodal LLMs, AI Agents), UPMC Professor of Computer Science at CMU, ex-Director of AI research at @Apple, co-founder Perceptual Machines (acquired by Apple)
https://thenumb.at
Computer Graphics, Programming, Math, OxCaml, C++
author of Blood in the Machine, tech writer, luddite
newsletter: https://www.bloodinthemachine.com/
books: https://www.hachettebookgroup.com/contributor/brian-merchant/
kofi link: https://ko-fi.com/brianmerchant
PhDing @UCSanDiego @NVIDIA @hillbot_ai on scalable robot learning and embodied AI. Co-founded @LuxAIChallenge to build AI competitions. @NSF GRFP fellow
http://stoneztao.com
Assistant Professor @Princeton. Developing robots that plan and learn to help people. Prev: @Cornell, @MIT, @Harvard.
https://tomsilver.github.io/
I build tools that propel communities forward
Professor @UCLA, Research Scientist @ByteDance | Recent work: SPIN, SPPO, DPLM 1/2, GPM, MARS | Opinions are my own
Marrying classical CV and Deep Learning. I do things, which work, rather than being novel, but not working.
http://dmytro.ai
AI & Transportation | MIT Associate Professor
Interests: AI for good, sociotechnical systems, machine learning, optimization, reinforcement learning, public policy, gov tech, open science.
Science is messy and beautiful.
http://www.wucathy.com
Graphics β© Vision, Electronic Elsewheres. AI-Mediated Reality & Interaction Research @ NVIDIA; previously: Omniverse XR and Robotics, much before that VR, VFX & TV.
Thinkpiece PiΓ±ata & Artist inbetween.
Opinions? No
πBrooklyn/TLV
https://omershapira.com
PhD candidate at UCSD. Prev: NVIDIA, Meta AI, UC Berkeley, DTU. I like robots π€, plants πͺ΄, and they/them pronouns π³οΈβπ
https://www.nicklashansen.com
Interpretable Deep Networks. http://baulab.info/ @davidbau
kashyap7x.github.io
Postdoc at NVIDIA. Previously at the University of TΓΌbingen and CMU. Robot Learning, Autonomous Driving.
Deep learning, computational chemistry, generative modeling, AI for Science. Principal Research Manager at Microsoft Research AI for Science.
Wisconsinite. Dad. Former Chair, @wisdems.org
Assistant Professor Stanford CS. Perception, learning and control for autonomous robotic manipulation. https://web.stanford.edu/~bohg/
Assistant Prof. at Georgia Tech | NVIDIA AI | Making robots smarter
Director, Princeton Language and Intelligence. Professor of CS.
Research fellow @OxfordStats @OxCSML, spent time at FAIR and MSR
Former quant π (@GoldmanSachs), former former gymnast π€ΈββοΈ
My opinions are my own
π§π¬-π¬π§ sh/ssh
Postdoc @ UC Berkeley. 3D Vision/Graphics/Robotics. Prev: CS PhD @ Stanford.
janehwu.github.io