Animesh Garg's Avatar

Animesh Garg

@animesh-garg.bsky.social

Foundation Models for Generalizable Autonomy. Assistant Professor in AI Robotics, Georgia Tech prev Berkeley, Stanford, Toronto, Nvidia

292 Followers  |  300 Following  |  40 Posts  |  Joined: 02.12.2024  |  2.4417

Latest posts by animesh-garg.bsky.social on Bluesky

Post image

We’re really excited about the speakers we have lined up, including Animesh Garg (Georgia Tech) @animesh-garg.bsky.social , Daniel Ho (1X Technologies), Hao Su (UCSD), Katerina Fragkiadki (CMU), Yilun Du (Harvard) @yilundu.bsky.social, and Russ Tedrake (TRI + MIT).

01.08.2025 18:11 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Preview
Learning to Simulate Robot Worlds Join the Learning to Simulate Robot Worlds workshop.

πŸš€ We’re pleased to announce our workshop at CoRL 2025: Learning to Simulate Robot Worlds! Workshop website: simulatingrobotworlds.github.io
The workshop aims to cover topics like physics-grounded simulation, photorealistic digital twins, AI-controlled simulators, to learned neural world models.

01.08.2025 18:11 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1
Adapt3R: Adaptive 3D Scene Representation for Domain Transfer in Imitation Learning Adapt3R

This was led by Albert Wilcox with support from Mohamed Ghanem, Masoud Moghani, Pierre Barroso, Benjamin Joffe and myself.
.
Learn more:
🌐 Website: pairlab.github.io/Adapt3R
πŸ“„ Paper: arxiv.org/abs/2503.04877
πŸ–₯️ Code: github.com/pairlab/Adap...

25.07.2025 16:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

On our real-world multitask IL benchmark, Adapt3R achieves a strong in-distribution success rate and sees the smallest performance loss by far under a dramatically new viewpoint.

25.07.2025 16:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Here we rotate the scene camera by ΞΈ radians about the EE starting position, and see that Adapt3R consistently achieves the strongest performance. Notably, it maintains >80% success rate on LIBERO, and has the only nonzero MimicGen success rate when ΞΈ β‰₯ 0.6.

25.07.2025 16:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

With an unseen embodiment, Adapt3R consistently outperforms its 2D counterparts and DP3, and outperforms 3D Diffuser Actor in two of our four settings. 2D features lifted into 3D are an effective representation for this scenario and Adapt3R makes good use of them!

25.07.2025 16:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

When we evaluate policies trained with Adapt3R on the multitask LIBERO benchmark or the high-precision tasks from the MimicGen paper, we see that they are just as performant as their RGB counterparts.

25.07.2025 16:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Adapt3R unprojects 2D features into a point cloud, transforms them into the end effector’s coordinate frame, and uses attention pooling to condense them into a single conditioning vector for IL. Notice that Adapt3R attends to the same points before and after the camera change!

25.07.2025 16:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Learning 3D representations is hard without 3D data
πŸ’‘ The key idea is to use a 2D foundation model to extract semantic features, and use 3D information to localize those features in a canonical 3D space without extracting any semantic information from the 3D data.

25.07.2025 16:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Albert Wilcox has been working on the using canonical 3D reps instead.
Yet, naive 3D alternatives don't work since most of the data is not easily featurized.

Adapt3R is a 3D backbone that works with your favorite robot learning method and generalizes to unseen embodiments & camera viewpoints!

25.07.2025 16:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Imitation learning frameworks are often with 2D inputs.
but 2D limits generalization even to camera poses.

This has been an ongoing challenge, especially for humanoids since the camera pose is not steady and need not match the training data.

We build Adapt3R to solve this problem!
Read on for more

25.07.2025 16:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@unitreerobotics.bsky.social great job.
Looking forward to more specs on the robot.

25.07.2025 13:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Unitree R1 is a new $6K humanoid with a onboard LLM/VLM!

The price & complexity is not a barrier for ML folks to enter robot learning.

Stable, low-cost developer platforms will accelerate humanoid development with new ideas coming from everywhere.

Unitree did it for quadrupeds & now bipeds!

25.07.2025 13:28 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

This summer, I've picked up my long-form writing again on my blog at praxiscurrents.substack.com.

I'm starting with a deep dive into the importance of data-driven methods in the robotics -- The age of empiricism in Physical AI

Feel free to visit and subscribe for regular updates.

25.06.2025 17:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

Huge thanks to the #ICLR2025 Organizing Committee (including many who couldn't make it to the conference) πŸ‘πŸ‘πŸ‘

27.04.2025 03:22 β€” πŸ‘ 50    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Come chat with us at ICLR 2025 today 3-5pm in Hall 3 + Hall 2B

#182 EgoSim: Egocentric Exploration in VirtualWorlds with
Multi-modal Conditioning
egosim.github.io/EgoSim/
Wei Yu

#401 PWM: Policy Learning with Multi-task World Models
imgeorgiev.com/pwm/
Varun Giridhar ncklashansen.bsky.social

25.04.2025 01:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

To all the folks arguing studying AI safety is scaremongering

Were you tracking the markets yesterday.

Unintended use cases by powerful yet uninformed users can cause immense irreversible damage nearly immediately!

#AI_SafetyMatters!

05.04.2025 11:06 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Humanoid Robotics Startup Agility is Raising $400 Million Agility Robotics, a humanoid robot maker led by former Microsoft executive Peggy Johnson, is raising $400 million at a pre-investment valuation of $1.75 billion, according to a person who has seen the...

A lot of kids who grew up on robotics science fiction are working hard to make it real.
The next few years in robotics πŸš€ πŸ”₯!

Agility raises 400m on 1.75B pre!

www.theinformation.com/articles/hum...

Congrats to Agilityrobotics team!
@cpaxton.bsky.social, Peggy Johnson, Pras V. , Jonathan H.

31.03.2025 17:27 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Are we there yet? No!
Humanoids will need all the human help they can get

I believe working on this problem might well be one of the coolest things you'll ever do!

I am leading our AI efforts for Humanoids.
I will elaborate in a blogpost in the upcoming weeks.

for now...to the stars and beyond!πŸš€

14.03.2025 16:29 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

It's a multifaceted decadal challenge.
There is a palpable excitement with tremendous progress - we are moving faster than ever before.

Apptronik has the ambition & heft to take on one of most impactful problems of this decade!
Thrilled to join Jeff Cardenas & Nick Paine on this journey

14.03.2025 16:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Andrew Barto and Richard Sutton are the recipients of the 2024 ACM A.M. Turing Award for developing the conceptual and algorithmic foundations of reinforcement learning. Andrew Barto and Richard Sutton are the recipients of the 2024 ACM A.M. Turing Award for developing the conceptual and algorithmic foundations of reinforcement learning. In a series of papers beginnin...

This is a moment to celebrate for the long standing foundational contributions to RL by Sutton and Barto

Congrats to them and all their collaborators over the years who made this long line of advancements possible

www.acm.org/media-center...

05.03.2025 13:10 β€” πŸ‘ 17    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
AnyPlace: Learning Generalized Object Placement for Robot Manipulation

Check out our website any-place.github.io & paper www.arxiv.org/abs/2502.04531 for more details!

πŸ“£ to all our authors & collaborators: @allanzhao.bsky.social, Miroslav Bogdanovic, Chengyuan Luo, Steven Tohme, Kourosh Darvish, @aspuru.bsky.social, Florian Shkurti, @animesh-garg.bsky.social [5/5]

03.03.2025 20:41 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
AnyPlace: Learning Generalized Object Placement for Robot Manipulation

πŸ” Check out more at
any-place.github.io
www.arxiv.org/abs/2502.04531

Yuchi Zhao Miroslav Bogdanovic Chengyuan Luo Steven Tohm Kourosh Darvish @aspuru.bsky.social Florian Skhurti & Animesh Garg

@uoftcompsci.bsky.social
@vectorinstitute.ai @accelerationc.bsky.social @gtresearchnews.bsky.social

24.02.2025 22:11 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

How well does AnyPlace perform?
πŸ† Simulation results: Outperforms baselines in
βœ” Success rate
βœ” Coverage of placement modes
βœ” Fine-placement precision
πŸ“Œ Real-world results: Our method transfers directly from synthetic to real-world tasks, succeeding where others struggle!

24.02.2025 22:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

To generalize across objects & placements, we generate a fully synthetic dataset with:
βœ… Randomly generated objects in Blender
βœ… Diverse placement configurations (stacking, insertion, hanging) in IsaacSim
This allows us to train our model without real-world data collection! πŸš€

24.02.2025 22:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Our key insight is that by leveraging a Vision-Language Model (VLM) to identify rough placement locations, we focus only on the relevant regions for local placement, which enables us to train the low-level placement-pose-prediction model to capture diverse placements efficiently.

24.02.2025 22:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Video thumbnail

How can robots reliably place objects in diverse real-world tasks?

πŸ€–πŸ” Placement is hard! - objects vary in shape & placement modes (such as stacking, hanging, insertion)

AnyPlace predicts placement poses of unseen objects in real-world with ony using synthetic training data!

Read onπŸ‘‡

24.02.2025 22:11 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1
Preview
The Ultra-Scale Playbook - a Hugging Face Space by nanotron The ultimate guide to training LLM on large GPU Clusters

sure you can read the cool new LLM paper.

but, you know what is needed to make it work behind the scenes?
Thousands of GPUs humming in perfect harmony.

Thx @thomwolf.bsky.social & @huggingface.bsky.social for writing such a wonderful treatise

huggingface.co/spaces/nanot...

19.02.2025 21:57 β€” πŸ‘ 14    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
a cartoon of peter griffin standing in front of a sign that says the beatles ALT: a cartoon of peter griffin standing in front of a sign that says the beatles

We must remind ourselves that those who hold views contrary to one’s own are rarely evil or stupid, and may know or understand things that we do not.

news.stanford.edu/stories/2017...

Resist intellectual monoculture πŸ™πŸΌπŸ’ͺ

16.02.2025 19:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I completely agree, however the synthesis often both requires a science and leads to more interesting questions

I often find myself defending CS systems research because of the outsized contributions they have in moving the field forward

It should not be relegated to industry

16.02.2025 10:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@animesh-garg is following 20 prominent accounts