Rana Hanocka

Rana Hanocka

@ranahanocka.bsky.social

Assistant Professor @uchicago @uchicagocs. PhD from @TelAvivUni. Interested in computer graphics, machine learning, & computer vision πŸ€–

267 Followers 28 Following 9 Posts Joined Dec 2024
4 months ago
Video thumbnail

Excited to share our #ICCV2025 work Reusing Computation in Text-to-Image Diffusion for Efficient Generation of Image Sets!

Our method generates large sets of images using significantly less compute than standard diffusion.

πŸ“Žhttps://ddecatur.github.io/hierarchical-diffusion/

1/

8 2 1 2
7 months ago
Post image

This work was led by Sining Lu and Guan Chen, in collaboration with Nam Anh Dinh, Itai Lang, Ari Holtzman, and me.

Check out our paper: arxiv.org/abs/2508.08228
We’re still actively developing LL3M, and we’d love to hear your thoughts! 7/

1 1 0 0
7 months ago
Post image

Another cool thing about LL3M: the Blender code it writes is actually readable. Clear structure, detailed comments, intuitive variable names. Easy to tweak a single parameter (e.g. key width) or even change the algorithmic logic (e.g. the keyboard pattern). 6/

0 0 1 0
7 months ago
Post image

LL3M generates 3D assets in 3 phases with specialized agents:
1️⃣ Initial Creation β†’ break prompt into subtasks, retrieve relevant code snippets (BlenderRAG)
2️⃣ Auto-refine β†’ critic spots issues, verification checks fixes
3️⃣ User-guided β†’ iterative edits via user feedback
5/

0 0 1 0
7 months ago
Post image

LL3M can create a wide range of shapes, without requiring specialized 3D datasets or fine-tuning. Every asset created is represented under the hood as editable Blender code. 4/

0 0 1 0
7 months ago
Post image

Even non-experts can jump right in and easily edit 3D shapes. Blender code created by LL3M generates a node graph that is packed with tunable parameters, enabling users to tweak colors, textures, patterns, lengths, heights, and more. 3/

0 0 1 0
7 months ago
Post image

What we ❀️ about LL3M: You're in the loop! If you want to make a tweak, LL3M can be your collaborative 3D design partner. And there's no need to regenerate the entire model each time - target a specific part, provide follow-up prompts, and the rest stays intact. 2/

0 0 1 0
7 months ago
Video thumbnail

We’ve been building something we’re π‘Ÿπ‘’π‘Žπ‘™π‘™π‘¦ excited about – LL3M: LLM-powered agents that turn text into editable 3D assets. LL3M models shapes as interpretable Blender code, making geometry, appearance, and style easy to modify. πŸ”— threedle.github.io/ll3m 1/

2 0 1 0
9 months ago
Video thumbnail

Our work β€œGeometry in Style” will be presented at #CVPR2025 on Sunday at 4pm in ExHall D, poster 219. Drop by and say hi!

Our technique is capable of performing expressive text-driven deformations that preserve the input shape identity.
1/

22 4 1 0
9 months ago
Post image

Big update from The Workshop on Computer Vision For Mixed Reality @ CVPR 2025:

πŸ“ Papers and schedule are now up!

Our Speakers:
Richard Newcombe (Meta)
Anjul Patney (NVIDIA)
Rana Hanocka (UChicago)
Laura Leal-Taixe β€ͺ(NVIDIA)
Margarita Grinvald (Meta)

πŸ“… June 11, 8AM πŸ“Room 109.

cv4mr.github.io

15 3 2 4
11 months ago

Congrats Brian and team on the best paper honorable mention at #3dv2025 πŸ₯³πŸŽ‰
Brian (hywkim-brian.github.io/site/) is going to start his PhD next year at Columbia with @silviasellan.bsky.social Make sure to watch out for their awesome works πŸ“ˆπŸ“ˆπŸ“ˆ

15 3 1 0
1 year ago
Tweet advertising that Rana Hanocka at UChicago CS is recruiting PhD students.

My colleague Rana Hanocka does exciting work on the frontier for 3D graphics and AI/ML, and she’s recruiting PhD students this cycle!

6 1 1 0