Ellis Brown @ NeurIPS 2024's Avatar

Ellis Brown @ NeurIPS 2024

@ellisbrown.bsky.social

CS PhD Student @ NYU w/ Profs Saining Xie & Rob Fergus Intern @ Ai2 | Prev: CMU, BlackRock, Vanderbilt https://ellisbrown.github.io

120 Followers  |  95 Following  |  5 Posts  |  Joined: 07.08.2023  |  1.6137

Latest posts by ellisbrown.bsky.social on Bluesky

Visual-spatial intelligence–we rely on it to perceive, interact, and navigate our everyday spaces. To what capacity do MLLMs possess it? Do they mirror how humans think and reason about space?

Presenting β€œThinking in Space: How Multimodal Models See, Remember, and Recall Spaces”! [1/n]

23.12.2024 22:45 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 7    πŸ“Œ 0
Preview
nyu-visionx (NYU VisionX) Org profile for NYU VisionX on Hugging Face, the AI community building the future.

πŸ€— huggingface.co/nyu-visionx

10.12.2024 20:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Cambrian-1: A Fully Open Vision-Centric Exploration of MLLMs Cambrian-1 is a family of multimodal LLMs with a vision-centric design. We also release CV-Bench, a new vision-centric benchmark, and Cambrian-10M, a multimodal instruction-tuning dataset.

🌎 cambrian-mllm.github.io

10.12.2024 20:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Cambrian-1: A Fully Open, Vision-Centric Exploration of Multimodal LLMs We introduce Cambrian-1, a family of multimodal LLMs (MLLMs) designed with a vision-centric approach. While stronger language models can enhance multimodal capabilities, the design choices for vision ...

πŸ“„ arxiv.org/abs/2406.16860

10.12.2024 20:05 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
x.com

TL;DR 🧡

x.com/_ellisbrown/...

10.12.2024 17:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Heading to #NeurIPS2024 to present Cambrian-1! Catch our oral presentation Friday @ 10am (Oral 5C) and our poster afterwards until 2pm (#3700 in East Hall A-C) πŸͺΌπŸŽ‰

10.12.2024 17:31 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

@ellisbrown is following 19 prominent accounts