in retrospect, we should have done more to temper expectations and highlight the difference between the method/architecture/open-source code and the actual pretrained weights, which we def want ppl to play with, but come with absolutely no guarantees.
28.11.2024 20:04 — 👍 9 🔁 0 💬 1 📌 0
(disclaimer: opinions are my own)
we used language that implied that the pretrained Octo weights would definitely actually be useful for YOU in YOUR lab. in reality, the real world is much bigger than OXE and some carefully-tuned evals. end-to-end robot learning is not even at the GPT-2 level
28.11.2024 20:04 — 👍 11 🔁 0 💬 2 📌 0
super happy to hear the first one, since we worked so hard on open-sourcing! the last one is the one that I'm less confident about -- it's very hard to beat a well-tuned from-scratch baseline. but if it works for you, then perhaps I've been too pessimistic!
27.11.2024 19:48 — 👍 1 🔁 0 💬 1 📌 0
don't worry I wasn't insulted :)
interestingly, this is the second time that public dunking on Octo has informed me of someone who actually does use it (like @tomdupuis.bsky.social), revising my opinion of its utility upwards rather than downwards.
(the first time I was the one doing the dunking)
27.11.2024 18:40 — 👍 8 🔁 0 💬 1 📌 1
yeah we 100% oversold Octo
27.11.2024 18:10 — 👍 15 🔁 0 💬 3 📌 0
tokenization is inductive bias
25.11.2024 18:10 — 👍 1 🔁 0 💬 0 📌 0
same, I've been trying to force myself to use btop but something about htop just feels... comfy... maybe it's just familiarity bias
22.11.2024 05:48 — 👍 1 🔁 0 💬 1 📌 0
Bringing the sergey posts until he does it himself.
Robotics. Reinforcement learning. AI.
I work at Sakana AI 🐟🐠🐡 → @sakanaai.bsky.social
https://sakana.ai/careers
PhD student @ UC Berkeley
forget this i didn’t know you can share block lists here. call me when you fixed this
hot takes, linear Algebra, JAX apologist, Raconteur
PhD student at UC Berkeley studying RL and AI safety.
https://cassidylaidlaw.com
Señor swesearcher @ Google DeepMind, adjunct prof at Université de Montréal and Mila. Musician. From 🇪🇨 living in 🇨🇦.
https://psc-g.github.io/
building the future
research at midjourney, deepmind. slinging ai hot takes 🥞at artfintel.com
Assistant Professor, Paul G. Allen School of Computer Science and Engineering, University of Washington
Visiting Faculty, NVIDIA
Ph.D. from Berkeley, Postdoc MIT
https://homes.cs.washington.edu/~abhgupta
I like robots and reinforcement learning :)
PhDing @UCSanDiego @NVIDIA @hillbot_ai on scalable robot learning and embodied AI. Co-founded @LuxAIChallenge to build AI competitions. @NSF GRFP fellow
http://stoneztao.com
Research @OpenAI. I study Reinforcement Learning. PhD from UT Austin. Previously FAIR Paris, Meta US, NVIDIA, CMU, and IIT Kharagpur.
Website: https://hari-sikchi.github.io/
Assistant Professor of CS @nyuniversity.
I like robots!
PhD Student at UC San Diego | LLM Agents, Reinforcement Learning, Human-AI Collaboration, Multi-Agent Systems
PhD student at @cmurobotics.bsky.social working on efficient algorithms for interactive learning (e.g. imitation / RL / RLHF). no model is an island. prefers email. https://gokul.dev/. on the job market!
Research Scientist at Google DeepMind, interested in multiagent reinforcement learning, game theory, games, and search/planning.
Lover of Linux 🐧, coffee ☕, and retro gaming. Big fan of open-source. #gohabsgo 🇨🇦
For more info: https://linktr.ee/sharky6000
A LLN - large language Nathan - (RL, RLHF, society, robotics), athlete, yogi, chef
Writes http://interconnects.ai
At Ai2 via HuggingFace, Berkeley, and normal places
Anti-cynic. Towards a weirder future. Reinforcement Learning, Autonomous Vehicles, transportation systems, the works. Asst. Prof at NYU
https://emerge-lab.github.io
https://www.admonymous.co/eugenevinitsky