induction vs transduction point holds
with induction you can search because you have a metric to optimise (% train examples correct)
with transduction there is no clear metric to guide search / brute force, so the model needs to get it right? or to come up with a way to guide its own search
22.12.2024 00:19 β π 0 π 0 π¬ 0 π 0
just checked, on semi-private set ryan got 43 (not that far i admit)
22.12.2024 00:15 β π 1 π 0 π¬ 1 π 0
ok he did use for loops so he didn't hill climb, but you can 'filter good candidates' by taking the solutions that solves 100% training examples, and submit only these as solutions
with transduction you can't filter
the challenge rules say you can submit only 2 (3?) solutions per pb
22.12.2024 00:09 β π 1 π 0 π¬ 1 π 0
2) he used program synthesis which allows hill climbing on the % of training examples correct
if the o3 prompt that circulates is correct, the o3 score uses transduction (predicting output grid directly), and you can't hill climb there
you can ensemble, but that doesn't help much for hard pbs
22.12.2024 00:03 β π 1 π 0 π¬ 1 π 0
this is a misleading comparison for two reasons
1) that guy got 50% on the public test set, which is easier than the private test set where o3 reached the 85%(87?)
22.12.2024 00:03 β π 1 π 0 π¬ 2 π 0
the official testing procedure is 2 or 3 solutions per problem iirc, don't think chollet would have let them brute force it
it seems they don't use program induction, so they can't hill climb on training examples either
21.12.2024 15:57 β π 2 π 0 π¬ 0 π 0
hmm is there no bookmarks over here? or did i miss them?
16.12.2024 17:06 β π 0 π 0 π¬ 1 π 0
imol-workshop.github.io
09.12.2024 19:48 β π 1 π 0 π¬ 0 π 0
hope to see you all at the IMOL worskhop on sunday!
09.12.2024 19:47 β π 1 π 0 π¬ 1 π 0
my work: scholar.google.com/citations?user=VBz8gZ4AAAAJ
09.12.2024 19:46 β π 1 π 0 π¬ 1 π 0
in vancouver for @neuripsconf.bsky.social
looking forward catching up with friends and meeting new ones!
reach out to chat about:
> open-ended learning
> intrinsic motivations
> exploration and diversity search
> social and cultural learning
> llm agents
> other?
09.12.2024 19:46 β π 1 π 0 π¬ 1 π 0
hi Melanie,
we have a cool workshop on intrinsically motivated open-ended learning with a blend of cogsci and ai on dec 15
@IMOLNeurIPS2024 on X
see program here: imol-workshop.github.io/pages/program/
29.11.2024 19:27 β π 2 π 0 π¬ 1 π 0
oh cool, what's the paper? i've been thinking it could be the case and was wondering who wrote about it
28.11.2024 01:04 β π 1 π 0 π¬ 1 π 0
Jobs - Flowers Laboratory
FLOWing Epigenetic Robots and Systems
find more info at flowers.inria.fr/jobs/ (other positions are open)
27.11.2024 17:43 β π 1 π 0 π¬ 0 π 0
balancing exploration and exploitation with autotelic rl
autotelic rl is usually concerned with open-ended exploration in the absence of external reward
how should we conduct an open-ended exploration *at the service* of an external task?
deep rl skills required
27.11.2024 17:43 β π 2 π 0 π¬ 1 π 0
llm-mediated cultural evolution
we wanna study how llm-based agents can be used to facilitate collective intelligence in controlled human experiment where groups of participant collectively find solutions to problems
this requires some background in cogsci + llms
27.11.2024 17:43 β π 3 π 0 π¬ 1 π 0
we are recruiting interns for a few projects with @pyoudeyer
in bordeaux
> studying llm-mediated cultural evolution with @nisioti_eleni
@Jeremy__Perez
> balancing exploration and exploitation with autotelic rl with @ClementRomac
details and links in π§΅
please share!
27.11.2024 17:43 β π 6 π 6 π¬ 1 π 0
AI/ML enthusiast. πσ
σ
σ
σ
σ ΅σ ½σ ͺσ σ
σ
σ
σ
σ
’σ
σ σ
σ
’σ
σ
¦σ
σ
σ
₯σ
£σ σ
σ
σ
’σ
σ
σ
€σ
σ
σ
σ
£σ σ σ σ
σ
σ
¦σ
σ σ
σ
σ
σ
σ
’σ
’σ
σ
£σ
£σ
σ
σ
σ σ
σ
σ
σ
σ
σ
σ
σ σ
’σ
σ
σ
σ
σ
A LLN - large language Nathan - (RL, RLHF, society, robotics), athlete, yogi, chef
Writes http://interconnects.ai
At Ai2 via HuggingFace, Berkeley, and normal places
DM me a post to bookmark it. Temporary solution by @matteason.me until we get real bookmarks. See pinned thread for more details.
Information and updates about RLC 2025 at the University of Alberta from Aug. 5th to 8th!
https://rl-conference.cc
computational cognitive science @ nyu. director NYU minds, brains, and machines initiative. https://gureckislab.org. Are you interested in research in my lab? https://intake.gureckislab.org/interest/
PhD student at INRIA in the Flowers team. https://grgkovac.github.io
Twitter: @KovacGrgur
PhD student working on the cognition of LLMs | HRL team - ENS Ulm | FLOWERS - Inria Bordeaux
PhD student @univgroningen, intern @InriaScool, @FlowersINRIA
. CogScI, AI. Inspired by brains (make RNNs behave like autograd) and behavior (open-ended goals)
PhD student at INRIA Flowers team. MVA master
reytuag.github.io/gautier-hamon/
Striving to make computers smarter.
PhD student @FlowersINRIA
Scientist @ DeepMind and Honorary Fellow @ U of Edinburgh.
RL, agency, philosophy, foundations, AI.
https://david-abel.github.io
Professor at Imperial College London and Principal Scientist at Google DeepMind. Posting in a personal capacity. To send me a message please use email.
Research Scientist at Google DeepMind, interested in multiagent reinforcement learning, game theory, games, and search/planning.
Lover of Linux π§, coffee β, and retro gaming. Big fan of open-source. #gohabsgo π¨π¦
For more info: https://linktr.ee/sharky6000
Large scale engineering @ Google Deepmind
research scientist @autodesk -- quality-diversity, generative design w/generative AI
SeΓ±or swesearcher @ Google DeepMind, adjunct prof at UniversitΓ© de MontrΓ©al and Mila. Musician. From πͺπ¨ living in π¨π¦.
https://psc-g.github.io/
research scientist @deepmind. language & multi-agent rl & interpretability. phd @BrownUniversity '22 under ellie pavlick (she/her)
https://roma-patel.github.io
AI policy researcher, wife guy in training, fan of cute animals and sci-fi. Started a Substack recently: https://milesbrundage.substack.com/
Professor of Computer Science at Oxford. Senior Staff Research Scientist at Waymo.
Primarily Robotics and AI. Distinguishing hype-notism from plausibility one press release at a time. rodneybrooks.com/blog people.csail.mit.edu/brooks