Dileep George @dileeplearning's Avatar

Dileep George @dileeplearning

@dileeplearning.bsky.social

AGI research @DeepMind. Ex cofounder & CTO Vicarious AI (acqd by Alphabet), Cofounder Numenta Triply EE (BTech IIT-Mumbai, MS&PhD Stanford). #AGIComics blog.dileeplearning.com

11,384 Followers  |  690 Following  |  319 Posts  |  Joined: 04.07.2023  |  2.1567

Latest posts by dileeplearning.bsky.social on Bluesky

๐ŸŽฏ

05.06.2025 21:53 โ€” ๐Ÿ‘ 10    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Ohh ok I realize that @tyrellturing.bsky.social mentioned evolution. Fine then. But then which neuroscientist believes this?

15.05.2025 18:04 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Hmmโ€ฆI donโ€™t think itโ€™s impossible.

Evolution could create structures in the brain that are in correspondence with structure in the world.

15.05.2025 18:02 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Clone-structured graph representations enable flexible learning and vicarious evaluation of cognitive maps - Nature Communications Higher-order sequence learning using a structured graph representation - clone-structured cognitive graphs (CSCG) โ€“ can explain how the hippocampus learns cognitive maps. CSCG provides novel explanati...

Here's the CSCG paper: www.nature.com/articles/s41...

And here' the CML paper:
www.nature.com/articles/s41...

15.05.2025 01:24 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Good conclusion :-).

15.05.2025 01:22 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

somehow chatGPT understand my opinion about successor representations? 4/

15.05.2025 01:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

I didn't mention partial observability specifically, so it is impressive that this was picked up. Looks like we did something right in our CSCG paper in making this explicit? 3/

15.05.2025 01:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

It is quite impressive that chatGPT picked up these nuances, picks up a relevant quote from the paper and even emphasizes portions of the response. 2/

15.05.2025 01:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

This paper turned up on a feed, I was intrigued by it and started reading...

..but then I was quite baffled because our CSCG work seem to have tackled many of these problems in a more general setting and it's not even mentioned!

So I asked ChatGPT... ...I'm impressed by the answer1. 1/๐Ÿงต

15.05.2025 01:22 โ€” ๐Ÿ‘ 11    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Graph schemas as abstractions for transfer learning, inference, and planning Transferring latent structure from one environment or problem to another is a mechanism by which humans and animals generalize with very little data. Inspired by cognitive and neurobiological insights...

Some of our work could explain this kind of latent graph learning and schema-like abstraction. 2/
arxiv.org/abs/2302.07350

29.04.2025 00:06 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Wow, very cool to see this work from Alla Karpova's lab. She had shown me the results when I visited @hhmijanelia.bsky.social and I was blown away.

www.biorxiv.org/content/10.1...

1/

29.04.2025 00:05 โ€” ๐Ÿ‘ 33    ๐Ÿ” 5    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

it takes a bit of getting immersed in the field to know this :-)

27.04.2025 21:15 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This comic would be an inside joke. Many neuroscience papers that study a brain region postulate that the current does a simple transformation of the input to make it easy for brain regions that are 'downstream' to solve the real problem effectively deferring the real problem. ...

27.04.2025 21:14 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Brain dynamics and spatiotemporal trajectories during threat processing

๐—›๐—ผ๐˜„ ๐˜€๐—ต๐—ผ๐˜‚๐—น๐—ฑ ๐˜„๐—ฒ ๐—ฑ๐—ฒ๐—ณ๐—ถ๐—ป๐—ฒ ๐—ฎ๐—ป๐—ฑ ๐—ฑ๐—ฒ๐˜๐—ฒ๐—ฟ๐—บ๐—ถ๐—ป๐—ฒ ๐—ฎ ๐—ฏ๐—ฟ๐—ฎ๐—ถ๐—ป ๐—ฟ๐—ฒ๐—ด๐—ถ๐—ผ๐—ป'๐˜€ ๐—ถ๐—บ๐—ฝ๐—ผ๐—ฟ๐˜๐—ฎ๐—ป๐—ฐ๐—ฒ?
We introduce the idea of "importance" in terms of the extent to which a region's signals steer/contribute to brain dynamics as a function of brain state.
Work by @codejoydo.bsky.social
elifesciences.org/reviewed-pre...

27.04.2025 17:17 โ€” ๐Ÿ‘ 55    ๐Ÿ” 17    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

It's kinda obvious. #AGIComics has already figured out which brain region is the most important. ๐Ÿ˜‡

27.04.2025 20:56 โ€” ๐Ÿ‘ 25    ๐Ÿ” 5    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0
Post image

And whether top-down't influence is multiplicative or not is very context-dependent. (this is also what is seen in neurobiology). www.science.org/doi/10.1126/...

26.04.2025 00:59 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

www.science.org/doi/10.1126/...

26.04.2025 00:57 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Explaining away is a concrete role for feedback computations, and here is one example showing its effect. ....there are many more examples in the paper.
www.science.org/cms/10.1126/...

26.04.2025 00:57 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
A detailed theory of thalamic and cortical microcircuits for predictive visual inference A generative vision model offers a detailed theory for cortical columns, thalamus, and computations in cortical circuits.

just want to re-point to this work and remind you that we have proposed and built models that incorporate topdown feedback, and have a theory about it.
www.science.org/doi/10.1126/...

26.04.2025 00:54 โ€” ๐Ÿ‘ 6    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

ohh...yes...this is exactly what I think after reading some of the "deep research" reports. ....written by a committee

30.03.2025 01:30 โ€” ๐Ÿ‘ 9    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

jumping on the Gemini 2.5 bandwagon... it's an incredible model. really feels like an(other) inflection point. talking to Claude 3.7 feels like talking to a competent colleague who knows about everything, but makes mistakes. Gemini 2.5 feels like talking to a world-class expert with A+ intuitions

28.03.2025 17:16 โ€” ๐Ÿ‘ 58    ๐Ÿ” 4    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 4

already being scaled up...

27.03.2025 16:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

We could run it to analyze a transformer...

27.03.2025 16:13 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

How about running an experiment that runs this process for a known system? Assume the limits of sampling and see what kinds of insights we can get?

27.03.2025 16:12 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I just donโ€™t understand the proposalโ€ฆ.what models are going to be the โ€˜foundationโ€™ for these brain models? Is it the transformer architecture? โ€ฆbecause that is the one that is proven to be scalable so far. If so, how are brain insights going to be extracted from a trained transformer?

26.03.2025 23:10 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

well, benchmarks are useful, greedy hill climbing on them might lead you to new opportunities even if it doesnโ€™t lead you to new insights. And the job markets can remain irrational longer than you can remain solvent :-).

26.03.2025 22:54 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Give me 10 billion dollars and Iโ€™ll do it. 1 billion for developing the hardware and 9 billion to pay for my opportunity cost ๐Ÿ˜‡

26.03.2025 22:15 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Sure. Can you give me a 10billion dollars?

26.03.2025 22:12 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

Whatโ€™s your beef with โ€œprocessingโ€? IMO, the specific architectural modifications are about โ€œprocessingโ€. The attention circuit is โ€œprocessingโ€. No?

26.03.2025 22:11 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Again, tell me which architecture+algorithm you want to scale.

26.03.2025 22:04 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@dileeplearning is following 19 prominent accounts