Curt Welch's Avatar

Curt Welch

@curtwelch.bsky.social

Robotics, AI, RL, Applied Impact Robotics, Nova Labs Metal Shop Steward, Software Engineer, Blacksmith, Maker, KZ4GD, old fart

68 Followers  |  187 Following  |  27 Posts  |  Joined: 16.11.2024  |  2.3287

Latest posts by curtwelch.bsky.social on Bluesky

I’m surprised it’s only around 20x. Tesla has far more employees than I would have guessed.

29.01.2025 14:42 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

I’ve got one stuck on my lap as well!

02.12.2024 02:15 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

As obvious as that idea is to most of us, the truth is that the voters don’t agree with it. In a democracy, ignorant voters get their turn to learn from there mistakes so they can become less ignorant. It’s their turn now.

23.11.2024 09:54 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The media didn’t do this. The people did it. They stopped paying for real news and decided all they wanted was sensationalism, rage and fires. All the real journalists lost their jobs; media was bought out by capitalists; the free press is now only a capitalistic click bait factory.

22.11.2024 13:29 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

My prime interest is solving the AGI puzzle and since human intelligence exists in a brain that is a real time reactive learning system I see no other options to duplicating it other than signal processing parallel networks. Aka neural networks.

21.11.2024 01:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

So do you understand how LeCun’s belief that the brain uses supervised learning works? I don’t but I have not followed his work that closely. Or do you have ideas of how that would work in the brain?

18.11.2024 14:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Interesting that i misread β€œruin” as β€œrun” and thought it was an odd idea. Then caught my error.

But in hindsight I see deeper truth in my error. People who aggressively run their own lives are the most blind to the the needs of others around them. And blame everyone else for their problems.

18.11.2024 14:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Of course that’s how it work. And anything bad that happens will because of some dem. Or brown skinned dude.

17.11.2024 18:04 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

But what I have not had had the time to do is feed it more complex real word data and see if the features it creates are actually useful and valid feature for driving behavior. Or did I just create a useless random function map? I’d love to work on this full time but life hasn’t allowed it.

17.11.2024 17:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I created this back in 2017, worked more on it during COVID but haven’t had time in my life to do more. The algorithm is fast an efficient for extracting features. And creating multilayers of nodes works great for creating higher level abstractions. It all seems good.

17.11.2024 17:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

It is using a Bayesian-like inferences logic to define the meaning of each node so the inputs to each node are treated as evidence as to the truth of the feature being part of the current environment. The algorithm uses temporal correlation as its main learning tool. It’s totally generic.

17.11.2024 17:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The network of these memory nodes are a self organizing map that balances activity across the net but in so doing it’s defining roughly equal probability features. Which means it’s information maximizing. It self organizes to represent as much low frequency state as it can.

17.11.2024 17:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

So what I’ve done for the first layer is to incorporate exponential decaying memory into the neurons that make them model a concept of recent past state. So if it’s a β€œcat” cell we can understand it to represents the odds that a cat is part of our environment.

17.11.2024 17:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The hard work of the first layer identified we just pressed a button. The second layer learns if we should do more or less of that in the future.

17.11.2024 17:03 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The first layer is temporal association learning. It’s classical conditioning and it’s what the cerebral cortex does for us. Once decided and simplified to a very large but sparse feature space the RL becomes simple and fast.

The second layer is the operant conditioning step.

17.11.2024 17:03 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I don’t have a full system working but I’m suspecting the RL step might be reduced to a single level linear map from state to actions. The first layer gives the brain its understand of how the world evolves but doesn’t decide how we respond. The RL level controls how we respond.

17.11.2024 17:03 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

But I’ve made good progress on that first layer and I’m seeing that it’s far more significant than I realized. After the first layer extracts a simplified representation of the environment from the raw data. The RL becomes simple.

17.11.2024 17:03 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

So I have for a very long time (20 years) believed we needed a two layer cake. But I was thinking the first layer would be something simple and small. Like the cake plate. But the real cake was all RL. or at best the two layers were equal in size and complexity.

17.11.2024 17:03 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The first layer of the cake is the last missing piece of the AGI puzzle. And it’s why we could use RL so effectively to solve problems that have an already simplified environment like board games but not apply it well to high dimensional problems. Get the first layer right then RL becomes easy.

17.11.2024 16:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The first step does all the heavy lifting of transforming the chaos of raw data into a highly simplified set of low frequency feature signals (grandmother neurons). Once the world is so simplified the RL step becomes so simple it is just a cherry on top. It’s why RL learning is so fast for us.

17.11.2024 16:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

As I see it we have two layers. The unsupervised perception snd prediction system which dynamically learns to translate raw sensory data into invariant state features and the second layer that maps these state signals into actions with RL.

17.11.2024 16:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

His second layer is just wrong however. We have no supervised learning hardware. He’s conflating how we use the hardware with what the hardware does. We teach each other in using supervision but the cake doesn’t have a mimic learning hardware layer.

17.11.2024 16:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Being a big RL believer since the 80s I was put off with his cake theme reducing it to an insignificant cherry. To me it was the real meat of the problem. But I’m far more in tune with his thinking now.

17.11.2024 16:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I mostly agree with the
cake idea but ChatGPT is not the cake. Not at all. But we aren’t far away from having the cake and ChatGPT 2.

What ChatGPT shows us is just how much human knowledge can be encoded in neural networks.

17.11.2024 16:24 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

But then the main training of ChatGPT is back prop training which the brain doesn’t use at all. Nothing is telling the brain that it computed the wrong muscle twitch

It’s all teal and error with rewards telling us what is right and wrong.

17.11.2024 16:24 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

ChatGPT doesn’t work like that as far as I understand it. It has no perception learning. The preprocessing step that converts text to tokens is the first two layers of the cake and I don’t think it’s learned and dynamic. It was hand crafted by the programmers. (Not sure about this).

17.11.2024 16:24 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Hello World!

16.11.2024 20:42 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

@curtwelch is following 20 prominent accounts