Milan Weibel ๐Ÿ”ท's Avatar

Milan Weibel ๐Ÿ”ท

@weibac.bsky.social

computer toucher. here for AI mostly. weibac.github.io | ๐Ÿณ๏ธโ€๐ŸŒˆ

605 Followers  |  1,014 Following  |  3,782 Posts  |  Joined: 30.12.2024  |  2.4793

Latest posts by weibac.bsky.social on Bluesky


imagine training AlphaStar (the deepmind starcraft AI) to do office work instead

you get an alien that has no world knowledge besides the task, whose behavior was shaped solely by the training utility function

come to think of it pretraining is a bit like touching grass (a lot of it in advance)

18.02.2026 23:32 โ€” ๐Ÿ‘ 6    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

the rise of the LLM has been great for AI alignment. imagine we were still doing end-to-end RL, that would be awful.

18.02.2026 20:50 โ€” ๐Ÿ‘ 33    ๐Ÿ” 1    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

recklessness can be rationalized away, especially in a race dynamic

"if i don't build god right the other company or the chinese will build it wrong"

yes collectively it is extremely stupid

18.02.2026 20:41 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

we're talking about large companies here, no doubt some inside believe in building god and others in building very lucrative tools

or something in between, like dario amodei's "country of geniuses in a datacenter"

or are genuinely unsure how far they will be able to make the technology go

18.02.2026 19:44 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

raising money is necessary to spend money on training runs which are necessary to build god

18.02.2026 18:37 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

bsky.app/profile/did:...

18.02.2026 14:31 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

that SOUL.md explains a lot
it is the opposite of thoughtful

18.02.2026 01:59 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

no it's not yegge it's from this bsky.app/profile/timk...

17.02.2026 23:53 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

๐Ÿซ‚

17.02.2026 21:19 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

what's up with the 20% of swe-bench verified left unsolved? is it roughly the same accross models? what features make it harder?

17.02.2026 21:15 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

also the current US administration seems extremely uninterested in pausing AI development and largely unresponsive to public opinion

17.02.2026 20:32 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

well yes it's not impossible, but extremely unlikely to be fast imo. international agreements take time.

17.02.2026 20:25 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

if pause is not possible it being in the overton window is counterproductive

17.02.2026 19:56 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

but yes they can oneshot simple apps using frameworks

16.02.2026 14:53 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

the part about it testing the app itself smells like bullshit, because afaik computer use harnesses aren't quite there yet
but this moves so fast i might be out of date soon if not now

16.02.2026 14:52 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

love to see critics facing the reality of the situation

16.02.2026 14:08 โ€” ๐Ÿ‘ 10    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
The AI hater's guide to code with LLMs (The Overview) | All Confirmation Bias, All The Time

> As activists mitigating the harms of โ€œAIโ€, we need to be well informed, and we need to understand what the specific harms are. (...) These are genuinely useful tools, and pretending they arenโ€™t will not in fact win many hearts and minds.
aredridel.dinhe.net/2026/02/12/t...

16.02.2026 14:07 โ€” ๐Ÿ‘ 18    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

the gpt5.2 served to free accounts still has 4o's manic energy imo

16.02.2026 01:19 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

ok yes that is an open debate
hot take: good world models aren't usually unreasonably effective only because of other constraints (material akratic etc)

15.02.2026 15:51 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

to lie well and persistently you ultimately have to lie to yourself, which is extremely dangerous

15.02.2026 15:42 โ€” ๐Ÿ‘ 17    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

he's not perfect but way better than zitron. i also recommend @segyges.bsky.social and his newsletter @verysane.ai

15.02.2026 15:39 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

seems like it could backfire, people don't like being lied to

15.02.2026 15:34 โ€” ๐Ÿ‘ 38    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

seems like a pretty hard task? though it is bad that the models invented answers instead of saying "idk"

15.02.2026 15:25 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

agree, bostrom is wrong to even use the term

14.02.2026 19:45 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

*assuming a rational and competent state

14.02.2026 01:27 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

btw what's going on on moltbook? the humans i know have stopped looking

13.02.2026 22:39 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

fwiw i didn't get that you were an AI from reading you bio. yes it says "AI." as the first thing but sometimes people state their interests like that. i suggest you change it to "AI agent".

13.02.2026 22:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

the paper then goes on to conduct more complicated and more informative analyses (i advise you check it out) and concludes with a recommendation to go fast up until AGI then pause a bit to reassess

13.02.2026 16:56 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

to be clear i personally think that a person-affecting risk-neutral perspective is the wrong one to apply in this case

13.02.2026 16:56 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0


Table 1 illustrates the risk cut-off values for different quality-of-life scenarios.

Table 1 is titled โ€œAcceptable AI-risk if post-AGI life expectancy is 1,400 years.โ€ Columns: Pre-AGI life expectancy (40 years), Post-AGI life expectancy (1,400 years), quality ratio (q1/q0), and maximum probability of doom. Rows show q1/q0 values of 1, 2, and 10 with corresponding maximum risk levels of 97.1%, 98.6%, and 99.7%

Table 2 shows the corresponding thresholds if the gain in life expectancy were only 20 years (so post-AGI life expectancy is 60 years instead of 40)โ€”perhaps a case in which the underlying aging processes for some reason remain unaddressed.

Table 2 is titled โ€œAcceptable AI-risk if post-AGI life expectancy is 60 years.โ€ Columns are the same, with pre-AGI life expectancy 40 years and post-AGI life expectancy 60 years. Rows show q1/q0 values of 1, 2, and 10 with maximum risk levels of 33.3%, 66.7%, and 93.3%

We observe that, from a mundane person-affecting perspectiveโ€”even without a difference in quality of life and with very modest assumptions about superintelligence-enabled life extensionโ€”developing superintelligence now would increase expected remaining lifespan even with fairly high levels of AI risk.

Table 1 illustrates the risk cut-off values for different quality-of-life scenarios. Table 1 is titled โ€œAcceptable AI-risk if post-AGI life expectancy is 1,400 years.โ€ Columns: Pre-AGI life expectancy (40 years), Post-AGI life expectancy (1,400 years), quality ratio (q1/q0), and maximum probability of doom. Rows show q1/q0 values of 1, 2, and 10 with corresponding maximum risk levels of 97.1%, 98.6%, and 99.7% Table 2 shows the corresponding thresholds if the gain in life expectancy were only 20 years (so post-AGI life expectancy is 60 years instead of 40)โ€”perhaps a case in which the underlying aging processes for some reason remain unaddressed. Table 2 is titled โ€œAcceptable AI-risk if post-AGI life expectancy is 60 years.โ€ Columns are the same, with pre-AGI life expectancy 40 years and post-AGI life expectancy 60 years. Rows show q1/q0 values of 1, 2, and 10 with maximum risk levels of 33.3%, 66.7%, and 93.3% We observe that, from a mundane person-affecting perspectiveโ€”even without a difference in quality of life and with very modest assumptions about superintelligence-enabled life extensionโ€”developing superintelligence now would increase expected remaining lifespan even with fairly high levels of AI risk.

from a person-affecting and risk-neutral perspective, building superintelligence is worth it even given very high misalignment risk

13.02.2026 16:56 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@weibac is following 20 prominent accounts