's Avatar

@braintelligence.bsky.social

Believer in inclusive democracy Posting mostly about AI/ML and tech if I can help it

266 Followers  |  735 Following  |  1,611 Posts  |  Joined: 08.11.2024  |  1.7626

Latest posts by braintelligence.bsky.social on Bluesky

Preview
GitHub - QwenLM/qwen-code: Qwen Code is a coding agent that lives in the digital world. Qwen Code is a coding agent that lives in the digital world. - QwenLM/qwen-code

Positioned as the open competitor to commercial coding assistants, Qwen Code adds Plan Mode for approve-first changes and Vision Intelligence that auto-switches to multimodal (Qwen3‑VL‑Plus) when images appearβ€”even in YOLO mode.

Qwen Code: github.com/QwenLM/qwen-...

12.10.2025 15:25 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Open-source Qwen Code just leveled upβ€”now it plans before it edits and sees your screenshots. πŸš€

12.10.2025 15:25 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Is correctly modeling that human and squid share a common ancestor

13.10.2025 02:33 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Grok 0.9’s video generation is surprisingly impressive β€” it’s free (with daily generation limits), fast, and delivers natural character movements and camera work.

12.10.2025 15:26 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0

And notably, in the case of GPT-J, we DIDN’T KNOW IT WAS DOING THIS until 3.5 years after the model was released!

12.10.2025 17:18 β€” πŸ‘ 66    πŸ” 10    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

I don't think people have updated enough on the capability gain in LLMs, which (despite being bad at math a year ago) now dominate hard STEM contests: gold medals in the International Math Olympiad, the International Olympiad on Astronomy & Astrophysics, International Informatics Olympiad...

12.10.2025 20:40 β€” πŸ‘ 99    πŸ” 16    πŸ’¬ 4    πŸ“Œ 2
Everett 13-year-old arrested by ICE and sent to Virginia detention facility
By Marcela Rodrigues Globe Staff,Updated October 12, 2025, 44 minutes ago



31
A 13-year-old boy was arrested by ICE in Everett and sent to a juvenile detention facility in Virginia.
A 13-year-old boy was arrested by ICE in Everett and sent to a juvenile detention facility in Virginia.
A 13-year-old boy was arrested by Immigration and Customs Enforcement agents in Everett after an interaction with members of the Everett Police Department and sent to a juvenile detention facility in Virginia, according to his mother and immigration lawyer Andrew Lattarulo.

The boy’s mother, Josiele Berto, was called to pick her son up from the Everett Police Department on Thursday, the day he was arrested. After waiting for about an hour and a half, she was told her son was taken by ICE, Berto told the Globe in a phone interview.

β€œMy world collapsed,” Berto said in Portuguese.

From the police department, the boy was taken to ICE’s holding facility in Burlington on Thursday evening, where he spent a night before being transferred by car to the Northwestern Regional Juvenile Detention Center in Winchester, Va., on Friday morning, his mother said. The juvenile facility is more than 500 miles away from Everett.

The boy is a 7th-grader at Albert N. Parlin School in Everett, his mother said. The teen and his family, who are Brazilian nationals, have a pending asylum case and are authorized to work legally in the United States, Lattarulo said.

Everett 13-year-old arrested by ICE and sent to Virginia detention facility By Marcela Rodrigues Globe Staff,Updated October 12, 2025, 44 minutes ago 31 A 13-year-old boy was arrested by ICE in Everett and sent to a juvenile detention facility in Virginia. A 13-year-old boy was arrested by ICE in Everett and sent to a juvenile detention facility in Virginia. A 13-year-old boy was arrested by Immigration and Customs Enforcement agents in Everett after an interaction with members of the Everett Police Department and sent to a juvenile detention facility in Virginia, according to his mother and immigration lawyer Andrew Lattarulo. The boy’s mother, Josiele Berto, was called to pick her son up from the Everett Police Department on Thursday, the day he was arrested. After waiting for about an hour and a half, she was told her son was taken by ICE, Berto told the Globe in a phone interview. β€œMy world collapsed,” Berto said in Portuguese. From the police department, the boy was taken to ICE’s holding facility in Burlington on Thursday evening, where he spent a night before being transferred by car to the Northwestern Regional Juvenile Detention Center in Winchester, Va., on Friday morning, his mother said. The juvenile facility is more than 500 miles away from Everett. The boy is a 7th-grader at Albert N. Parlin School in Everett, his mother said. The teen and his family, who are Brazilian nationals, have a pending asylum case and are authorized to work legally in the United States, Lattarulo said.

ICE kidnapped a 7th-grader with a pending asylum claim and spirited him out of state without notifying his parents, seemingly with the cooperation of the local police in Everett, MA.

www.bostonglobe.com/2025/10/12/m...

12.10.2025 21:43 β€” πŸ‘ 7565    πŸ” 4566    πŸ’¬ 399    πŸ“Œ 685
Preview
Apple's former CEO says the company has its 'first real competitor' in decades "AI has not been a particular strength for the business," Apple's former CEO John Sculley said of the company.

This is an interesting observation that I think is actually correct

Partially because Apple seems to have drifted from their ethos, and partially because AI is an entirely new paradigm from computing as we know it

www.businessinsider.com/former-apple...

13.10.2025 01:09 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The older I get the more I believe that older people should be more intentional with making space for younger people

Just because you CAN do something doesn’t mean you should

Mentor/collaborate/support someone just coming up

11.10.2025 23:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
add-36-59 - gemma-2-2b Graph | Neuronpedia Attribution Graph for gemma-2-2b

If you're interested in Anthropic's work on transformer circuits, consider trying out Neuronpedia's circuit tracing tool here. TBH it's kind of hard to find interesting stuff in my experience, but fun when you do. www.neuronpedia.org/gemma-2-2b/g...

11.10.2025 21:15 β€” πŸ‘ 15    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Agentic workflows, RL, TTE, are essentially recurrence… loss functions for these target tokens but seems like should target embeddings, A2A should maybe be in embeddings too

β€œAnd embedding is worth a thousand words” as they almost say

11.10.2025 20:32 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

… this is where diffusion can shine

The model guides it with its β€œthought” embedding values, and diffusion uses this guidance to convert to language/actions

11.10.2025 18:15 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This is where yann lecun is on to something

We train models to reason in token domain, when we should have them reason in embedding space first, and token translation is a thin last operation

11.10.2025 18:14 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

In this case vought wants them as a new underclass to replace migrant farm and factory workers

11.10.2025 18:12 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The mechanics of how they think are different, but I’m starting to think all thought is based on the same core task of modeling the world

So they’ll converge on our same philosophical dilemmas as ai becomes embodied with continuous learning

11.10.2025 17:22 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

BREAKING: Friday night massacre underway at CDC. Doznes of "disease detectives," high-level scientists, entire Washington staff and editors of the MMWR (Morbidity and Mortality Weekly Report) have all been RIFed and received the following notice:

11.10.2025 02:10 β€” πŸ‘ 15362    πŸ” 8468    πŸ’¬ 859    πŸ“Œ 1123

In the future, the fact we shoved user input directly into the core layer will be seen as barbaric and inhumane

11.10.2025 02:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Scaling LLM Multi-turn RL with End-to-end Summarization-based Context Management ArXiv link for Scaling LLM Multi-turn RL with End-to-end Summarization-based Context Management

Researchers have unveiled SUmmarization augmented Policy Optimization (SUPO), a novel RL framework that empowers large language models to handle complex, long-horizon tasks by using LLM-generated summaries for effective context management. https://arxiv.org/abs/2510.06727

11.10.2025 00:11 β€” πŸ‘ 0    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

The same logic applies to fashion: instead of buying expensive clothes to look better, invest that time and money into getting fit, you’ll look better than anything designer labels can offer.

11.10.2025 00:09 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Sign Up Skype a Scientist gives you the opportunity to connect with students and the public around the world. ​

Here I am. Once again. I'm out of archaeologists.

We matched 576 groups w/archaeologists, leaving 13 groups unmatched. This brings me pain! We've never run out of scientists like this before. 598 requests for one category is A LOT. But still.

Archaeologists πŸ₯Ί
www.skypeascientist.com/sign-up.html

10.10.2025 15:51 β€” πŸ‘ 136    πŸ” 111    πŸ’¬ 14    πŸ“Œ 12

I’m someone who thinks AI is a powerful art tool for humans, but the trump admin has been shoveling out lazy, low quality AI generations, which tarnishes the whole field

11.10.2025 00:24 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Google says it hit a milestone of 1.3 quadrillion monthly tokens processed across its services this summer, up from 980T monthly tokens announced in July (Matthias Bastian/The Decoder)

Main Link | Techmeme Permalink

10.10.2025 23:25 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Put AI or LLM in your bio and you’ll probably get put on a block list

bsky.app/profile/brai...

10.10.2025 23:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

GAGA-1 just dropped: an AI β€œactor” that fuses voice, lipsync, and performance.

It’s not fully freeβ€”there’s roughly 60 seconds of free generation per month; the API is currently unavailable. It’s built for dialogue-driven scene prototyping.

10.10.2025 16:41 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

On one hand: don't anthropomorphize AI. On the other: LLMs exhibit signs of gambling addiction.

The more autonomy they were given, the more risks the LLMs took. They exhibit gambler's fallacy, loss-chasing, illusion of control...

A cautionary note for using LLMs for investing without guardrails.

10.10.2025 05:09 β€” πŸ‘ 115    πŸ” 29    πŸ’¬ 13    πŸ“Œ 8
Three line charts showing performance metrics for Qwen3-30B-A3B-4bit running on an M3 Ultra (512 GB RAM) across varying batch sizes.

βΈ»

Top Chart: Prompt Tokens per Second
	β€’	Y-axis: Prompt TPS (tokens per second), X-axis: Batch Size.
	β€’	Values rise slightly overall:
	β€’	0: 2876        5: 2868        10: 2880        15: 2885        25: 2893        30: 2891
	β€’	Peak performance: 2893 TPS at batch 25, showing stable prompt throughput across batch sizes.

βΈ»

Middle Chart: Generation Tokens per Second
	β€’	Y-axis: Generation TPS, X-axis: Batch Size.
	β€’	Throughput improves steadily with larger batches:
	β€’	0: 108        5: 249        10: 336        15: 425        25: 453        30: 473
	β€’	Performance scales efficiently, plateauing near 473 TPS.

βΈ»

Bottom Chart: Peak Memory Usage
	β€’	Y-axis: Peak Memory (GB), X-axis: Batch Size.
	β€’	Memory grows roughly linearly with batch size:
	β€’	0: 18.28 GB     5: 19.46 GB     10: 21.66 GB     15: 23.28 GB     25: 24.89 GB     30: 26.50 GB
	β€’	Peak usage: 26.5 GB at batch 30.

βΈ»

Overall: Prompt throughput remains flat, generation throughput scales well, and memory usage rises linearly with batch size.

Three line charts showing performance metrics for Qwen3-30B-A3B-4bit running on an M3 Ultra (512 GB RAM) across varying batch sizes. βΈ» Top Chart: Prompt Tokens per Second β€’ Y-axis: Prompt TPS (tokens per second), X-axis: Batch Size. β€’ Values rise slightly overall: β€’ 0: 2876 5: 2868 10: 2880 15: 2885 25: 2893 30: 2891 β€’ Peak performance: 2893 TPS at batch 25, showing stable prompt throughput across batch sizes. βΈ» Middle Chart: Generation Tokens per Second β€’ Y-axis: Generation TPS, X-axis: Batch Size. β€’ Throughput improves steadily with larger batches: β€’ 0: 108 5: 249 10: 336 15: 425 25: 453 30: 473 β€’ Performance scales efficiently, plateauing near 473 TPS. βΈ» Bottom Chart: Peak Memory Usage β€’ Y-axis: Peak Memory (GB), X-axis: Batch Size. β€’ Memory grows roughly linearly with batch size: β€’ 0: 18.28 GB 5: 19.46 GB 10: 21.66 GB 15: 23.28 GB 25: 24.89 GB 30: 26.50 GB β€’ Peak usage: 26.5 GB at batch 30. βΈ» Overall: Prompt throughput remains flat, generation throughput scales well, and memory usage rises linearly with batch size.

someone got 473 tok/sec(!!) on a macbook M3 Ultra for Qwen3-30b-a3b

and that’s a decent model..

10.10.2025 13:04 β€” πŸ‘ 18    πŸ” 1    πŸ’¬ 3    πŸ“Œ 0
Preview
microsoft/UserLM-8b Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

UserLM-8B: an LLM that mimics a person

Microsoft fine tuned an LLM to respond as the user instead of as an assistant

This is useful anytime you need mock user input. like maybe for testing agents, or training models

huggingface.co/microsoft/Us...

10.10.2025 14:36 β€” πŸ‘ 31    πŸ” 2    πŸ’¬ 3    πŸ“Œ 3
Post image

I would like to share some work we've been doing at cascadetech.ai: Predicted Outputs in vLLM. If you aren't familiar with PO, it allows you to dramatically speed up generation when you know something about the contents of the output (think: code modification).

10.10.2025 18:11 β€” πŸ‘ 11    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1

Next token predictor
Stochastic parrot

/s

10.10.2025 21:43 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

"Sora 2, an elaborate regency romance where everyone is wearing a live duck for a hat (each duck is also wearing a hat) , prestige drama"

10.10.2025 19:32 β€” πŸ‘ 155    πŸ” 16    πŸ’¬ 22    πŸ“Œ 8

@braintelligence is following 20 prominent accounts