Joe Bak-Coleman's Avatar

Joe Bak-Coleman

@jbakcoleman.bsky.social

Research Scientist at the University of Washington based in Brooklyn. Also: SFI External Applied Fellow, Harvard BKC affiliate. Collective Behavior, Statistics, etc..

9,863 Followers  |  1,898 Following  |  4,191 Posts  |  Joined: 28.04.2023
Posts Following

Posts by Joe Bak-Coleman (@jbakcoleman.bsky.social)

Its possible big discoveries and important steps lie more in the unpredictable combinations than the predicable ones. Collective behavior is one example (computer graphics techniques + anima behavior questions)

06.03.2026 00:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

It’s more the difficulty of altering dynamics without degrading functionality. Collective behavior is so fickle and fragile it’s quite strange science works at all, in some ways.

For example, how does noisy and finite but diverse individual reads of the literature differ from all getting summaries

06.03.2026 00:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Stewardship of global collective behavior | PNAS Collective behavior provides a framework for understanding how the actions and properties of groups emerge from the way individuals generate and sh...

We wrote about the general idea here. From what we know about collective behavior, it’s highly unlikely that altering it for short term profits is going to produce desirable outcomes at scale.

www.pnas.org/doi/10.1073/...

05.03.2026 21:17 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I think the termite analogy gets us to nearly the opposite conclusion. Animal collective behavior often relies on highly selected rules for interaction, that emergently produce functionality. Altering that with robot termites that don’t follow those rules would almost certainly collapse functioning.

05.03.2026 21:15 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
A social media post from Donald J. Trump on Truth Social, posted 9 minutes ago. The text reads: "Iran tried to interfere in 2020, 2024 elections to stop Trump, and now faces renewed war with United States:" followed by a link to justthenews.com. Below is a link preview showing the article headline "Iran tried to interfere in 2020, 2024 elections to stop Trump, and now faces possible war with U.S." accompanied by a photo from what appears to be an Iranian protest or rally, showing demonstrators holding anti-Trump signs and posters with Persian text.

A social media post from Donald J. Trump on Truth Social, posted 9 minutes ago. The text reads: "Iran tried to interfere in 2020, 2024 elections to stop Trump, and now faces renewed war with United States:" followed by a link to justthenews.com. Below is a link preview showing the article headline "Iran tried to interfere in 2020, 2024 elections to stop Trump, and now faces possible war with U.S." accompanied by a photo from what appears to be an Iranian protest or rally, showing demonstrators holding anti-Trump signs and posters with Persian text.

I am going to need every single person who said that misinformation was a moral panic to feel enough shame and regret that they leave public intellectual life.

28.02.2026 16:30 β€” πŸ‘ 1392    πŸ” 277    πŸ’¬ 32    πŸ“Œ 28

Related point, if you are in misinformation research and have been launching study after study to see what Uncle Fred thinks after seeing misinformation PLEASE listen to what @katestarbird.bsky.social and I have been saying for years: effects on elite and official action is what matters most.

28.02.2026 16:36 β€” πŸ‘ 271    πŸ” 81    πŸ’¬ 9    πŸ“Œ 4

I suspect it is! There’s been some work on ai companies and research, but I think largely focused on ai research rather than its consequences.

I do think the field will turn up messy results unless we start getting cois straightened out.

27.02.2026 13:10 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I'm with @devezer.bsky.social, you *absolutely* cannot test for p-hacking.

These tests universally assume p-hacking, and nothing else under the sun, explain deviations from idealized distributions and then claim to have found it.

26.02.2026 21:56 β€” πŸ‘ 21    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

Fantastic thread.

26.02.2026 15:10 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

For example if someone wants to tell me about space travel, I should treat that information differently if they are a friend I trust who has a PhD in astrophysics, versus some random person on the internet, versus a psychotic billionaire trying to sell a space travel company

9/

26.02.2026 13:18 β€” πŸ‘ 109    πŸ” 12    πŸ’¬ 1    πŸ“Œ 2

If the sign up bonus is good, I’d even give it a depression survey and plot the results over time.

Stupidest job ever, but like… I’d take it.

26.02.2026 13:15 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

For $400k a year, I’d ask Claude how it’s doing each day and send Dario Amodei a thumbs up emoji on slack.

26.02.2026 13:13 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Preview
Living the metascience dream (or nightmare) with AI for science What happens when we go from replication crisis to robustness extremes?

AI makes continuous reproducibility and robustness testing trivial. What happens to science under new levels of scrutiny and stress-testing by default?

Some thoughts on how this could play out, informed by watching open science play out over the last decade.

23.02.2026 18:17 β€” πŸ‘ 58    πŸ” 20    πŸ’¬ 1    πŸ“Œ 10
Preview
AI, peer review and the human activity of science When researchers cede their scientific judgement to machines, we lose something important.

Good piece. Carl and I reached similar conclusions about ai in peer review. It’s good to see the hype tamed with what it is we actually do when we’re doing science.

www.nature.com/articles/d41...

25.02.2026 01:32 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

My experience with trying to get LLMs to write statistical models is that they will happily and silently bury an implausible assumption about the data generating process in an unobjectionable, conventional structure.

24.02.2026 23:34 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Of course we shouldn’t condition our model choices on the inferences we draw, but the models form is the questions they askβ€”based on what we know, assume, value etc..

I truly don’t see how that can be automated, it’s where our knowledge and data actually meet.

24.02.2026 23:33 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

It’s as if when we choose the terms in a model or its form we’re guessing and somewhere out there is the right answer, the biggest threat to getting it wrong is that we might choose the model based on our subjectivity.

So we want to hedge across options, or ask a machine to do it.

24.02.2026 23:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

More seriously, a lot of the LLMs in analysis discourse seems to be stuck in the same sort of thinking that motivated avetsging full specification curves of every possible model, or many analysts studies.

24.02.2026 23:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Less seriously, I’m gonna be big mad when the EC2 instance checking my reproducibility OOMs and rejects submission.

24.02.2026 23:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Many are appropriately outraged by Altman’s comments here implying that raising a human child is akin to β€œtraining” an AI model.

This is part of a broader pattern where AI industry leaders use language that collapses the boundary between human and machine.

🧡/

22.02.2026 19:29 β€” πŸ‘ 491    πŸ” 199    πŸ’¬ 28    πŸ“Œ 22

"There are deeper strata of truth…there is such a thing as poetic, ecstatic truth. It is mysterious and elusive, and can be reached only through fabrication and imagination and stylization." - Werner Herzog

22.02.2026 19:10 β€” πŸ‘ 90    πŸ” 20    πŸ’¬ 2    πŸ“Œ 0

Frankly if you can’t figure out how to use a proper environment for your code it raises deeper questions about the analysis.

20.02.2026 02:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Definitely inspired by it. I don’t understand conceptually adding something non reproducible to the problem of reproducibility when it’s a small infrastructure investment to ensure code compiles the way we do with latex.

20.02.2026 02:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

Most code I’ve struggled to reproduce could be addressed with a little template repository, even just using package management.

Containers make it all easier, but it’s weird how rarely someone makes it clear which version of (say) R they used.

20.02.2026 01:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

I’ve lost track of downloading someone’s R code and spending three hours figuring out what version of which packages don’t cause things to break, while replacing hardcoded paths everywhere.

20.02.2026 01:15 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I think the benefits sort of pay for themselves in terms of avoiding dependency hell when writing code. There is a learning curve but not particularly steep re: eveything else in science. For example, even using like pip freeze > requirements.txt and relative paths would go a long way.

20.02.2026 01:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

My old boss and current boss happened to meet halfway across the world today, and im truly hoping they took the piss out of me.

20.02.2026 01:11 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

β€œHave the agentic robot do it”

Vs.

β€œUse nix, GitHub, docker, or even just decent dependency management”

19.02.2026 23:55 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

People excited about computational reproducibility with LLMs are gonna lose their minds when the llm tells them that it’s largely a solved problem with dozens of suitable tools and it just requires knowing basic software development.

19.02.2026 23:54 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 2    πŸ“Œ 0

There's also plenty of complexity inherent to physics (the beast of social systems) but like lots of complex systems, you get emergent forms of simplicity amenable to models. I'm not sure we've done a good job integrating theory and empirics because of the fallout over homo-economicus.

18.02.2026 20:45 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0