I have, and he is almost certainly incorrect about Anthropicβs profitability
05.03.2026 13:34 β π 1 π 0 π¬ 1 π 0@alexanderhoyle.bsky.social
Postdoctoral fellow at ETH AI Center, working on Computational Social Science + NLP. Previously a PhD in CS at UMD, advised by Philip Resnik. Internships at MSR, AI2. he/him. On the job market this cycle! alexanderhoyle.com
I have, and he is almost certainly incorrect about Anthropicβs profitability
05.03.2026 13:34 β π 1 π 0 π¬ 1 π 0I'm so sorry to hear this. Your work is consistently thoughtful, distinctive, and impactful. I have so appreciated your mentorship and perspective, and am sure you will find a place where your voice is valued
03.03.2026 21:46 β π 1 π 0 π¬ 1 π 0hoping to develop something like this :)
23.02.2026 13:49 β π 2 π 0 π¬ 1 π 0You might find this interesting rycolab.io/classes/phil...
20.02.2026 06:53 β π 2 π 0 π¬ 0 π 0
Already a thing!
www.cs.umd.edu/~jbg//docs/2...
aclanthology.org/2023.acl-lon...
(Iβm sure thereβs more)
I appreciate the shout-out to our work showing that LDA tends to work better, although it would have been nice to include it as a comparison :)
28.01.2026 09:54 β π 5 π 0 π¬ 1 π 0
I thought these were deeper explorations of LM βunderstandingβ than the Stochastic Parrots paper:
direct.mit.edu/coli/article...
julianmichael.org/blog/2020/07...
direct.mit.edu/coli/article...
Also see Will Merrillβs work for more formal treatments
Yeah, that denominator was what I was wondering about, since my understanding is usually it's something like 5-10 onsites per season?
I did a "toe-dipping" last cycle (first year of postdoc), and I wasn't sure how to consider the proportion of onsites I got relative to screenings
I was discussing first rounds with a colleague recently---do you have a sense of the typical "conversion" rate to onsites?
[Also, sorry for dropping into your mentions for the second time today]
Does it create good diagrams?
20.01.2026 11:58 β π 0 π 0 π¬ 1 π 0Oh amazing! Do you have the datasets you used to test it? Iβve been playing around with a project related to this topic
20.01.2026 10:45 β π 3 π 0 π¬ 2 π 0This is super great---a nice, illustrative collection of measurement issues. Thanks for sharing
19.01.2026 14:38 β π 1 π 0 π¬ 0 π 0FWIW, auto completion models are worse because they are cheap and run all the time. the larger models/integrated systems (like Claude code) are significantly better
09.01.2026 12:51 β π 6 π 0 π¬ 2 π 0
I think you are relying on two assumptions: (a) they used a commercial LLM to begin with and (b) that commercial LLMs train on all data that is passed to them
I don't know about (a), but (b) is not the case when using the API endpoints for the major providers
platform.openai.com/docs/concepts
Sincere question: how do you distinguish your paper being passed to a commercial LLM from your paper being indexed by commercial search engines?
08.01.2026 16:10 β π 0 π 0 π¬ 0 π 0Itβs also weirdly dependency-averse (maybe it has been RLHFβd away from using outdated/old libraries?). I was vibe coding an iOS app and it reimplemented a full photo editor. I know nothing about the Swift ecosystem, but I figured there had to have been some existing library (there is of course)
08.01.2026 10:40 β π 1 π 0 π¬ 0 π 0Now I'm late to replying---thanks, this means a lot coming from you! For future work, are looking at ways to apply it to mech interp, but it's still in the early stages. I'm also interested in the "reverse" problem: transformations that improve similarities along a given dimension (e.g., syntax)
07.01.2026 11:25 β π 0 π 0 π¬ 1 π 0Happy to be in the second round!
03.01.2026 09:33 β π 2 π 0 π¬ 0 π 0Was just browsing βFor Youβ and didnβt expect to see one of ours here! Some here I havenβt seen, thanks for the pointers :)
02.01.2026 20:51 β π 2 π 0 π¬ 0 π 0Obsidian does calendar management?
02.01.2026 15:00 β π 1 π 0 π¬ 1 π 0Interestingβwhatβs the difference from Claude Code/Codex?
29.12.2025 01:55 β π 0 π 0 π¬ 1 π 0Ah l see you wrote about this in the original thread, apologies!
19.12.2025 16:27 β π 1 π 0 π¬ 1 π 0thatβs fair, the author is definitely putting forward a position. Your argument is that cog sci anticipated these developments somewhat, or? (May have not gathered it in the thread)
19.12.2025 16:10 β π 0 π 0 π¬ 1 π 0
Just read a nice New Yorker piece that interviews mostly cognitive scientists/neuroscientists who are (mostly) very impressed:
www.newyorker.com/magazine/202...
Itβs unfortunate because Iβve found LDA (mallet/tomotopy, not gensim) is actually very good for tweets
09.12.2025 11:20 β π 1 π 0 π¬ 1 π 0my poor heart π
09.12.2025 11:19 β π 1 π 0 π¬ 0 π 0Realizing neither of you have John Williams' Stoner, which is a beautiful book
06.12.2025 21:27 β π 0 π 0 π¬ 0 π 0I enjoyed Changing Places, and there's two more in the "campus trilogy". I don't recall much about the third (if I even got to it?), but Small World was fun---more in the left quadrant
06.12.2025 21:24 β π 1 π 0 π¬ 1 π 0It's funny, when I've spoken to some ML people---including, just recently, Dave Blei himself---the question is often, "who's still using topic models?" Meanwhile in DH/Social science, it's "please, enough topic models!"
06.12.2025 21:04 β π 2 π 1 π¬ 0 π 0
Some collections of work in these directions:
openreview.net/pdf?id=gJcEM...
arxiv.org/abs/2210.13382
iclr.cc/virtual/2025...
www.worldmodelworkshop.org