YouTube video by Machine Learning Street Talk
Do you think that ChatGPT can reason?
Everybody talking about the "new" apple paper might find this MLST interview with @rao2z.bsky.social interesting. "Reasoning" and "inner thoughts" of LLMs were exposed as self-mumblings and fumblings long ago. #LLMs #AI
www.youtube.com/watch?v=y1Wn...
08.06.2025 19:40 β π 5 π 0 π¬ 0 π 0
yes - people will still need a phone, and a lot of AI products, services, and UI will need a screen. and a touchable one at that.
03.06.2025 01:46 β π 1 π 0 π¬ 0 π 0
Backtracing: Retrieving the Cause of the Query
Many online content portals allow users to ask questions to supplement their understanding (e.g., of lectures). While information retrieval (IR) systems may provide answers for such user queries, they...
Interesting - could #LLMs in search capture context missed when googling?
"backtracing ... retrieve the cause of the query from a corpus. ... targets the information need of content creators who wish to improve their content in light of questions from information seekers."
arxiv.org/abs/2403.03956
14.05.2025 14:55 β π 0 π 0 π¬ 0 π 0
They mostly test whether they can steer pos/neg responses. But given Shakespeare was a test also, wld be interesting to extract style vectors from any number of authors then compare generations. (Is this approach used in those "historical avatars?" No idea.)
14.05.2025 14:42 β π 0 π 0 π¬ 0 π 0
Style Vectors for Steering Generative Large Language Model
This research explores strategies for steering the output of large language models (LLMs) towards specific styles, such as sentiment, emotion, or writing style, by adding style vectors to the activati...
@tedunderwood.me In case you haven't seen this paper, you might find interesting. Researchers extract style vectors (incl from Shakespeare) and apply to an LLM internal layers instead of training on original texts. Generations can then be "steered" to a desired style.
arxiv.org/abs/2402.01618
14.05.2025 14:37 β π 1 π 0 π¬ 1 π 0
But design will need to focus on tweaking model interactions so that they track conversational content and turns over time. For example with bi-directional prompting: models prompt users to keep conversations on track.
This seems a rich opportunity for interaction design #UX #IxD #LLMs #AI
14.05.2025 13:38 β π 1 π 0 π¬ 0 π 0
to sustain dialog. Social interaction face to face or online is already vulnerable to misunderstandings and failures, and we have use of countless signals, gestures, etc w which to rescue our interactions.
A communication-first approach to LLMs for conversation makes sense, as talk is not writing.
14.05.2025 13:38 β π 0 π 0 π¬ 1 π 0
"when LLMs take a wrong turn in a conversation, they get lost and do not recover."
Interaction design is going to be necessary to scaffold LLMs for talk, be it voice or single user chat or multi-user (e.g. social media).
It's one thing to read/summarize written documents, quite another ...
14.05.2025 13:38 β π 0 π 0 π¬ 1 π 0
LLMs Get Lost In Multi-Turn Conversation
Large Language Models (LLMs) are conversational interfaces. As such, LLMs have the potential to assist their users not only when they can fully specify the task at hand, but also to help them define, ...
"LLMs tend to (1) generate overly verbose responses, leading them to (2) propose final solutions prematurely in conversation, (3) make incorrect assumptions about underspecified details, and (4) rely too heavily on previous (incorrect) answer attempts."
arxiv.org/abs/2505.06120
14.05.2025 13:38 β π 0 π 0 π¬ 1 π 0
Attention Mechanisms Perspective: Exploring LLM Processing of Graph-Structured Data
Attention mechanisms are critical to the success of large language models (LLMs), driving significant advancements in multiple fields. However, for graph-structured data, which requires emphasis on to...
"LLMs ... recognize graph-structured data... However... we found that even when the topological connection information was randomly shuffled, it had almost no effect on the LLMsβ performance... LLMs did not effectively utilize the correct connectivity information."
www.arxiv.org/abs/2505.02130
14.05.2025 13:11 β π 0 π 0 π¬ 0 π 0
Perhaps one could fine tune on Lewis Carroll, then feed the model with philosophical paradoxes, and see whether the model produces more imaginative generations.
12.05.2025 17:21 β π 0 π 0 π¬ 0 π 0
I think because this isn't making the model trip, synesthetically, but is simply giving it juxtapositions. So what is studied is a response to these paradoxical and conceptually incompatible prompts, not a measure of any latent conceptual activations or features.
12.05.2025 17:21 β π 0 π 0 π¬ 1 π 0
Yes and the label applied says as much about the person as it does about the model. In the world of creatives, the most-used term now is "slop," derived perhaps from enshitification. The latter capturing corporate malice where the "slop" is AI-generated byproduct unfit for human consumption...
10.05.2025 17:08 β π 1 π 0 π¬ 0 π 0
Thread started w your second post so yes I missed the initial post. Never mind.
10.05.2025 16:53 β π 0 π 0 π¬ 0 π 0
Assuming alignment using synthetic data is undesirable, one route is to complement global alignment (alignment to some "universally" preferred human values) w local, contextualized alignment, via feedback and use by the user. Tune the LLM's behavior to user preferences.
10.05.2025 16:43 β π 0 π 0 π¬ 1 π 0
Customized LLMs use the feedback obtained from the individual user interactions and align to those.
10.05.2025 16:34 β π 0 π 0 π¬ 1 π 0
Staying power of ceasefires becoming a proxy for multilateral resilience amid baseline rivalries?
10.05.2025 16:33 β π 1 π 0 π¬ 0 π 0
I think this will be one accelerant for individualized/personally customized AI - e.g. personal assistants. The verifiers can use the user's preferences and tune to those rather than apply globally aligned behavioral rules.
10.05.2025 16:29 β π 0 π 0 π¬ 1 π 0
It's also a problem of use cases and user adoption. Though it may turn out that Transformer-based AI does indeed fail to meet expectations.
There's a lot of misunderstanding and anthropomorphism of AI's reasoning, for example, that might not turn out well.
10.05.2025 16:27 β π 1 π 0 π¬ 0 π 0
Coincidentally many startups of that time set up in loft & warehouse spaces w exposed concrete & steel beams.... I like this analogy especially for Social Interaction Design/Social UX, where "social architecture" is exposed for users to take up in norms, behaviors, expectations for how to engage
10.05.2025 16:24 β π 2 π 0 π¬ 0 π 0
I can't disagree w that. Reflection through reading employs more critical thinking skills than conversation; bots solicit unserious interaction & even attempts to "hack" guardrails. I'm a huge reader but I do have lengthy convos w ChatGPT, likely because I read/reflect.
06.05.2025 15:04 β π 0 π 0 π¬ 1 π 0
Agree w you. Tariffs as targeted protections of domestic ind, as reciprocity, as reshoring incentives, as embargoes - all these are different & neglect unintended consequences as we're seeing in markets & bonds & dollar.
Regardless of motives it's now a matter of game theory - who moves, when, etc
06.05.2025 14:53 β π 2 π 0 π¬ 0 π 0
For now I can see that chatbots likely would fail to provide accurate or probable reasoning if prompted for explanations of historical choices, actions, etc, for lack of proper historical context. But this too could be improved w training on secondary lit.
It's admittedly all rather Black Mirror.
06.05.2025 14:19 β π 0 π 0 π¬ 0 π 0
To learn a historical figure from a book however is to imagine their reasons, motives, actions in abstract. (Which is fine.) To have them personified as chatbots seems absurd and kitschy - but might reach some students who simply don't engage by reading.
06.05.2025 14:19 β π 0 π 0 π¬ 2 π 0
These bots likely are built on texts but not graphs, w which they could better hew to facts, etc. They might be trained to better interact - but this would be to layer pedagogical learning methods onto the bot's conversation style (is still interesting). On moral view, you're absolutely right.
06.05.2025 14:19 β π 1 π 0 π¬ 1 π 0
Co-founder & CEO at Searchcraft. We help digital product teams ship faster with accessible, enterprise-grade content discovery tools.
Background in UI/UX. Former agency owner. Lifelong creative.
πDenver, Colorado, USA
Digital Transformation Programme Manager at Queenβs University Belfast.
Intelligent Interfaces for Frontier Technologies. schemadesign.com
Founding Partner @schemadesign.com. Exploring the edge of AI, spatial computing, & data.
Design Researcher curious about complex systems
life is short and I am in a hurry.
tkukurin.github.io
Postdoc @ TakeLab, UniZG | previously: Technion; TU Darmstadt | PhD @ TakeLab, UniZG
Faithful explainability, controllability & safety of LLMs.
π On the academic job market π
https://mttk.github.io/
Cognitive scientist, postdoc at Justus Liebig University, Giessen. Natural/artificial vision/cognition.
Assistant Professor of Computational Linguistics @ Georgetown; formerly postdoc @ ETH Zurich; PhD @ Harvard Linguistics, affiliated with MIT Brain & Cog Sci. Language, Computers, Cognition.
π₯ LLMs together (co-created model merging, BabyLM, textArena.ai)
π₯ Spreading science over hype in #ML & #NLP
Proud shareLM㪠Donor
@IBMResearch & @MIT_CSAIL
Postdoc at Northeastern and incoming Asst. Prof. at Boston U. Working on NLP, interpretability, causality. Previously: JHU, Meta, AWS
Asst Prof. @ UCSD | PI of LeMπN Lab | Former Postdoc at ETH ZΓΌrich, PhD @ NYU | computational linguistics, NLProc, CogSci, pragmatics | he/him π³οΈβπ
alexwarstadt.github.io
NLP, Linguistics, Cognitive Science, AI, ML, etc.
Job currently: Research Scientist (NYC)
Job formerly: NYU Linguistics, MSU Linguistics
Studying digital service development.
In Finnish:
ViestintÀÀ, markkinointia, ohjelmistokehitystÀ, valokuvausta, graafista suunnittelua, ja ties mitÀ muuta opiskellut.
English and Finnish posts. Language filter in the Bluesky settings can be useful.
Coming in January, 2026: The Global Casino: How Wall St Gambles with People and Planet. In the meantime check out my newsletter: System Change, https://annpettifor.substack.com/
I teach philosophy part time, and I work full time as a creative producer in NYC.
Data management SME, author of "SQL in a Nutshell", grandpa to 3, dad to 7, husband to 1, master gardener, indie rock enthusiast. Ex-NASA, DoD, Deloitte, AWS. All opinions are my own.
Writing a book about Joseph Stalin's management philosophy: how it manifests in tyrants great and small, and how the rest of us can deal with the BS. Newsletter here: https://masonpelt.substack.com/
Ithaca | prev Chicago | interested in interconnected systems and humans+computers | past and future: academic and industry research | currently: gardening