Dr. Jonathan Mall's Avatar

Dr. Jonathan Mall

@jonathanmall.bsky.social

LLMs, AI, Psychology and Speaking - Seeking to infuse technology with empathy. That's why I founded www.neuroflash.com. An AI platform driving brand-aligned marketing, and helping people connect and understand each other's perspectives.

1,714 Followers  |  2,075 Following  |  23 Posts  |  Joined: 21.11.2024  |  1.8781

Latest posts by jonathanmall.bsky.social on Bluesky

Post image

I made an interactive conversation analysis of the (world changing?) #Trump #Zelenskyy #Vance conversation. (Transcript from @apnews.com)

claude.site/artifacts/f8...

Try it yourself β€” What stands out to you? πŸ‘€

03.03.2025 07:16 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Claude Artifact Try out Artifacts created by Claude users

Interactive version: claude.site/artifacts/f8...

02.03.2025 12:15 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Indeed, looks like an export mistake. It's correct in the dynamic version.

02.03.2025 12:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

The words we choose reveal more than we think. Language isn't just communication β€” it's a reflection of power, strategy, and worldview.

02.03.2025 11:19 β€” πŸ‘ 1    πŸ” 2    πŸ’¬ 2    πŸ“Œ 0
Post image

- Vance emphasized American interests and positioned himself as defending diplomacy.

02.03.2025 11:19 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

- Zelensky focused on historical context ("2014," "occupied") and used "we" to suggest collective identity "We are staying in our country, staying strong."

02.03.2025 11:19 β€” πŸ‘ 0    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

- Trump dominated the conversation (48.1% of total words) and used "you" 28 times, creating a confrontational tone "You don't have the cards. You're buried there."

- Trump's language was heavy on negation and control ("don't," "not," "no").

02.03.2025 11:19 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

I analyzed the Trump-Zelensky-Vance Transcript from @apnews.com, to see what stands behind their words - speaking volumes about power dynamics. πŸ“Š

Key Insights in thread

02.03.2025 11:19 β€” πŸ‘ 6    πŸ” 4    πŸ’¬ 4    πŸ“Œ 0

Very interesting topic, will give it a read next year :)

31.12.2024 12:56 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Great approach. A good feedback signal is worth a lot in any of these complex agentic workflows

10.12.2024 11:16 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Paper page - Agent-as-a-Judge: Evaluate Agents with Agents Join the discussion on this paper page

What is better than an LLM as a Judge? Right, an Agent as a Judge! Meta created an Agent-as-a-Judge to evaluate code agents to enable intermediate feedback alongside DevAI a new benchmark of 55 realistic development tasks.

Paper: huggingface.co/papers/2410....

10.12.2024 09:53 β€” πŸ‘ 27    πŸ” 1    πŸ’¬ 2    πŸ“Œ 1
Post image

There is ZERO connection between vaccines and autism!

NONE!

10.12.2024 03:00 β€” πŸ‘ 18619    πŸ” 3803    πŸ’¬ 565    πŸ“Œ 130
Video thumbnail

Why is the government not capable wiping the floor with Nigel Farage every single day like this? Alastair Campbell effortlessly destroys him over his Brexit legacy which by almost every economic and financial measure has been disastrous for the country

06.12.2024 06:41 β€” πŸ‘ 9506    πŸ” 2374    πŸ’¬ 560    πŸ“Œ 166
Post image

OpenAIs o1-pro model is only marginally better than the o1 model. But 10x the price ($20 to $200) to get access, not sure that's worth the qual increase.

Alternatively, do your own agentic or chain of thought prompting. Or go full professional with #dspy or #adalflow for auto prompt optimization.

06.12.2024 09:43 β€” πŸ‘ 18    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1
Post image

When predicting human sentiment responses for (german) words, Qwen 2.5 Coder 32B is doing worse than gpt4oMini.

R squared is
0.845 for GPT4oMini
0.714 for Qwen2.4 Coder 32B

Many more comparisons to follow :)

Do you know a good LLM for European Languages?

02.12.2024 10:26 β€” πŸ‘ 12    πŸ” 0    πŸ’¬ 3    πŸ“Œ 0
Post image

Using Perplexity instead of prompting to predict the more likely outcome of neuroscience experiments. Beating human experts by roughly 20% - wow.

Could help to make research more efficient Imho. Especially since they've older models to achieve this!

www.nature.com/articles/s41...

30.11.2024 09:17 β€” πŸ‘ 9    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Menlo Ventures shows that enterprise usage of anthropic has doubled in the last year.

Could it be because it's often better than openai? - for coding it's my go to model.

30.11.2024 08:40 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

A year ago, Claude 2.1 had a 200K token context window with 27% accuracy on needle-in-haystack tests. Now, Alibaba's Qwen2.5-Turbo claims a 1M token window with 100% recall on similar tests.
@alanthompson.net

30.11.2024 08:33 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Koala

30.11.2024 08:24 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Test-Time Training on Nearest Neighbors for Large Language Models Many recent efforts augment language models with retrieval, by adding retrieved data to the input context. For this approach to succeed, the retrieved data must be added at both training and test time...

SIFT introduces test-time learning by selecting highly relevant and non-redundant data for fine-tuning. Some call it Q-star 2.0. It optimizes information gain, reducing prediction uncertainty while using minimal compute - neat.

paper: arxiv.org/abs/2305.18466
Video: www.youtube.com/watch?v=vei7...

30.11.2024 07:53 β€” πŸ‘ 4    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Post image

Performant and actually open LLM was just released. (with training data and recipes etc.) will be testing OLMO soon. allenai.org/olmo

30.11.2024 07:50 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Been trying that (with contemporary people) and it works. You can test it's accuracy of you have responses from your copied subjects. For historical figures, that could be decisions or text they have written.
Not saying it'll work, but at least you could estimate whether it does or doesn't.

30.11.2024 07:20 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Are instruct models becoming "too obedient"? When I prompt them with brand guidelines, and leave certain phrasen in, the llm will try to put those into virtually every content :D
Prompt engineering can mitigate, but its certainly different from the DaVinci days (gpt3)...old school completion.

30.11.2024 07:16 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Happy Thanksgiving, everyone!

While you are drinking wine and drowning in turkey, I would encourage folks to think about:

What if Bluesky really is our shot at breaking the web2/socmedia stranglehold on collective sensemaking...

What kinds of trillionaire and sovereign attacks are coming?

28.11.2024 14:36 β€” πŸ‘ 89    πŸ” 10    πŸ’¬ 11    πŸ“Œ 1
Post image

πŸš€ Just ran a simulation on 1,000 word ratings using human data + our improved LLM prompt for simulating responses. Results? We explain almost 80% of the variance! This approach could revolutionize how we model human emotionsβ€”customized for specific groups. πŸ”₯ #AI #EmotionModeling

28.11.2024 12:08 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

It's great, but we find the setup to be great too.. :)
Will try textgrad alone as well to compare.
How Creating good eval prompts is still a part I don't quite get.

23.11.2024 11:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Try adalflow, it combines dspy and textgrad... But it's a bit more complex to setup.

21.11.2024 21:33 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@jonathanmall is following 20 prominent accounts