Ah, havenโt mastered the art of a sunny side up yet I see
04.08.2025 19:19 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0@aidangcw.bsky.social
Psychopathology | Personality | Quant Methods Professor of Psychology and Psychiatry | Eisenberg Family Depression Center | University of Michigan Editor | Journal of Psychopathology and Clinical Science Founder | www.smart-workshops.com
Ah, havenโt mastered the art of a sunny side up yet I see
04.08.2025 19:19 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Has she tried the madame version?
04.08.2025 13:13 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0Tempting, isnโt it?
04.08.2025 13:08 โ ๐ 5 ๐ 0 ๐ฌ 0 ๐ 0The timing on this post from Twitterโฆ
03.08.2025 01:08 โ ๐ 7 ๐ 0 ๐ฌ 0 ๐ 0This is peak lakehouse
03.08.2025 00:11 โ ๐ 7 ๐ 0 ๐ฌ 0 ๐ 01:1 correspondence seems deeply unlikely. I really love this brief editorial that talks about it in terms of description and mechanism.
To use their eg of a candle and light, thereโs no obvious link between the chemical combustion and the concept of illumination.
journals.asm.org/doi/full/10....
If I understand you correctly I totally agree. This absolutely should be hard. Which is why I think itโs so absurd when people act like it isnโt.
30.07.2025 21:42 โ ๐ 2 ๐ 0 ๐ฌ 2 ๐ 0Interesting way to say it. I guess thatโs probably how they think about it, as assumed. I think the best version of RDoC would be designed at adressing the multilevel/multimodal coherence issue instead of being a neural reductionism model
30.07.2025 20:45 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0These are good examples that illustrate the issue.
Though another way to frame it is who was audacious (hubristic?) enough to think they could capture a psychological construct with peripheral psychophys?
I still see this in the logic of some research studies.
For sure. I canโt think of a more important issue.
30.07.2025 17:15 โ ๐ 5 ๐ 1 ๐ฌ 1 ๐ 0This is the question of our times. Though I think answers are likely to be at least somewhat concept specific
30.07.2025 16:59 โ ๐ 13 ๐ 0 ๐ฌ 1 ๐ 0I was thinking the other day about how much peopleโs ignorance of survivor bias influences their perspectives (including politics) much to everyoneโs detriment
30.07.2025 16:57 โ ๐ 3 ๐ 0 ๐ฌ 0 ๐ 0IIRC the average was ~100 words/ minute, which is what you expect when reading out loud, but there was real variability here.
Would be happy to chat more if you're interested in pursuing collaboratively, but if not we include our prompts in the manuscript, etc.
I don't have a great sense for the exact minimum. No doubt it will depend on eliciting prompt, construct, and model. But we do see that absolute error is modestly associated with length and at lower number of diaries/minutes (2 samples) we see things stabilize at some point.
27.07.2025 16:50 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0You could also create multi informant latent variables, with either self and LLM or multiple LLMs. Trait scores are highly correlated across models but not perfectly so.
27.07.2025 15:46 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0For sure. A limitation of the data we had available for this study is that we lacked non-self-report outcomes. So, due to method overlap self-report almost always would โwinโ. To show incremental validity ideally weโd try to predict unprivileged outcomes.
27.07.2025 15:45 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0Itโs already longer than we were shooting for at first. It would be good to add that though on revision or rejection.
27.07.2025 12:53 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0I think this is fair criticism and easy to add. True story. We had that as a focus early on when we first did this with just ChatGPT. Then we thought it would be better not as an ad for openAI and went for many models. Then the output started ballooning and it sort of got unwieldy and lost track
27.07.2025 12:53 โ ๐ 3 ๐ 1 ๐ฌ 1 ๐ 0Ha
27.07.2025 12:19 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0To say nothing of the sand orcs
27.07.2025 12:08 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0I draw this distinction because I think there are other ways to set this sort of thing up where you train this or another model up to specifically predict self report. Here we ask the LLM to rate personality based on the text we give it. We discuss in context of self-other agreement literature
27.07.2025 12:07 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0Thanks. A nuance, perhaps, but we donโt treat self-reports as โtrueโ in an absolute sense. The personality lit has dispelled the notion of self-as-true a long time ago. We think of this as a different rater based on different modality of data. I think this nuance is critical, others might disagree.
27.07.2025 12:07 โ ๐ 6 ๐ 1 ๐ฌ 2 ๐ 0There is also potential for getting info that isnโt easily captured by questionnaire. Here we use the big5 as the target because we had self-reports to compare but also because the info contained is so so broad. Iโm pretty convinced we can start extrapolating.
27.07.2025 12:02 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0I particularly think this holds potential for taking narrative summaries collected in real- or near-time to events (as we do in Study 2) before they are filtered through days of interpretation and meaning making by folks. Also that is a context where we canโt give long lists of questionnaire items.
27.07.2025 12:02 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0If I may, this is sort of a double barreled question. 1. I think we should typically trust our patients. At least that their reports reflect their experiences accurately. 2. The value of this method isnโt in getting more trustworthy info, but in aggregating and distilling info from other data.
27.07.2025 12:02 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0It is wild. Thatโs an apt description. When we started doing this I almost fell out of my chair I was so surprised.
27.07.2025 11:53 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Thanks for engaging, Michael! Great idea. As you noted, one of the big takeaways/demonstrations here was how much this interface can democratize the use of LLMs. But also weโre very interested in understanding this all better and figuring out what works best. Let me chat with team and get back to u
27.07.2025 11:51 โ ๐ 1 ๐ 1 ๐ฌ 0 ๐ 0So, ironically perhaps, the spell check AI didnโt catch the error in the figure caption. As much as I wish we had, we did not have โsand elf-reportsโ in this study. As you all know, sand elves are impossible to work with. It was just human self-report.
H/t @gregdepow.bsky.social for catching this
The robots know us so well.
We show pretrained LLM ratings of personality from brief snippets of natural language achieve agreement with self-reports that are comparable to family and friends.