Hye Sun Yun's Avatar

Hye Sun Yun

@hyesunyun.bsky.social

PhD candidate in CS at Northeastern University | NLP + HCI for health | she/her πŸƒβ€β™€οΈπŸ§…πŸŒˆ

119 Followers  |  383 Following  |  15 Posts  |  Joined: 02.12.2024  |  1.6937

Latest posts by hyesunyun.bsky.social on Bluesky

Preview
As AI expands into medicine, Northeastern study finds AI models influenced by medical biasΒ  - Khoury College of Computer Sciences Humans can be easily influenced by language that is one-sided, especially in complex fields like medicine. But a new Khoury-led study shows that large language models, too, can be tricked […]

Thrilled to share our research showing how LLM models can be influenced by bias from "spun" medical literature is now featured in Northeastern's Khoury news! This shows critical insights as AI enters healthcare.
The full paper can be found at arxiv.org/abs/2502.07963

25.08.2025 15:36 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

I am at CHIL this week to present my poster (Caught in the Web of Words: Do LLMs Fall for Spin in Medical Literature?) on Thursday June 26.

Looking forward to connecting and sharing our work on spin with the CHIL community!

26.06.2025 00:52 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

I am at CHI this week to present my poster (Framing Health Information: The Impact of Search Methods and Source Types on User Trust and Satisfaction in the Age of LLMs) on Wednesday April 30

CHI Program Link: programs.sigchi.org/chi/2025/pro...

Looking forward to connecting with you all!

29.04.2025 00:50 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Online Health Information–Seeking in the Era of Large Language Models: Cross-Sectional Web-Based Survey Study Background: As large language model (LLM)–based chatbots such as ChatGPT (OpenAI) grow in popularity, it is essential to understand their role in delivering online health information compared to other...

LLM-based chatbots are changing how people search for health informationβ€”but how do users perceive their quality and trustworthiness compared to other online sources? Our survey study explores these questions. Check it out! www.jmir.org/2025/1/e68560

01.04.2025 01:06 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Oxford Word of the Year 2024 - Oxford University Press The Oxford Word of the Year 2024 is 'brain rot'. Discover more about the winner, our shortlist, and 20 years of words that reflect the world.

I'm searching for some comp/ling experts to provide a precise definition of β€œslop” as it refers to text (see: corp.oup.com/word-of-the-...)

I put together a google form that should take no longer than 10 minutes to complete: forms.gle/oWxsCScW3dJU...
If you can help, I'd appreciate your input! πŸ™

10.03.2025 20:00 β€” πŸ‘ 10    πŸ” 8    πŸ’¬ 0    πŸ“Œ 0
Preview
Caught in the Web of Words: Do LLMs Fall for Spin in Medical Literature? Medical research faces well-documented challenges in translating novel treatments into clinical practice. Publishing incentives encourage researchers to present "positive" findings, even when empirica...

Huge thanks to my amazing co-authors!
Karen Y.C. Zhang, @ramezkouzy.bsky.social, @ijmarshall.bsky.social, @jessyjli.bsky.social, & @byron.bsky.social [7/7]

Check out our full findings here: arxiv.org/abs/2502.07963

15.02.2025 02:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Can we fix this? We tested zero-shot prompts to reduce LLMs' susceptibility to spin.
Good news: prompts that encouraged reasoning reduced their tendency to overstate trial results! πŸ› οΈ
Careful design is key to improving evidence synthesis for clinical decisions. [6/7]

15.02.2025 02:34 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

When we asked LLMs to simplify abstracts into plain language, they often propagated spin into their summaries. This means LLMs could unintentionally mislead patients and non-experts about the effectiveness of treatments. 😱 [5/7]

15.02.2025 02:34 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

We asked LLMs how favorably they perceived a treatment’s results (0-10 scale). Even though LLMs could detect spin, they were far more influenced by it than human experts.
Meaning: LLMs believed spun abstracts presented more favorable results! 😬 [4/7]

15.02.2025 02:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

When we prompted 22 LLMs to identify spin in medical abstracts, we found that they were moderately to strongly capable of detecting spin.
However, things got interesting when we asked LLMs to interpret the results… [3/7]
πŸ”½

15.02.2025 02:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

So what is spin?
Spin refers to reporting strategies that make experimental treatments appear more beneficial than they actually areβ€”often distracting from nonsignificant results.

Example:
❌ β€œThe treatment shows a promising trend toward significance…”
βœ… β€œNo significant difference was found.”
[2/7]

15.02.2025 02:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🚨 Do LLMs fall for spin in medical literature? πŸ€”

In our new preprint, we find that LLMs are susceptible to biased reporting of clinical treatment benefits in abstractsβ€”more so than human experts. πŸ“„πŸ” [1/7]

Full Paper: arxiv.org/abs/2502.07963

πŸ§΅πŸ‘‡

15.02.2025 02:34 β€” πŸ‘ 63    πŸ” 25    πŸ’¬ 3    πŸ“Œ 4

As someone interested in an academic position post-PhD, I found this post very helpful. Thank you for sharing your wisdom and advice.

24.01.2025 17:45 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Awesome! Thank you

07.01.2025 15:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Thank you!

07.01.2025 01:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The application form says it is no longer accepting responses. Is the application closed now?

06.01.2025 23:54 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@hyesunyun is following 19 prominent accounts