Eric Martinez's Avatar

Eric Martinez

@ericmtztx.bsky.social

UTRGV School of Medicine / UTHealth RGV Asst Dir Business Intelligence & Enterprise Engineering Adjunct Lecturer @ UTRGV CS Jiu-Jitsu Black Belt

87 Followers  |  163 Following  |  14 Posts  |  Joined: 23.11.2024  |  1.6943

Latest posts by ericmtztx.bsky.social on Bluesky

deciding between "actually taking some time off" and "work on personal projects and call it relaxing"

deciding between "actually taking some time off" and "work on personal projects and call it relaxing"

hmm i feel attacked

24.12.2024 20:27 โ€” ๐Ÿ‘ 96    ๐Ÿ” 12    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 0

Your 'AI Engineer' can't deploy their own model? Your 'Data Engineer' can't write a web app? Stop with narrow job titles that create silos. In healthcare tech, we need full-stack problem solvers who can ship end-to-end. One capable generalist > five specialists.

30.11.2024 19:27 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Why Isnโ€™t Healthcare More Personalized? Currently, inefficiencies in healthcare practices can frustrate patients and lead to unnecessary pre-surgical tests or boilerplate after-visit instructions, among other issues. Personalized health car...

Healthcare (system) is rarely personalised. In this HBR article the authors predict that *eventually healthcare will be as personalised as online shopping or banking. See poll below to weigh in.
#healthcare #personalisation #pt

hbr.org/2024/11/why-...

28.11.2024 12:26 โ€” ๐Ÿ‘ 4    ๐Ÿ” 3    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

๐Ÿงต Following our healthcare AI journey? I'll share:

- Implementation & design details
- Evaluation methodology and dataset development
- AI engineering process
- Data platform infrastructure/ HIPAA compliance

Learn how we're making healthcare more accessible in the RGV, one search at a time. (7/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Step 5: Address Bias Head-On
We're expanding our evaluation dataset to detect unwanted outcomes that could impact patients, providers, or departments. No system is perfect, but measuring bias lets us iterate and improve responsibly. (6/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Step 4: Systematic Development

- Build comprehensive evaluation dataset
- Perfect basic provider matching
- Enable structured query rewriting with LLMs

Each step verified through testing to ensure reliable results. (5/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Step 3: Measure What Matters
Key metrics we're tracking:

- Query-to-(provider/specialty/department) match accuracy
- Cross-lingual search performance
- Response time (targeting <300ms)

These guide our improvements where they matter most and can be sliced across descriptors to measure bias. (4/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Step 2: Fast Iteration Cycles
We built an evaluation pipeline measuring precision and other key metrics. This enables rapid testing of different embedding models, scoring methods, and chunking strategies - focusing on getting provider retrieval right before adding complexity. (3/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Step 1: Real Data, Synthetic Queries
We extract actual patient-provider matches from our EHR, then use LLMs to generate natural search queries. This creates a robust test dataset that reflects real patient needs - from "my toe hurts" to "doctor for diabetes control" (2/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We're building an AI-powered search engine at UTHealth RGV that connects patients with the right healthcare providers. Our focus: making it easier for our community to find and access the care they need. Here's how we're building responsibly: (1/7)

28.11.2024 18:06 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Itโ€™s almost a game for people to race to take the latest open weight model and Med* it. But anyone who has tried them can tell you that they donโ€™t seem very good and their performance on anything real world is still far below state of the art closed source models.

27.11.2024 05:31 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I do think small task oriented models trained on real world data could be very useful (speed, cost, task performance). But skeptical that throwing PubMed articles and things of that sort will improve performance on real world clinical tasks, especially without significant curation and synthesis.

27.11.2024 05:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Great paper. Really skeptical of the fine-tuning techniques and training data being used in many of these Med* models. Not much effort seems to be going into building legit high quality datasets for clinical tasks. I yawn at yet another Med* model evalโ€™d on low quality QA datasets.

27.11.2024 05:22 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Assuming biased data means biased models is lazy.

Measure bias by quantifying disparities in your modelโ€™s behavior.

Hereโ€™s how:

1. Define your AI's tasks and metrics

2. Measure performance across critical subgroups

3. Iterate based on real results

Act like a scientistโ€”test, measure, repeat.

26.11.2024 06:38 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Iโ€™d be curious to see if this phenomenon occurs with 1) larger variants of the same model teaching smaller models 2) The effect of prompting to curate high quality data than the original. It would be interesting to see how long it takes for collapse to outweigh the benefits.

23.11.2024 17:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Neat RAG library we developed for our in-house Ruby on Rails applications at UTRGV School of Medicine. Weโ€™ll be using this to power our upcoming โ€œFind a Doctorโ€ tool. No dependencies other than pgvector and Azure OpenAI.

23.11.2024 07:35 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@ericmtztx is following 19 prominent accounts