Michael Dinitz's Avatar

Michael Dinitz

@mdinitz.bsky.social

Associate Professor, Department of Computer Science, Johns Hopkins University. https://www.cs.jhu.edu/~mdinitz/

480 Followers  |  119 Following  |  31 Posts  |  Joined: 11.11.2023  |  1.9528

Latest posts by mdinitz.bsky.social on Bluesky

And if you're faculty: I beg you, for the sake of everyone else in the world, strongly consider creating your own public webpages for your course instead of having everything locked up in Canvas.

09.10.2025 00:04 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 1    πŸ“Œ 1

Congrats!

27.08.2025 22:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

First day of class for the semester. Told my 2.5-year-old that "Daddy is going to be a teacher today, just like your teachers!". Response: "Daddy you're so silly". Apparently I lack the gravitas of a daycare teacher.

26.08.2025 15:17 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Glad you liked our lightness paper! I'm pretty excited about that direction.

25.08.2025 13:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Incredibly well deserved!!

31.05.2025 08:53 β€” πŸ‘ 6    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Awesome, congrats!!

18.02.2025 23:45 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Escaping Collapse: The Strength of Weak Data for Large Language Model Training Synthetically-generated data plays an increasingly larger role in training large language models. However, while synthetic data has been found to be useful, studies have also shown that without proper...

Synthetic Data is all the rage in LLM training, but why does it work? In arxiv.org/abs/2502.08924 we show how to analyze this question through the lens of boosting. Unlike boosting, however, our assumptions on the data and the learning method are inverted.

14.02.2025 13:48 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

A proof is a logical argument written to convince a skeptical audience. A corollary is that the best way to read a proof is to roleplay as a skeptical audience.

06.02.2025 14:33 β€” πŸ‘ 10    πŸ” 2    πŸ’¬ 2    πŸ“Œ 1

I think there’s a lot to be said for going all in on something. It was neat being at Google and doing everything on Google systems. On the other hand, faculty autonomy is one of the nice things about academia compared to industry.

26.01.2025 19:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

We used zoom for lectures, canvas for communicating with students and organizing classes. We have a zoom license but also we’re a Microsoft campus. So classes are zoom & canvas, official stuff with admin is on teams, and most departments have an internal slack.

26.01.2025 19:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Yeah, for cross-department you’re stuck with whatever the university is set up on. We’re also on teams, but no professors use it, so we all just use email still.

26.01.2025 18:20 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Might be even more of a long shot in math, but I think that Zulip is actually better than slack, and it’s free for academics. I use it internally in my research group. I know the category theorists like it, so maybe that will convince other math people?

26.01.2025 18:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Google meet is surprisingly good too, but Teams-only is crazy. Strongly recommend getting your department to use slack if possible. We just use the free version, and even that is great. Aside from messages, having a #teaching channel and an #advising channel to ask questions is super useful.

26.01.2025 17:22 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We have a department slack that is very active for faculty. That’s now how I mainly communicate with other faculty in my department. Across departments it’s still email, though - no one I know is willing to use teams.

26.01.2025 17:10 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I may or may not have some Hagoromo chalk in my office, if you want to make the best use of those blackboards :)

20.01.2025 18:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Interviewer: can you explain this gap in your resume?

Networking researcher: UDP

20.12.2024 23:07 β€” πŸ‘ 12    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Preview
Breaking the Barrier: A Polynomial-Time Polylogarithmic Approximation for Directed Steiner Tree The Directed Steiner Tree (DST) problem is defined on a directed graph $G=(V,E)$, where we are given a designated root vertex $r$ and a set of $k$ terminals $K \subseteq V \setminus {r}$. The goal is ...

Just saw a new result on the arxiv: polylog approximation for directed steiner tree! arxiv.org/abs/2412.10744 . Amazing breakthrough in a fundamental approximation algorithms problem. Congrats Bundit!

17.12.2024 14:59 β€” πŸ‘ 14    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Complications: The Ethics of the Killing of a Health Insurance CEO (guest post) - Daily Nous The killing of UnitedHealthcare CEO Brian Thompson earlier this month has ignited moral debate around the world. Many have condemned the killing as they would any other murder. Others, though, have th...

A statistic I've seen repeated in the discussion of health insurance is that UHC used an "AI with a 90% error rate" to deny claims. Recently I saw it repeated twice here: dailynous.com/2024/12/15/c... in which a moral philosopher struggles to figure out if murdering health insurance CEOs is moral.

15.12.2024 22:10 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Prompt: Draw a picture of an American family at Thanksgiving dinner.

03.12.2024 02:15 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 3    πŸ“Œ 0

Very interesting! Somehow I completely missed this self-improving literature, but it looks really neat. We're definitely *not* self-improving in this sense -- we're not "learning" the true distribution as we go and modifying our strategy, just continuing to use our fixed (but robust) search tree.

27.11.2024 02:30 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I have similar problems with activation energy :). Not sure if this helps or not, but maybe a point I should have advertised: the entire construction and proof is ~2 pages. The rest of the paper is lower bounds, experiments, discussion, etc.

26.11.2024 20:36 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I always love it when new settings (like algorithms with predictions) end up leading us back to classical problems. Hopefully there's more to be done on "robust" versions of classical optimal data structures!

26.11.2024 20:21 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

I think that this is super cool. Somehow, despite optimal BSTs being a super classical question (introduced by Knuth in the 70's and taught in undergrad algorithms), no one seems to have asked the question "what if our distribution is wrong?"

26.11.2024 20:21 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

And since H(p) is a known lower bound, our search trees are almost optimal when EMD(p,q) is reasonably small!

26.11.2024 20:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Our result can be interpreted as giving "robust" bounds for this problem: we build a search tree that has expected query complexity of O(H(p) + log(EMD(p,q))), even without any bound on EMD(p,q)!

26.11.2024 20:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

But what if we're given a probability distribution q which is *not* equal to the true distribution p? Then things can get quite bad if you use the classical DP algorithm for q (or other standard approaches).

26.11.2024 20:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We're given n items, and a probability distribution p over them. We want to build a binary search tree for the items with minimum expected query complexity for p. Classical result of Knuth is a dynamic programming algorithm for this problem, simple enough that I teach it in undergrad algorithms!

26.11.2024 20:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Optimal binary search tree - Wikipedia

But it turns out that there's a more interesting interpretation (imho). Think about a different problem: optimal binary search trees en.wikipedia.org/wiki/Optimal... .

26.11.2024 20:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We also have matching lower bounds, extensions to collections of distributions, and empirical tests, so check out the paper! Hopefully this leads to interesting followup work on distributional predictions.

26.11.2024 20:21 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@mdinitz is following 20 prominent accounts