Calvin McCarter's Avatar

Calvin McCarter

@calvinmccarter.bsky.social

calvinmccarter.com

217 Followers  |  454 Following  |  60 Posts  |  Joined: 15.11.2024  |  2.5086

Latest posts by calvinmccarter.bsky.social on Bluesky

reporting for duty here on bluesky to fight for anti-blueskyism. (i'm not fighting against blueskyism because i'm not a hater like that.)

18.11.2025 13:47 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

i first read that as:

"power washing" but for the feet. idea to fill out later

03.10.2025 01:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Slightly off-topic, but one of my biggest pet peeves is when people say about baby sadness, "Don't worry: they won't remember this." In the long run, we are all memoryless!

29.09.2025 00:36 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
The idiosyncrasies of tabular data The things that make tabular data different

What makes tabular data unique (and interesting!) is not merely that it's arranged into rows and columns. New blogpost: calvinmccarter.substack.com/p/the-idiosy...

19.08.2025 03:06 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

yea

01.08.2025 01:22 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

what are the reasons for this from your perspective?

26.07.2025 19:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

maybe not more expensive to manufacture in an absolute sense. but if chinese internal demand rises, and it no longer needs us exports, then china may no longer see the value in exporting to the us. in which case, the cost of consumer goods *in the us* would rise.

27.06.2025 16:14 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

also, immigrants will also do less of our low-wage / low-status work for us (declining LatAm TFR). meanwhile, AI will (initially) compete with high-wage / high-status workers. so we're still looking at a combustible political situation.

27.06.2025 15:33 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

tbf, while AI is going to start doing more of our work for us, East Asia will start doing less of our work for us (their trade surpluses will decline due to their aging demographics and the end of the USD as reserve currency)

27.06.2025 15:31 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

In principle I agree, but federal lands are relatively poorly managed wrt fire prevention and management. Obviously it would be preferable to just fix that, but I'm not sure whether that's realistic.

20.06.2025 00:57 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Related to your earlier assessment of the probability of the US striking Iran, how would you assess the probability of an Israeli strike? Is it harder to predict, because Israel doesn't need to bring assets into the region?

12.06.2025 02:36 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

"autoregression for training, diffusion for inference"

21.05.2025 20:38 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

alphaxiv does this for all arxiv papers -- just s/ar/alpha/ in the url -- and i've been told that it's coming soon to {bio,chem,med}rxiv as well.

18.03.2025 00:44 β€” πŸ‘ 4    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

I don't disagree with you exactly, but if an institution's natural defenders are AWOL (or even siding with its enemies) due to a litany of grievances, then it's probably already too late to save that institution.

16.03.2025 21:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(not that there's anything wrong with that)

16.03.2025 17:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I just delete my old repo, then give myself a new username...

11.03.2025 18:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Flipping the switch on far-UVC We’ve known about far-UVC’s promise for a decade. Why isn't it everywhere?

Has anyone tried far-UVC in their home? It's now dropped into the ~$300 price range where I'm interested in trying it for myself. substack.com/home/post/p-...

11.03.2025 14:43 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

it's definitely a Michigan thing

12.02.2025 04:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I am slightly cynical about the Clean Label Project, given that it seems to be "pay to play". Also, afaict, it was started by the founder of ByHeart formula, and theirs was like the first thing to get certification. Not that this is necessarily bad -- ByHeart and Kendamil are the best formula IMO.

12.01.2025 23:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

the website has a list: cleanlabelproject.org/product-cate... which is specific and helpful, though oddly it slightly differs from the brands in the report:

12.01.2025 23:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Here's a link to the report: cleanlabelproject.org/wp-content/u... (TLDR heuristics: whey is better than plant-based, non-organic is better than organic, unflavored is better than chocolate-flavored)

12.01.2025 20:15 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Ultra exciting! And it's gratifying to see that this method uses the kernel density integral preprocessing method that I published in @tmlr-pub.bsky.social (2023). (One takeaway: even if your ML research focus isn't deep learning, pursue directions that complement rather than compete with it.)

08.01.2025 19:03 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

i cooked him

23.12.2024 03:26 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model We show that BERT (Devlin et al., 2018) is a Markov random field language model. This formulation gives way to a natural procedure to sample sentences from BERT. We generate from BERT and find that it...

There are ways around this, eg arxiv.org/abs/1902.04094, but I think uniform rate MLMs are the cleaner approach. Plus, as shown in Eq 5 of the Factorization Curse paper, it's what you get after marginalizing over all XLNet-style autoregressive permutations.

20.12.2024 19:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This creates a train-inference gap if one is training at any fixed masking rate. Of course, with MLMs people don't even try that, and instead use pseudo-likelihood as an approximation for likelihood. Besides being approximate, the problem with this is that this takes L separate forward-passes.

20.12.2024 19:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

When one evaluates log-likelihood of a sequence of length L via the chain rule of probability, the first term has missingness fraction of 1, the second has missingness of (L-1)/L, etc. So the inference-time masking rate is ~ Uniform[0, 1].

20.12.2024 19:52 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0

The other advantage of uniform rate is that your MLM would now be able to provide likelihoods not just pseudo-likelihoods (unlike standard MLMs) and would be able to do so for any generation order (unlike causal LMs).

19.12.2024 22:59 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
The Factorization Curse: Which Tokens You Predict Underlie the Reversal Curse and More Today's best language models still struggle with hallucinations: factually incorrect generations, which impede their ability to reliably retrieve information seen during training. The reversal curse, ...

Besides trying a fixed 30% masking rate, did you also consider using random ~ Uniform[0, 1] rate? This is akin to XLNet and order-agnostic autoregression, and recent work from Meta also found that it improves retrieval performance: arxiv.org/abs/2406.05183

19.12.2024 21:43 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

How do you go from a hit in your antibody screen to a suitable drug? Now introducing CloneBO: we optimize antibodies in the lab by teaching a generative model how we optimize them in our bodies!
w/ Nat Gruver, Yilun Kuang, Lily Li, @andrewgwils.bsky.social and the team at Big Hat! 1/7

17.12.2024 16:01 β€” πŸ‘ 11    πŸ” 1    πŸ’¬ 1    πŸ“Œ 1

this treaty already exists (Prometheus was punished by the alien-gods for uplifting hominids)

15.12.2024 01:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@calvinmccarter is following 20 prominent accounts