A city of drivers expands, relentlessly and inevitably, into sprawl.
A city of pedestrians, cyclists, and transit-users settles, gradually and elegantly, into itself.
@ysawej.bsky.social
Deep Learning, NLP, multimodal models and all things neural network. Also, math guy and computer scientist.
A city of drivers expands, relentlessly and inevitably, into sprawl.
A city of pedestrians, cyclists, and transit-users settles, gradually and elegantly, into itself.
Still the best two sentences on climate action.
29.12.2024 17:42 โ ๐ 7591 ๐ 1565 ๐ฌ 50 ๐ 62An illustrated guide to never learning anything
25.12.2024 00:26 โ ๐ 146 ๐ 20 ๐ฌ 6 ๐ 3โThe best part of waking up, is your house being quiet and your kids not tearing shit upโ
29.11.2024 23:22 โ ๐ 7 ๐ 1 ๐ฌ 1 ๐ 0The next coffee drinker whom awakens, understands the joy, of a hot pot ready and waiting.
29.11.2024 11:34 โ ๐ 11 ๐ 1 ๐ฌ 0 ๐ 0There is no greater joy than everyone in the house being asleep as you drink your morning coffee.
29.11.2024 11:00 โ ๐ 7305 ๐ 442 ๐ฌ 304 ๐ 85Code written with box characters used on old old software to make fake UIs
Youโre still arguing about tabs vs. spaces? May I presentโฆ
25.12.2024 18:37 โ ๐ 5327 ๐ 1293 ๐ฌ 157 ๐ 149I didn't embark in machine learning thinking of it as an ideological project to disenfranchise human beings.
But we need to face reality, machine learning can easily become the driver of such change, shifting power structures.
We, actors of tech, can modulate this effect.
1/3
๐
21.12.2024 01:42 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Cartoon by singer sarcastically showing that people wrongly perceive car dependent infrastructure as โpublic investmentโ (it isnโt) and investment in public transit as a โwasteful subsidyโ (it isnโt, it has an excellent return on investment and actually saves public money).
"Public investment" vs "Wasteful Subsidy." The only problem with this clever Singer cartoon is that some people might actually not get that itโs sarcastically illustrating the perception problem, NOT telling the truth. Just in case it actually needs to be said, the truth is the opposite.
18.12.2024 08:21 โ ๐ 7019 ๐ 923 ๐ฌ 127 ๐ 33Dynamic programming alternatives to dynamic programming for optimal control. Replace your Bellman equation with backpropagation. www.argmin.net/p/twirling-t...
21.11.2024 15:36 โ ๐ 41 ๐ 3 ๐ฌ 2 ๐ 0Bookmark
16.12.2024 17:08 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Bookmark
16.12.2024 06:33 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Bookmark
16.12.2024 06:31 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0boundary2 just published a forum on the gordian knot of finance, where @stefeich.bsky.social, @aminsamman.bsky.social, @thisblue.bsky.social, Janet Roitman, Dick Bryan, and myself reflect on the infuriating hold of finance on economic policy (and how to break it)
www.boundary2.org/the-gordian-...
The tech bro fascination with eugenics is so on brand. The idea that you could make accurate, actionable predictions from individual genotypes (with all their complex, non-linear interactions) from averaged, linearly modeled population-level genomic statistics is just another big data fantasy
12.12.2024 16:13 โ ๐ 3070 ๐ 483 ๐ฌ 130 ๐ 101jacob sansbury @jsnnsa having kids in the next 5 years might be a tragic mistake every smart bio founder/scientist iโve talked to seems to think embryo editing for things like short sleeper, reduced cancer risk, etc is possible on a near term horizon imagine having two kids a couple years apart, one is a super human and the other isnโt โsorry jim, little timmy wonโt get cancer, only needs 4 hours of sleep and youโre normal you were just born in the wrong order ๐คท๐ป โ
Tech bros know what every parent wants: children who sleep less
11.12.2024 14:39 โ ๐ 8650 ๐ 1211 ๐ฌ 515 ๐ 720noice!
09.12.2024 10:49 โ ๐ 14 ๐ 5 ๐ฌ 0 ๐ 0Insane lack of transparency from OpenAI. Saying the models people use are different than the "final" evals they released. Do better. Especially enterprises will move elsewhere because of this stuff.
06.12.2024 22:00 โ ๐ 56 ๐ 5 ๐ฌ 5 ๐ 0Zuck is developing 2GW+ data center.
"Last big AI update of the year:
โขโ โ Meta AI now has nearly 600M monthly actives
โขโ โ Releasing Llama 3.3 70B text model that performs similarly to our 405B
โขโ โ Building 2GW+ data center to train future Llama models
Next stop: Llama 4. Let's go!"
Old quant types (some base model types require these): - Q4_0: small, very high quality loss - legacy, prefer using Q3_K_M - Q4_1: small, substantial quality loss - legacy, prefer using Q3_K_L - Q5_0: medium, balanced quality - legacy, prefer using Q4_K_M - Q5_1: medium, low quality loss - legacy, prefer using Q5_K_M New quant types (recommended): - Q2_K: smallest, extreme quality loss - not recommended - Q3_K: alias for Q3_K_M - Q3_K_S: very small, very high quality loss - Q3_K_M: very small, very high quality loss - Q3_K_L: small, substantial quality loss - Q4_K: alias for Q4_K_M - Q4_K_S: small, significant quality loss - Q4_K_M: medium, balanced quality - recommended - Q5_K: alias for Q5_K_M - Q5_K_S: large, low quality loss - recommended - Q5_K_M: large, very low quality loss - recommended - Q6_K: very large, extremely low quality loss - Q8_0: very large, extremely low quality loss - not recommended - F16: extremely large, virtually no quality loss - not recommended - F32: absolutely huge, lossless - not recommended
Learning about quantization suffixes while `ollama pull llama3.3` download completes (fyi, quantization for the default 70b is q4_K_M)
โข make-ggml .py: github.com/ggerganov/ll...
โข pull request: github.com/ggerganov/ll...
Bookmark
06.12.2024 02:49 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Will be at #NeurIPS2024 Dec 10-13. Looking forward to run into everyone in #AI all at once๐ค The 2019 Vancouver one was the largest conf I ever attended--not sure how they plan to cram even more this time.. ๐ฑ
[Will be at our poster on 12/11 morning openreview.net/forum?id=kPB... ]
๐ฃI'm hiring PhD interns for combined theory+empirical projects in: exploration in post-training, multi-task learning in autoregressive models, distillation, reasoning beyond CoT.
Apply on the link below. If you're at #NeurIPS2024, message me to chat.
jobs.careers.microsoft.com/global/en/jo...
The paper proposes a single-channel Deep Cascade Fusion of Diarization and Separation (DCF-DS) framework for back-end speech recognition, achieving first place in the realistic single-channel track of the CHiME-8 NOTSOFAR-1 challenge.
12.11.2024 10:03 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0๐ Summary:
The project is about Speaker Diarization using OpenAI Whisper, combining ASR capabilities with VAD and Speaker Embedding to identify speakers in transcriptions. The pipeline corrects timestamps, extracts vocals, and aligns timestamps for accurate speaker identification. (1/3)
I'm excited to share that we've released v1.0 of our podcast corpus, SPoRC, led by my PhD student Ben Litterer! This first dataset is a slice of time, comprising over one million episodes from May and June 2020, including transcripts, diarization, and extracted audio features.
15.11.2024 15:03 โ ๐ 52 ๐ 16 ๐ฌ 1 ๐ 4Thrilled to share our NeurIPS spotlight on uncertainty disentanglement! โจ We study how well existing methods disentangle different sources of uncertainty, like epistemic and aleatoric. While all tested methods fail at this task, there are promising avenues ahead. ๐งต ๐ 1/7
๐: arxiv.org/abs/2402.19460
It wasn't even the transformer paper that was first to show attention was all you need. Everyone forgets how aggressively folks were working on faster alternatives to RNNs ~2016, and another paper from Google did a pure attention model first: arxiv.org/abs/1606.01933
02.12.2024 03:48 โ ๐ 23 ๐ 3 ๐ฌ 3 ๐ 1