Ujan's Avatar

Ujan

@ujaan.bsky.social

Learning NLP. PhD student @UNSW Sydney

41 Followers  |  625 Following  |  6 Posts  |  Joined: 15.08.2024  |  1.78

Latest posts by ujaan.bsky.social on Bluesky

The problem with most machine-based random number generators is that they’re not TRULY random, so if you need genuine randomness it is sometimes necessary to link your code to an external random process like a physical noise source or the current rate of US tariffs on a given country.

09.04.2025 19:15 β€” πŸ‘ 19192    πŸ” 3653    πŸ’¬ 385    πŸ“Œ 245
Surveillance shows Tufts graduate student detained
YouTube video by WCVB Channel 5 Boston Surveillance shows Tufts graduate student detained

This is horrifying. She co-wrote an OpEd at her school, then was grabbed by masked people, put into an unmarked black van, is now detained in an ICE facility over 1000 miles away.
youtu.be/PuFIs7OkzYY
Trying to limit the topics I post about, but this is my country and this one shook me. =(

26.03.2025 23:05 β€” πŸ‘ 60    πŸ” 12    πŸ’¬ 4    πŸ“Œ 1
Post image Post image Post image

Major weaknesses raised by Reviewer 2

14.02.2025 19:14 β€” πŸ‘ 19    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0

any international/intergovernmental report on β€œAI safety” that fails to acknowledge/ground recommendation on the growing unprecedented corporate power and monopoly is not a friend of the public, civil society, and definitely of marginalised and impacted groups

03.02.2025 12:23 β€” πŸ‘ 103    πŸ” 19    πŸ’¬ 0    πŸ“Œ 0
Post image

In fact, I have coauthored a whole paper arguing that the networked nature of modern production can amplify the harm of a small or moderate shock...

way beyond what standard models predict.

bengolub.net/snff-2/

4/

02.02.2025 12:59 β€” πŸ‘ 217    πŸ” 25    πŸ’¬ 5    πŸ“Œ 1

Ep 3 was a banger

01.02.2025 06:22 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

It turns out that = trust me bro

29.01.2025 11:51 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Preview
Bridging the Human-AI Knowledge Gap: Concept Discovery and Transfer in AlphaZero Artificial Intelligence (AI) systems have made remarkable progress, attaining super-human performance across various domains. This presents us with an opportunity to further human knowledge and improv...

One of my grand interpretability goals is to improve human scientific understanding by analyzing scientific discovery models, but this is the most convincing case yet that we CAN learn from model interpretation: Chess grandmasters learned new play concepts from AlphaZero's internal representations.

27.01.2025 21:43 β€” πŸ‘ 108    πŸ” 22    πŸ’¬ 2    πŸ“Œ 1

I recently talked to someone doing a PhD in multivariate calculus.

Or, as they would call it, a grad student

26.01.2025 09:54 β€” πŸ‘ 49    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

When they introduced group projects in school they did not tell us that life would be a continuous string of group projects until we retire

25.01.2025 22:37 β€” πŸ‘ 27    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0
Email inbox subject line stating "Your application has moved to the next stage"

Email inbox subject line stating "Your application has moved to the next stage"

Text of the email, stating that the next stage it ... declined.

Text of the email, stating that the next stage it ... declined.

After several rounds, to the granting agency that sent me this email, with the following subject line:
"Your application has moved to the next stage"

Well done.

That is some Grade 'A' level trolling.

You got me.

24.01.2025 16:04 β€” πŸ‘ 51    πŸ” 3    πŸ’¬ 7    πŸ“Œ 2

β€œI don’t know when this will happen but probably in 10 years” is researcher-speak for β€œI have no clue”

24.01.2025 13:47 β€” πŸ‘ 26    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

I hate what deep learning has done to the word "inference". Now we are forced to write sentences in our papers like "approximate inference at inference time"β€”I'm all for nonsense Dr Seuss language, but not in a scientific paper!

23.01.2025 12:02 β€” πŸ‘ 37    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

Every time I go to a public library I begin to believe again that we can create amazing things that are just there for all to enjoy and use for free.

03.01.2025 20:42 β€” πŸ‘ 5697    πŸ” 723    πŸ’¬ 33    πŸ“Œ 41

If you have ever wondered what colonization was likeβ€”the violence, the dispossession, the massacres, the genocide, the displacement, the famine, the dehumanization, the racismβ€”you are watching it in real time in Palestine. It is among the most depraved of all possible acts.

25.12.2024 09:34 β€” πŸ‘ 537    πŸ” 162    πŸ’¬ 6    πŸ“Œ 5
Post image

With the new OpenAI O3 moving performance from 5% up to 25% on FrontierMath it’s time to push open-source models upwards!

We're super happy to release FineMath, the best open math dataset yet. A strong baseline to start training your own models

Find it in the trending section of HuggingFace ;)

23.12.2024 13:20 β€” πŸ‘ 61    πŸ” 9    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

"a blue ball is on a wooden table, perfectly motionless. There is no motion in the scene. The camera does not move or pan in any way."

17.12.2024 13:27 β€” πŸ‘ 59    πŸ” 10    πŸ’¬ 5    πŸ“Œ 3

ethernet? you mean the wifi cable

09.12.2024 17:01 β€” πŸ‘ 943    πŸ” 141    πŸ’¬ 37    πŸ“Œ 18

AI influencers "this <big company name / Stanford / Berkeley / MIT>" person/paper says ..... Yes, and? there is always so much context.
But social media incentivizes non-nuanced bite-sized panic/optimism inducing takes. How do we increase the context with which information is disemminated in media?

05.12.2024 06:11 β€” πŸ‘ 11    πŸ” 1    πŸ’¬ 3    πŸ“Œ 0

Two stages of writing:

1) This shouldn't take too long
2) Oh no

04.12.2024 19:36 β€” πŸ‘ 37484    πŸ” 7102    πŸ’¬ 567    πŸ“Œ 787
Preview
GitHub - allenai/OLMo: Modeling, training, eval, and inference code for OLMo Modeling, training, eval, and inference code for OLMo - allenai/OLMo

We just updated the OLMo repo at github.com/allenai/OLMo!
There are now several training configs that together reproduce the training runs that lead to the final OLMo 2 models.
In particular, all the training data is available, tokenized and shuffled exactly as we trained on it!

02.12.2024 20:13 β€” πŸ‘ 54    πŸ” 11    πŸ’¬ 0    πŸ“Œ 0

Let's cycle through the memes for this one until it stops... πŸ˜‡πŸ˜…πŸ™

02.12.2024 01:23 β€” πŸ‘ 38    πŸ” 6    πŸ’¬ 1    πŸ“Œ 1

It is?

30.11.2024 14:06 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
The unofficial logo of the Adaptive Agents Lab (Adage). It shows a cylindrical robot with some gauges and buttons on it.

The unofficial logo of the Adaptive Agents Lab (Adage). It shows a cylindrical robot with some gauges and buttons on it.

πŸŽ‰Good news, everyone! πŸŽ‰
I will recruit graduate students on the algorithmic and theoretical aspects of Reinforcement Learning.
You will join Adage, @mila-quebec.bsky.social and @polymtl.bsky.social.
More info on why and how you should apply:
academic.sologen.net/2024/11/22/g...
Deadline: Dec 1st

29.11.2024 03:23 β€” πŸ‘ 36    πŸ” 10    πŸ’¬ 2    πŸ“Œ 0

Space: *amorphous blob*

Astronomers: ah yes the choo choo train nebula

24.11.2024 08:54 β€” πŸ‘ 18    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Ok, last two papers for this week!

A final game-theoretic RLHF method and a different take on RLHF altogether inspired by prospect theory.

1. 🧲 Magnetic Preference Optimization (MPO).

2. Kahneman-Tversky Optimization (KTO).

🧡 1/3.

22.11.2024 12:43 β€” πŸ‘ 30    πŸ” 7    πŸ’¬ 2    πŸ“Œ 0

I have become a fan of the game-theoretic approaches to RLHF, so here are two more papers in that category! (with one more tomorrow πŸ˜…)

1. Self-Play Preference Optimization (SPO).

2. Direct Nash Optimization (DNO).

🧡 1/3.

21.11.2024 12:30 β€” πŸ‘ 74    πŸ” 9    πŸ’¬ 2    πŸ“Œ 2
A screenshot from the linked paper's figure 1. The figure is a pretty-complicated three column figure, but --- in essence, it sketches out how the authors compare llm sequences to the pretraining data / human authors to the pretraining data. Humans write more novel n-gram sequences.

A screenshot from the linked paper's figure 1. The figure is a pretty-complicated three column figure, but --- in essence, it sketches out how the authors compare llm sequences to the pretraining data / human authors to the pretraining data. Humans write more novel n-gram sequences.

LLMs generate novel word sequences not contained in their pretraining data. However, compared to humans, models generate significantly fewer novel n-grams.

RLHF = 30% *more* copying than base!

Awesome work from the awesome Ximing Lu (gloriaximinglu.github.io) et al. 🀩

arxiv.org/pdf/2410.04265

22.11.2024 06:14 β€” πŸ‘ 314    πŸ” 46    πŸ’¬ 6    πŸ“Œ 2

I've started putting together a starter pack with people working on Speech Technology and Speech Science: go.bsky.app/BQ7mbkA

(Self-)nominations welcome!

19.11.2024 11:13 β€” πŸ‘ 82    πŸ” 34    πŸ’¬ 44    πŸ“Œ 3
Post image

Doing good science is 90% finding a science buddy to constantly talk to about the project.

09.11.2024 22:53 β€” πŸ‘ 885    πŸ” 215    πŸ’¬ 22    πŸ“Œ 65

@ujaan is following 20 prominent accounts