Ross Taylor's Avatar

Ross Taylor

@rosstaylor90.bsky.social

Building new thing. Same handle at the other place

104 Followers  |  44 Following  |  7 Posts  |  Joined: 14.11.2024  |  1.3326

Latest posts by rosstaylor90.bsky.social on Bluesky

Post image 26.03.2025 20:45 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Friend of the pod Ross Taylor (ex reasoning lead for Llama) launched a platform for open reasoning data, gr.inc, seems promising!

21.02.2025 17:28 β€” πŸ‘ 30    πŸ” 5    πŸ’¬ 4    πŸ“Œ 0

β€œAnd yet it reasons”

05.12.2024 07:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
The Hitchhiker's Guide to Reasoning
YouTube video by Ross Taylor The Hitchhiker's Guide to Reasoning

(πŸŽ₯45m) The Hitchhiker's Guide to Reasoning

A talk about LLM reasoning, covering various methods, core problems, and future research directions!

www.youtube.com/watch?v=S5l5...

02.12.2024 11:33 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Taming Transformers for High-Resolution Image Synthesis Designed to learn long-range interactions on sequential data, transformers continue to show state-of-the-art results on a wide variety of tasks. In contrast to CNNs, they contain no inductive bias tha...

IMO VQGAN is why GANs deserve the NeurIPS test of time award. Suddenly our image representations were an order of magnitude more compact. Absolute game changer for generative modelling at scale, and the basis for latent diffusion models.

28.11.2024 00:09 β€” πŸ‘ 104    πŸ” 15    πŸ’¬ 2    πŸ“Œ 2

Really simple principle that would go a long way: assume good faith. Most people are trying to do the right thing.

28.11.2024 06:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The toxicity towards that HF employee for sharing that Bluesky post dataset was really bad, and reminded me of 2022 anti-LLM sentiment.

Not a great look for πŸ¦‹.

28.11.2024 06:26 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I personally vote text is self supervised learning (SSL) for the same reason image is SSL.

Text tokenization is a modification but so are patches in vision. If you want you can go all the way down to bits or raw audio samples. We know text bits work as (again modded) it's used in LLM compression.

19.11.2024 09:20 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Is it right to call large-scale pre-training on text self-supervised learning?

Seems more like large-scale, multi-task supervised learning. Labels are present as semantic units (tokens) in the corpus, unlike SSL for vision and other modalities?

18.11.2024 09:01 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Giving this place a go…

17.11.2024 17:57 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@rosstaylor90 is following 19 prominent accounts