Raju Penmatsa's Avatar

Raju Penmatsa

@rajuptvs.bsky.social

ML Research Engineer @ Hitachi America R&D. also on x at https://x.com/iam_rajuptvs

97 Followers  |  417 Following  |  30 Posts  |  Joined: 27.10.2024  |  2.2865

Latest posts by rajuptvs.bsky.social on Bluesky

It has always been an adversarial game, and will always be.

28.11.2024 19:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Sick!!! 🀣

28.11.2024 17:24 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Using Machine Learning to Aid Survivors and Race through Time We’re on a journey to advance and democratize artificial intelligence through open source and open science.

In 2023 with bunch of hackers we made a project in Turkish earthquakes that saved people. Powered by HF compute with open-source models by Google

I went to my boss @julien-c.hf.co asked that day if I could use company's compute and he said "have whatever you need".
hf.co/blog/using-ml-for-disasters

27.11.2024 15:33 β€” πŸ‘ 69    πŸ” 1    πŸ’¬ 1    πŸ“Œ 2

It's pretty sad to see the negative sentiment towards Hugging Face on this platform due to a dataset put by one of the employees. I want to write a small piece. 🧡

Hugging Face empowers everyone to use AI to create value and is against monopolization of AI it's a hosting platform above all.

27.11.2024 15:23 β€” πŸ‘ 462    πŸ” 72    πŸ’¬ 29    πŸ“Œ 8

FYI, here's the entire code to create a dataset of every single bsky message in real time:

```
from atproto import *
def f(m): print(m.header, parse_subscribe_repos_message())
FirehoseSubscribeReposClient().start(f)
```

28.11.2024 09:56 β€” πŸ‘ 443    πŸ” 62    πŸ’¬ 20    πŸ“Œ 10

The thing is, there's already a dataset of 235 MILLION posts from 4 MILLION users available for months. Not sure why @hf.co is a target of abuse

zenodo.org/records/1108...

28.11.2024 01:32 β€” πŸ‘ 116    πŸ” 13    πŸ’¬ 7    πŸ“Œ 0
Post image Post image Post image Post image

google released another product focused on learning.
it is called "learn about" (realized this through google ai studio and learnlm model)
this is like cousin to notebooklm, but more open-ended and interactive.
for me learning using ai, is my favorite usecase.
learning.google.com/experiments/...

25.11.2024 23:40 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Napkin Math For Fine Tuning Pt. 1 w/Johno Whitaker
YouTube video by Hamel Husain Napkin Math For Fine Tuning Pt. 1 w/Johno Whitaker

just wanted to share this super practical video for anyone who is dealing with OOM errors, and want to understand various optimization techniques for fine-tuning.Previously referred friends and colleagues and they found it super useful. my favorite class in the course
youtu.be/-2ebSQROew4?...

22.11.2024 16:12 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

But I think we can still change the default from concise to other. I definitely remember doing that.

definitely worth a shot.

22.11.2024 03:32 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
a man is sitting at a desk working on a computer . ALT: a man is sitting at a desk working on a computer .

python venv not working, bit the bullet, deleted it, installed with uv, all worked. ????

21.11.2024 03:36 β€” πŸ‘ 96    πŸ” 5    πŸ’¬ 13    πŸ“Œ 1
Post image Post image

I've spent the last two years scouring all available resources on RLHF specifically and post training broadly. Today, with the help of a totally cracked team, we bring you the fruits of that labor β€” TΓΌlu 3, an entirely open frontier model post training recipe. We beat Llama 3.1 Instruct.

Thread.

21.11.2024 17:01 β€” πŸ‘ 212    πŸ” 42    πŸ’¬ 8    πŸ“Œ 10
Preview
RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning Large language models (LLMs) deployed as agents solve user-specified tasks over multiple steps while keeping the required manual engagement to a minimum. Crucially, such LLMs need to ground their gene...

this has a lot of parallels to this.
arxiv.org/abs/2410.02089

21.11.2024 17:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
TΓΌlu 3: The next era in open post-training We give you open-source, frontier-model post-training.

sorry my bad. just saw this on the post,
looks like this is gonna be explored in future.
open.substack.com/pub/robotic/...

21.11.2024 17:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Code might have a lot of overhead in computation, but since code has shown to increase model generalization capabilities over time.
Also this might help model learn, why some code was wrong if there is error and can correct itself.

21.11.2024 17:43 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

looks very interesting, and on quick glance makes a lot of sense. especially the verifiable rewards part of it.
Is there an extension to this where, it includes code generation and execution feedback is taken into account for RL.

21.11.2024 17:39 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

for me i really think, this preview is a way to collect user data and usage pattern, and hone in the RL policy that was used during training on user queries.

this for me is a typical ml practice.. where you deploy the model, collect user feedback and iterate and curate similar datasets and iterate.

21.11.2024 17:35 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

keyboard looks dope !!

21.11.2024 17:31 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Please tell me more about your incredible SWE-bench score

21.11.2024 17:25 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

pymupdf4llm from pyMuPDF is really good in parsing pdfs and converting them to markdown.

embedding image link in .md is really handy

18.11.2024 21:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Teleop of in-home robot using a low-cost setup (all open sourced soon)

18.11.2024 19:47 β€” πŸ‘ 51    πŸ” 9    πŸ’¬ 4    πŸ“Œ 1

thanks a lot for this.. will check it out..

18.11.2024 19:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

thanks for this much needed atm !! Kudos to the team!!

18.11.2024 18:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

at first glance, looks inefficient (i maybe wrong).. looks like the native scaled decoder is trying to cover up for the small image encoder and insufficient signals from them.

But hey.. if it works, it works πŸ˜…

18.11.2024 18:58 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Lol, so true.. are there any promising papers that show the effect of scaling image encoder.
This seems to be quite disproportionate, image encoder vs other params.

18.11.2024 18:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

super impressed by Qwen2vl,
both 7b and 72B are just awesome.
if the problem is broken into subtasks,
7b performance significantly increases.

In my limited evaluation,
7b beats the new sonnet too for image based extraction.

Kudos to the team!!

11.11.2024 22:42 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Note to my future self:

THINK OUT LOUD,
AND SHARE MORE IN PUBLIC (can be in various ways)

09.11.2024 03:13 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

kind of don't want to publicize much about this platform,
already feel anxious that people will start flooding here and might lose the current vibes that I am loving here.

09.11.2024 02:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

If you listen to podcasts and like infrastructure, databases, cloud, or open source you should check it out

09.11.2024 02:43 β€” πŸ‘ 6    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Ship It! Always ships on Friday 😎

Let us know if you like the occasional news/articles episode. Trying to find a balance with interviews

@withenoughcoffee.bsky.social and I obviously recorded this before this week

08.11.2024 22:23 β€” πŸ‘ 20    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

another great find on 🟦☁️. thanks πŸ™.

09.11.2024 02:53 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@rajuptvs is following 19 prominent accounts