Jonathan Ross

Jonathan Ross

@jonathan-ross.bsky.social

CEO + Founder @ Groq, the Most Popular API for Fast Inference | Creator of the TPU and LPU, Two of the World’s Most Important AI Chips | On a Mission to Double the World's AI Compute by 2027

980 Followers 16 Following 42 Posts Joined Nov 2024
5 months ago
Preview
Groq CEO: Our mission is to provide over half of the world’s inference compute Jonathan Ross, CEO and founder of Groq, joins CNBC’s 'Squawk on the Street' to discuss the AI chip startup’s $750 million funding round, its push to deliver faster, lower-cost inference chips, and why...

Fantastic insight on the massive demand for AI inference infrastructure “The demand for AI compute is insatiable” @groq.com CEO @jonathan-ross.bsky.social, “Our mission is to provide over half of the world’s inference compute” - @cnbc.com

cnb.cx/4nG7Pcm #AI

5 4 0 0
6 months ago

Founder Tip #2: You have to spend time to make time.

Hiring, re-organizing, calendar clean up (across the team), preparation for meetings (internal and external), etc. Half my day is available for whatever I find important - because the other half is spent freeing up time.

2 1 0 0
6 months ago

Clearly China doesn't have enough compute for scaled AI today:
- GPT-OSS, Llama [US]: optimized for cheaper inference
- R1, Kimi K2, Qwen [China]: optimized for cheaper training

With China's population reducing inference costs is more important, and that means more training.

1 1 0 0
10 months ago
Post image

Transcribe audio with @groq.com.

6 2 1 0
11 months ago

I spent the weekend hanging out with a group of friends. A question we asked was what dreams did we have that we gave up on?

When I was 18, I had two dreams:

1) Be an astronaut
2) Build AI chips

I didn’t give up on one of them. 😀

2 0 0 0
1 year ago
Preview
Mistral Saba Added to GroqCloud™ Model Suite - Groq is Fast AI Inference GroqCloud™ has added another openly-available model to our suite – Mistral Saba. Mistral Saba is Mistral AI’s first specialized regional language model,

Big news! Mistral AI Saba 24B is on GroqCloud! The specialized regional language model is perfect for Middle East and South Asia-based devs and enterprises building AI solutions that need fast inference.
Learn more: groq.com/mistral-saba...

7 1 1 0
1 year ago
YouTube
Jonathan Ross, Founder & CEO @ Groq: NVIDIA vs Groq - The Future of Training vs Inference | E1260 YouTube video by 20VC with Harry Stebbings

YouTube: www.youtube.com/watch?v=xBMR...

Spotify: open.spotify.com/episode/30np...

Try Groq: console.groq.com

6 1 1 0
1 year ago
Video thumbnail

It was a pleasure being back on 20VC with Harry Stebbings. His craft of interviewing is second to none and we went deep.

This is the interview after we just launched 19,000 LPUs in Saudi Arabia. We built the largest inference cluster in the region.

Link to the interview in the comments below!

66 7 4 0
1 year ago
Post image

We built the region’s largest inference cluster in Saudi Arabia in 51 days and we just announced a $1.5B agreement for Groq to expand our advanced LPU-based AI inference infrastructure.

Build fast.

8 1 0 0
1 year ago
Preview
a close up of a man 's face with the words inconceivable written on it . ALT: a close up of a man 's face with the words inconceivable written on it .
2 0 0 0
1 year ago
Video thumbnail

My emergency episode with @harrystebbings.bsky.social at 20VC just launched on the impact of #DeepSeek on the AI world

156 31 7 2
1 year ago
Post image

Yesterday at the World Economic Forum in Davos, I joined a constructive discussion on AGI alongside @andrewyng.bsky.social, @yejinchoinka.bsky.social, @jonathan-ross.bsky.social , @thomwolf.bsky.social and moderator @nxthompson.bsky.social. Full discussion here: www.weforum.org/meetings/wor...

46 6 1 1
1 year ago
Post image
14 2 0 0
1 year ago

Thank you! 🙏

3 0 0 0
1 year ago

Over the next decade, we want to drive the cost down for generative AI 1,000x making a lot more activities profitable. And we think that that will cause a 100x spend increase.

🧵(5/5)

8 0 0 0
1 year ago

Over the last 60 years, almost like clockwork, every decade compute gets about 1000x cheaper, people buy 100,000x as much of it, spending 100x times more overall. 

Our mission at Groq is to drive the cost of compute towards zero;The cheaper we make compute the more people spend.

🧵(4/5)

5 0 1 0
1 year ago

- The answer is when you make a steam engine more efficient, it reduces the OpEx;

- When you reduce the OpEx, it increases the number of activities that are profitable;

- Therefore, people will do more things using steam engines and coal demand rises.

The same paradox applies to compute.

🧵(3/5)

2 0 1 0
1 year ago

It’s a paradox because if they're more efficient, why are they buying more coal?

🧵(2/5)

1 0 1 0
1 year ago
Post image

When you make compute cheaper do people buy more?

Yes. It's called Jevons Paradox and it's a big part of our business thesis.

In the 1860s, an Englishman wrote a treatise on coal where he noted that every time steam engines got more efficient people bought more coal.

🧵(1/5)

9 1 1 0
1 year ago
Post image

This is insane, Groq is the #4 API on this list! 😮

OpenAI, Anthropic, and Azure are the top 3 LLM API providers on LangChain

Groq is #4, and close behind Azure

Google, Amazon, Mistral, and Hugging Face are the next 4.

Ollama is for local development.

Now add three more 747's worth of LPUs 😁

17 2 1 0
1 year ago
YouTube
2025 Predictions with bestie Gavin Baker YouTube video by All-In Podcast

www.youtube.com/watch?v=HxNU...

1 0 0 0
1 year ago

Groq just got a shout out on the All-In pod as one of the big winners for 2025 alongside Nvidia. It’s the year of the AI chip and ours is the fastest 😃

5 0 1 0
1 year ago
Video thumbnail

Welcome to Shipmas - Groq Style.

Groq's second B747 this week. How many LPUs and GroqRacks can we load into a jumbo jet? Take a look.

Have you been naughty or nice?

12 1 0 0
1 year ago
Post image Post image Post image

Santa rented two full 747s this week to make his holiday deliveries of GroqRacks. Ho ho ho! 🎅

18 2 2 1
1 year ago

(5/5) Learning: product-led growth works; even when your product is too large and expensive to let people have it for free, you just have to be more creative about it.

2 0 0 0
1 year ago

(4/5) We're not shipping anyone millions of dollars of hardware.
It’s not a big ask for them to try it.
And when they try it, they love it.

2 0 1 0
1 year ago

(3/5) By making Groq easy and low cost to try, we got 60,000 developers on our developer console in 30 days. Less than a year after that, and we're at 645,000 developers and growing.

1 0 2 0
1 year ago

(2/5) That makes it almost impossible to do counterintuitive things. Like, “Try this new chip called an LPU", when everything in the zeitgeist is talking about GPUs.

And if you're a startup? Forget it.

That realization is why we made the strategic decision to put up our own cloud.

0 0 1 0
1 year ago
Post image

(1/5) One of the reasons why chips are so hard to innovate in is because if you're asking someone to put up a 10 million, 100 million, or a billion dollar check they need to know that what they're buying is going to work.

6 1 1 0
1 year ago

(5/5) The new Llama-3.3-70B model launched and is now available to all 645,000 GroqCloud™ developers as of this morning. Go cook, and don't forget to share what you build here.

Thank you for making GroqCloud™ the #1 API for fast inference! This is only just the beginning.

4 1 0 0