magnesit

magnesit

@magnesit.bsky.social

Fan of privacy | thinks people should care less about how others wanna live | has love-hate-relationship with IPv6 | advocate of using AI in a reasonable way.

93 Followers 53 Following 907 Posts Joined Nov 2024
4 days ago

Why in the name of fuck is Gemini 3.1 Flash Lite priced at $1.50/M output tokens? lil bro it's not that good πŸ₯€πŸ₯€

0 0 0 0
6 days ago

Volla... πŸ™„

funny how it's the same 3 companies that come up with stuff like that every single time

8 0 0 0
1 week ago

I'm really, really looking forward to Deepseek V4! Let's just hope it releases soon, because the competition is evolving a lot right now...

1 0 0 0
1 week ago
Preview
Motorola News | Motorola's new partnership with GrapheneOS Motorola announces three new B2B solutions at MWC 2026, including GrapheneOS partnership, Moto Analytics and more.

We're happy to announce a long-term partnership with Motorola. We're collaborating on future devices meeting our privacy and security standards with official GrapheneOS support.

motorolanews.com/motorola-thr...

649 192 50 65
3 weeks ago

There's just no better option security-wise. But rest assured, they are working on their own phone together with a large OEM, (likely) coming 2027.

0 0 0 0
1 month ago

Flash (Fast) avoids mistakes a pure instruct model would never be able to avoid with the current SOTA. Even situations where way too much relevance would be put onto the first token of the response - and where every other instruct model fails - are handled well by Flash.

0 0 0 0
1 month ago

Gemini 3 Flash (Fast mode) is literally just a reasoning model that pretends like it's not and any comparison between instruct models is inherently unfair. Even minimal reasoning is still reasoning and you can clearly feel the difference in the response quality my opinion.

0 0 1 0
1 month ago

Graphene Is All You Need.

0 0 0 0
1 month ago

Every VLM Implementation except for Qwen's and Gemini's feels botched.

0 0 0 0
2 months ago

Update, support for the model improved with newer versions of llama.cpp and hits >60 t/s decode speed now.

I still don't believe this thing will run well on a phone though.

0 0 0 0
2 months ago

Doubling the number of active parameters but cutting the number of experts in half feels arbitrary and has not shown an improvement in output quality so far. The model runs slower though.

We'll have to see what Magistral can make out of it.

0 0 0 0
2 months ago
Preview
Introducing Mistral 3 | Mistral AI A family of frontier open-source multimodal models

They did: mistral.ai/news/mistral-3

3B, 8B, 14B, 675B.

1 0 0 0
2 months ago

I'm glad the RL slop has been reduced in the Ministral series of models. The models don't perform absolutely SOTA, but at least they do not seem to spam \boxed for every single problem now.

Their Magistral pipeline could probably continue to scale extremely well in the future though.

0 0 0 0
2 months ago

With enough modifications made to it, I'm certain Mistral could create an entirely new high-performance LLM from it.

They could also try to begin tackling problems like hallucinations with their own research, one of the most pressing issues in LLMs.

0 0 1 0
2 months ago

What Mistral needs to replicate is the quality of ML research; there is little use in training a European LLM just for it to perform worse than the counterparts it took it's technologies from.

Furthermore, efforts in reverse-engineering GPT-OSS seem to be worth a shot.

0 0 1 0
2 months ago

I'm slightly disappointed in Mistral Large 3 being solely based on a recycled Deepseek architecture with minor changes.

Mistral has a lot of potential, and while trying out custom architectures is risky for smaller ML startups, it's the only way to remain independent in the long term.

0 0 2 0
2 months ago

Meta will probably be making a much bigger comeback with their future Llama-series LLMs than Google did with Gemini 3, at least towards the technical community.

My stance on OpenAI for building the most "reliable" LLMs prevails. They just work most of the time.

0 0 0 0
2 months ago

I genuinely thought Google could do better. They have an absurdly good vertical integration. They are literally the perfect candidate for building LLMs due to their sheer dataset size, compute capability and top-tier researchers.

0 0 1 0
2 months ago

The model still makes such stupid mistakes I do not find myself using it at all anymore, not even for Nano Banana Pro whose capabilities have been severely over-hyped.

Nothing might come close to Gemini 3 in helping the AI bubble pop.

0 0 1 0
2 months ago

I can't help but think that Gemini 3 Flash has been even more benchmaxxed than other models...

Besides that, they waited way too with publishing Gemini 3. The models are only barely SOTA a few weeks after their initial release. What was the point of all that?!

0 0 1 0
3 months ago

Brace,

Mistral might be dropping bombshell LLMs very soon. 3B, 8B, and one proprietary.

0 0 1 0
3 months ago

THEY FUCKING FIXED IT

0 0 0 0
3 months ago

Yep, that's what I meant to say - device support will come eventually, but Android 16 QPR1 releasing doesn't go hand in hand with that immediately. :)

4 0 0 0
3 months ago

You will likely need to wait for a bit longer than for the generic QPR1 release until everything regarding the Pixel 10 is sorted out. I might be wrong though.

1 0 0 0
3 months ago

As far as I am aware, Android 16 QPR1 being ported doesn't necessarily mean Pixel 10 device support is coming. Google dropped the AOSP device source trees for their new phones, which not only made the jump to Android 16 harder but also hindered Pixel 10 adoption.

2 0 2 0
4 months ago
Preview
a man in a suit and tie says oh my god okay it 's happening .. Alt: The scene from "The Office" where the protagonist screams "Oh, my God! Okay, it's happening!"
10 0 0 0
4 months ago

Still the case btw

0 0 1 0
4 months ago
Chart showing Magistral Medium 1.2 outperforming Gemini 2.5 Flash and Deepseek R1 2508.

I know the Artificial Analysis Index can be inaccurate as hell

But boy, is that a hell of a beautiful sight.

1 0 0 0
4 months ago

Still the case btw

0 0 1 0
5 months ago

YouTube is rage baiting everyone, yet again:

They finally introduced a selector on MOBILE WEB for the "Audio track" which allows you to get rid of the terrible auto-translation... BUT you can only use this for Shorts. Not for full videos. They don't have that option there.

How the fuck.

1 0 0 0