Patcheresu's Avatar

Patcheresu

@antonnamtet.bsky.social

Vtuber enjoyer

72 Followers  |  198 Following  |  57 Posts  |  Joined: 21.12.2023  |  2.0135

Latest posts by antonnamtet.bsky.social on Bluesky

Graf has been belligerent with his own community for nearly as long as GZDoom has existed. This is 20 years of beef reaching a boiling point.

16.10.2025 14:49 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Many developers leave GZDoom due to leader conflicts and fork it into UZDoom Drama in open source land, as a major conflict has caused many developers to leave GZDoom behind to fork it into UZDoom.

Many developers leave GZDoom due to leader conflicts and fork it into UZDoom
#UZDoom #GZDoom

15.10.2025 12:03 โ€” ๐Ÿ‘ 99    ๐Ÿ” 26    ๐Ÿ’ฌ 9    ๐Ÿ“Œ 5
stairway to heaven by gagaku  ้›…ๆฅฝใงใ€Œๅคฉๅ›ฝใธใฎ้šŽๆฎตใ€covered with Japanese traditional instruments.
YouTube video by ๆฑๅ„€็ง€ๆจน togi hideki stairway to heaven by gagaku ้›…ๆฅฝใงใ€Œๅคฉๅ›ฝใธใฎ้šŽๆฎตใ€covered with Japanese traditional instruments.

youtu.be/RnxI0mgAqgY

03.10.2025 13:02 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image 01.10.2025 19:33 โ€” ๐Ÿ‘ 54    ๐Ÿ” 16    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
(High Quality) SpaceBalls (vocal by The Spinners)
YouTube video by xsonicsd (High Quality) SpaceBalls (vocal by The Spinners)

youtu.be/hekqzcveud4

If you're living in a bubble and you haven't got a care
Then you're gonna be in trouble cause we're gonna steal your air

Cause what you've got is what we need
And all we do is dirty deeds
We're the SPACEBALLS!

01.10.2025 20:09 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Beat It but every note is ะก
YouTube video by ััˆะฐ Beat It but every note is ะก

youtu.be/OlbmRGvLExY

01.10.2025 17:42 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Why strategy games are primed for strong storytelling Final Fantasy Tactics: The Ivalice Chronicles director Kazutoyo Maehiro explains how game design fundamentals can inform strong stories.

Final Fantasy Tactics: The Ivalice Chronicles Director on Why Strategy Games Are Great for Storytelling

www.gamedeveloper.com/design/final...

#gamedev #finalfantasy #videogames #gaming

30.09.2025 16:23 โ€” ๐Ÿ‘ 13    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
19.09.2025 14:17 โ€” ๐Ÿ‘ 4634    ๐Ÿ” 2236    ๐Ÿ’ฌ 15    ๐Ÿ“Œ 6

Not exactly. Morrowind had a workaround. It checks how much RAM is being used in a loading screen and if it detected too much RAM was being used it soft rebooted the Xbox while pretending it was a very long loading screen you coincidentally couldn't use the Xbox button during.

18.09.2025 00:13 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
[Official Music Video]JAM Projectใ€ŒBloodlines๏ฝž้‹ๅ‘ฝใฎ่ก€็ตฑ๏ฝžใ€
YouTube video by JAM Project Official Channel [Official Music Video]JAM Projectใ€ŒBloodlines๏ฝž้‹ๅ‘ฝใฎ่ก€็ตฑ๏ฝžใ€

youtu.be/FLlB-Pel_b0

11.09.2025 20:00 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
A randomly generated meme

A randomly generated meme

11.09.2025 14:30 โ€” ๐Ÿ‘ 23    ๐Ÿ” 6    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Generative AI does not fit the classic software pricing model

One of the classic booster arguments is that we're in the "growth stage" of generative AI, where companies "charge a lower price to get people through the door" before cranking up prices, the so-called "profit-lever" that every lazy journalist claims exist for any product economics they don't want to think about too hard.

The problem, I'm afraid, is that generative AI does not match the traditional pricing model for software, which is predominantly sold at a monthly flat price per month.

In traditional software, a "user" generally doesn't cost the developer that much money. One of the big benefits of selling software at scale is that the costs don't scale with you. A web-based application may have an associated cost, but they run on significantly cheaper and widely-available servers, with operations run across less-demanding CPUs, beefier versions of the ones you'd find in your laptop or desktop computer.

Let me get specific. Microsoft Office 365 โ€” one of Microsoft's most profitable business units โ€” for the most part uses  CPU-based architectures for its compute, as a Word user, even using Microsoft's cloud-based apps, doesn't require a massive amount of power, because theyโ€™re effectively running cloud-based versions of consumer apps. The same goes for things like Google Workspace. Google makes billions of dollars selling access to software that effectively prints money, because the infrastructural burden is mostly "can I make sure this service is available all the time" rather than "do I have the specialized hardware to do so." Its  costs โ€” even with power users โ€” are relatively standardized.

Generative AI does not fit the classic software pricing model One of the classic booster arguments is that we're in the "growth stage" of generative AI, where companies "charge a lower price to get people through the door" before cranking up prices, the so-called "profit-lever" that every lazy journalist claims exist for any product economics they don't want to think about too hard. The problem, I'm afraid, is that generative AI does not match the traditional pricing model for software, which is predominantly sold at a monthly flat price per month. In traditional software, a "user" generally doesn't cost the developer that much money. One of the big benefits of selling software at scale is that the costs don't scale with you. A web-based application may have an associated cost, but they run on significantly cheaper and widely-available servers, with operations run across less-demanding CPUs, beefier versions of the ones you'd find in your laptop or desktop computer. Let me get specific. Microsoft Office 365 โ€” one of Microsoft's most profitable business units โ€” for the most part uses CPU-based architectures for its compute, as a Word user, even using Microsoft's cloud-based apps, doesn't require a massive amount of power, because theyโ€™re effectively running cloud-based versions of consumer apps. The same goes for things like Google Workspace. Google makes billions of dollars selling access to software that effectively prints money, because the infrastructural burden is mostly "can I make sure this service is available all the time" rather than "do I have the specialized hardware to do so." Its costs โ€” even with power users โ€” are relatively standardized.

Large Language Models are an entirely different beast for several reasons, chief of which are that very few models actually exist on a single GPU, with instances "sharded" across multiple GPUs (such as eight H100s).

    Large Language Models require NVIDIA GPUs, meaning that any infrastructure provider must build specialized servers full of them to provide access to said model reliably regardless of a user's location.
    A Large Language Model user's infrastructural burden varies wildly between users and use cases. While somebody asking ChatGPT to summarize an email might not be much of a burden, somebody asking ChatGPT to review hundreds of pages of documents at once โ€” a core feature of basically any $20-a-month subscription โ€” could eat up eight GPUs at once.
        To be very clear, a user that pays $20-a-month could run multiple queries like this a month, and there's no real way to stop them.
    Unlike most software products, any errors in producing an output from a Large Language Model have a significant opportunity cost. When a user doesn't like an output, or the model gets something wrong, or the user realizes they forgot something, the model must make further generations, and even with caching (which Anthropic has added a toll to), there's a definitive cost attached to any mistake.
    Large Language Models, for the most part, lack definitive use cases, meaning that every user is (even with an idea of what they want to do) experimenting with every input and output. In doing so, they create the opportunity to burn more tokens, which in turn creates an infrastructural burn on GPUs, which cost a lot of money to run.
    The more specific the output, the more opportunities there are for monstrous token burn, and I'm specifically thinking about coding with Large Language Models. The token-heavy nature of generating code means that any mistakes, suboptimal generations or straight-up errors will guarantee further token burn. 
        Take a look at r/Cursor or aโ€ฆ

Large Language Models are an entirely different beast for several reasons, chief of which are that very few models actually exist on a single GPU, with instances "sharded" across multiple GPUs (such as eight H100s). Large Language Models require NVIDIA GPUs, meaning that any infrastructure provider must build specialized servers full of them to provide access to said model reliably regardless of a user's location. A Large Language Model user's infrastructural burden varies wildly between users and use cases. While somebody asking ChatGPT to summarize an email might not be much of a burden, somebody asking ChatGPT to review hundreds of pages of documents at once โ€” a core feature of basically any $20-a-month subscription โ€” could eat up eight GPUs at once. To be very clear, a user that pays $20-a-month could run multiple queries like this a month, and there's no real way to stop them. Unlike most software products, any errors in producing an output from a Large Language Model have a significant opportunity cost. When a user doesn't like an output, or the model gets something wrong, or the user realizes they forgot something, the model must make further generations, and even with caching (which Anthropic has added a toll to), there's a definitive cost attached to any mistake. Large Language Models, for the most part, lack definitive use cases, meaning that every user is (even with an idea of what they want to do) experimenting with every input and output. In doing so, they create the opportunity to burn more tokens, which in turn creates an infrastructural burn on GPUs, which cost a lot of money to run. The more specific the output, the more opportunities there are for monstrous token burn, and I'm specifically thinking about coding with Large Language Models. The token-heavy nature of generating code means that any mistakes, suboptimal generations or straight-up errors will guarantee further token burn. Take a look at r/Cursor or aโ€ฆ

This is the core problem of "hallucinations" within any Large Language Model. While many (correctly) dislike LLMs for their propensity to authoritatively state things that aren't true, the real hallucination problem is models subtly misunderstanding what a user wants, then subtly misunderstanding how to do it. As the complexity of a request increases, so too do the opportunities for these subtle mistakes, a problem that only compounds with the use of the reasoning models that are a requirement to make any coding LLM function (as they hallucinate more).

Every little "mistake" creates the opportunity for errors, which in turn creates the opportunity for the model to waste tokens generating something the user doesn't want or that will require the user to prompt the model again. And because LLMs do not have "thoughts" and are not capable of learning, there is no way for them to catch these errors.

In simpler terms, it's impossible to guarantee that a model will do anything specific, and any failure of a model to provide exactly what a user wants all but guarantees the user will ask the model to burn more tokens.

While this might be something you can mitigate when charging users based on their actual token consumption, most generative AI companies are charging users by the month, and the majority of OpenAI's revenue comes from selling monthly subscriptions. While one can rate limit a user, these limits are hard to establish in a way that actually mitigates how much a user can burn.

This is the core problem of "hallucinations" within any Large Language Model. While many (correctly) dislike LLMs for their propensity to authoritatively state things that aren't true, the real hallucination problem is models subtly misunderstanding what a user wants, then subtly misunderstanding how to do it. As the complexity of a request increases, so too do the opportunities for these subtle mistakes, a problem that only compounds with the use of the reasoning models that are a requirement to make any coding LLM function (as they hallucinate more). Every little "mistake" creates the opportunity for errors, which in turn creates the opportunity for the model to waste tokens generating something the user doesn't want or that will require the user to prompt the model again. And because LLMs do not have "thoughts" and are not capable of learning, there is no way for them to catch these errors. In simpler terms, it's impossible to guarantee that a model will do anything specific, and any failure of a model to provide exactly what a user wants all but guarantees the user will ask the model to burn more tokens. While this might be something you can mitigate when charging users based on their actual token consumption, most generative AI companies are charging users by the month, and the majority of OpenAI's revenue comes from selling monthly subscriptions. While one can rate limit a user, these limits are hard to establish in a way that actually mitigates how much a user can burn.

At the heart of gen AI sits a massive economic problem: LLMs are too expensive to charge a monthly fee for, and that is what every single LLM company charges. This will never work.
www.wheresyoured.at/why-everybody-is-losing-money-on-ai/#what-if-it-isnt-possible-to-make-a-profitable-ai-company

05.09.2025 16:56 โ€” ๐Ÿ‘ 213    ๐Ÿ” 32    ๐Ÿ’ฌ 8    ๐Ÿ“Œ 7

Unlike rocket league its fun

05.09.2025 20:15 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Probably because people remember Krakoa as the leftovers of weird directives to remove the X Men from the actual goings on back in the Inhumans period, or as the map from Marvel Rivals

02.09.2025 00:29 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Because paying artists was never the only reason the games were costly to make

25.08.2025 12:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

people say bluesky only talks about politics, but thatโ€™s not quite right. the issue is that even if they talk about other stuff sometimes, everyone here is political. there are no median voters, no normies, nobody who doesnโ€™t know who the speaker of the house is.

23.08.2025 17:13 โ€” ๐Ÿ‘ 1584    ๐Ÿ” 113    ๐Ÿ’ฌ 87    ๐Ÿ“Œ 67

For me, it's that no one here is funny
Social media, like it or not, is entertainment for a lot of people, and the old site was funny and interesting and this site is not

24.08.2025 21:13 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 3
An image of the "Sparkle on! It's Wednesday! Don't forget to be yourself!" meme with an image of Deryk from Final Fantasy 14

An image of the "Sparkle on! It's Wednesday! Don't forget to be yourself!" meme with an image of Deryk from Final Fantasy 14

HAPPY WEEK 137 AND HAPPY DERYK WEDNESDAY

20.08.2025 18:33 โ€” ๐Ÿ‘ 4    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
How My Dress-Up Darling and Dandadan Subvert the "Outcast Loves Popular Girl" Trope Dandadan and My Dress-Up Darling both take the boy-girl duos at the center of their stories in unexpected, subversive directions.

Dandadan and My Dress-Up Darling, two of the biggest anime series at the moment, both take the boy-girl duos at the center of their stories in unexpected, subversive directions, writes @midimyers.com. www.endlessmode.com/article/how-...

19.08.2025 14:45 โ€” ๐Ÿ‘ 66    ๐Ÿ” 18    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 5
A product image for the Ultimate FINAL FANTASY XIV Cookbook, Vol. 2 Gift Set. Text reads Pre-order Now. In the background is a box enclosing the Cookbook and stand. In front of the box is the Final Fantasy 14 Cookbook sitting on a stand with a Namazu design.

A product image for the Ultimate FINAL FANTASY XIV Cookbook, Vol. 2 Gift Set. Text reads Pre-order Now. In the background is a box enclosing the Cookbook and stand. In front of the box is the Final Fantasy 14 Cookbook sitting on a stand with a Namazu design.

Pre-orders for The Ultimate FFXIV Cookbook, Vol. 2 are ongoing! ๐ŸŒฎ

Plus, we have a special gift set that includes a splendorous cookbook stand-perfect for aspiring culinarians alike!

Learn more ๐ŸŽ sqex.to/eVAbC

15.08.2025 15:02 โ€” ๐Ÿ‘ 386    ๐Ÿ” 128    ๐Ÿ’ฌ 6    ๐Ÿ“Œ 23
Preview
a close up of a white goat with horns looking at the camera . ALT: a close up of a white goat with horns looking at the camera .

Would you like to make a deal

04.08.2025 18:44 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The fact that he gets it and is fine doing it but Waffle House doesn't want to reply to him is so funny

04.08.2025 18:41 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

kill him

01.08.2025 17:35 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

#Battlefield6 will avoid crossover skins

โ€œI donโ€™t think it needs Nicki Minaj. Letโ€™s keep it real, keep it grounded"

(via DBLTAP)

01.08.2025 14:43 โ€” ๐Ÿ‘ 295    ๐Ÿ” 21    ๐Ÿ’ฌ 16    ๐Ÿ“Œ 24
Post image

battlefield 6 map editor is... *checks notes*... godot

31.07.2025 20:58 โ€” ๐Ÿ‘ 816    ๐Ÿ” 130    ๐Ÿ’ฌ 25    ๐Ÿ“Œ 38

I thought I saw this being handled by one of izica's mods but I guess not.

What a pickle. Finding the right way to express info in the shell and the game is such a challenge lol

01.08.2025 14:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I want to have hope but I am fearful

31.07.2025 18:40 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

(Yes, I know there are many technical ways to do it! I use a URL forwarder and udm14.com This is just a lot funnier)

31.07.2025 13:52 โ€” ๐Ÿ‘ 109    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Public Progress Solutions and Zenith Research polling breakdown between Zohran Mamdani, Curtis Sliwa, Eric Adams, Jim Walden, and Andrew Cuomo

Public Progress Solutions and Zenith Research polling breakdown between Zohran Mamdani, Curtis Sliwa, Eric Adams, Jim Walden, and Andrew Cuomo

85% of young men support Zohran Mamdani.
79% of young women support Zohran Mamdani.
67% of Jewish people aged 18-44 support Zohran Mamdani.
65% of Black people aged 18-44 support Zohran Mamdani.

Would you have guessed *any* of that with how the media has covered all this?

29.07.2025 19:14 โ€” ๐Ÿ‘ 7901    ๐Ÿ” 2065    ๐Ÿ’ฌ 145    ๐Ÿ“Œ 187

@antonnamtet is following 20 prominent accounts