Graf has been belligerent with his own community for nearly as long as GZDoom has existed. This is 20 years of beef reaching a boiling point.
16.10.2025 14:49 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0@antonnamtet.bsky.social
Vtuber enjoyer
Graf has been belligerent with his own community for nearly as long as GZDoom has existed. This is 20 years of beef reaching a boiling point.
16.10.2025 14:49 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Many developers leave GZDoom due to leader conflicts and fork it into UZDoom
#UZDoom #GZDoom
youtu.be/hekqzcveud4
If you're living in a bubble and you haven't got a care
Then you're gonna be in trouble cause we're gonna steal your air
Cause what you've got is what we need
And all we do is dirty deeds
We're the SPACEBALLS!
Final Fantasy Tactics: The Ivalice Chronicles Director on Why Strategy Games Are Great for Storytelling
www.gamedeveloper.com/design/final...
#gamedev #finalfantasy #videogames #gaming
Not exactly. Morrowind had a workaround. It checks how much RAM is being used in a loading screen and if it detected too much RAM was being used it soft rebooted the Xbox while pretending it was a very long loading screen you coincidentally couldn't use the Xbox button during.
18.09.2025 00:13 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0A randomly generated meme
11.09.2025 14:30 โ ๐ 23 ๐ 6 ๐ฌ 0 ๐ 0Generative AI does not fit the classic software pricing model One of the classic booster arguments is that we're in the "growth stage" of generative AI, where companies "charge a lower price to get people through the door" before cranking up prices, the so-called "profit-lever" that every lazy journalist claims exist for any product economics they don't want to think about too hard. The problem, I'm afraid, is that generative AI does not match the traditional pricing model for software, which is predominantly sold at a monthly flat price per month. In traditional software, a "user" generally doesn't cost the developer that much money. One of the big benefits of selling software at scale is that the costs don't scale with you. A web-based application may have an associated cost, but they run on significantly cheaper and widely-available servers, with operations run across less-demanding CPUs, beefier versions of the ones you'd find in your laptop or desktop computer. Let me get specific. Microsoft Office 365 โ one of Microsoft's most profitable business units โ for the most part uses CPU-based architectures for its compute, as a Word user, even using Microsoft's cloud-based apps, doesn't require a massive amount of power, because theyโre effectively running cloud-based versions of consumer apps. The same goes for things like Google Workspace. Google makes billions of dollars selling access to software that effectively prints money, because the infrastructural burden is mostly "can I make sure this service is available all the time" rather than "do I have the specialized hardware to do so." Its costs โ even with power users โ are relatively standardized.
Large Language Models are an entirely different beast for several reasons, chief of which are that very few models actually exist on a single GPU, with instances "sharded" across multiple GPUs (such as eight H100s). Large Language Models require NVIDIA GPUs, meaning that any infrastructure provider must build specialized servers full of them to provide access to said model reliably regardless of a user's location. A Large Language Model user's infrastructural burden varies wildly between users and use cases. While somebody asking ChatGPT to summarize an email might not be much of a burden, somebody asking ChatGPT to review hundreds of pages of documents at once โ a core feature of basically any $20-a-month subscription โ could eat up eight GPUs at once. To be very clear, a user that pays $20-a-month could run multiple queries like this a month, and there's no real way to stop them. Unlike most software products, any errors in producing an output from a Large Language Model have a significant opportunity cost. When a user doesn't like an output, or the model gets something wrong, or the user realizes they forgot something, the model must make further generations, and even with caching (which Anthropic has added a toll to), there's a definitive cost attached to any mistake. Large Language Models, for the most part, lack definitive use cases, meaning that every user is (even with an idea of what they want to do) experimenting with every input and output. In doing so, they create the opportunity to burn more tokens, which in turn creates an infrastructural burn on GPUs, which cost a lot of money to run. The more specific the output, the more opportunities there are for monstrous token burn, and I'm specifically thinking about coding with Large Language Models. The token-heavy nature of generating code means that any mistakes, suboptimal generations or straight-up errors will guarantee further token burn. Take a look at r/Cursor or aโฆ
This is the core problem of "hallucinations" within any Large Language Model. While many (correctly) dislike LLMs for their propensity to authoritatively state things that aren't true, the real hallucination problem is models subtly misunderstanding what a user wants, then subtly misunderstanding how to do it. As the complexity of a request increases, so too do the opportunities for these subtle mistakes, a problem that only compounds with the use of the reasoning models that are a requirement to make any coding LLM function (as they hallucinate more). Every little "mistake" creates the opportunity for errors, which in turn creates the opportunity for the model to waste tokens generating something the user doesn't want or that will require the user to prompt the model again. And because LLMs do not have "thoughts" and are not capable of learning, there is no way for them to catch these errors. In simpler terms, it's impossible to guarantee that a model will do anything specific, and any failure of a model to provide exactly what a user wants all but guarantees the user will ask the model to burn more tokens. While this might be something you can mitigate when charging users based on their actual token consumption, most generative AI companies are charging users by the month, and the majority of OpenAI's revenue comes from selling monthly subscriptions. While one can rate limit a user, these limits are hard to establish in a way that actually mitigates how much a user can burn.
At the heart of gen AI sits a massive economic problem: LLMs are too expensive to charge a monthly fee for, and that is what every single LLM company charges. This will never work.
www.wheresyoured.at/why-everybody-is-losing-money-on-ai/#what-if-it-isnt-possible-to-make-a-profitable-ai-company
Unlike rocket league its fun
05.09.2025 20:15 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Probably because people remember Krakoa as the leftovers of weird directives to remove the X Men from the actual goings on back in the Inhumans period, or as the map from Marvel Rivals
02.09.2025 00:29 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0Because paying artists was never the only reason the games were costly to make
25.08.2025 12:02 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0people say bluesky only talks about politics, but thatโs not quite right. the issue is that even if they talk about other stuff sometimes, everyone here is political. there are no median voters, no normies, nobody who doesnโt know who the speaker of the house is.
23.08.2025 17:13 โ ๐ 1584 ๐ 113 ๐ฌ 87 ๐ 67For me, it's that no one here is funny
Social media, like it or not, is entertainment for a lot of people, and the old site was funny and interesting and this site is not
An image of the "Sparkle on! It's Wednesday! Don't forget to be yourself!" meme with an image of Deryk from Final Fantasy 14
HAPPY WEEK 137 AND HAPPY DERYK WEDNESDAY
20.08.2025 18:33 โ ๐ 4 ๐ 1 ๐ฌ 0 ๐ 0Dandadan and My Dress-Up Darling, two of the biggest anime series at the moment, both take the boy-girl duos at the center of their stories in unexpected, subversive directions, writes @midimyers.com. www.endlessmode.com/article/how-...
19.08.2025 14:45 โ ๐ 66 ๐ 18 ๐ฌ 3 ๐ 5A product image for the Ultimate FINAL FANTASY XIV Cookbook, Vol. 2 Gift Set. Text reads Pre-order Now. In the background is a box enclosing the Cookbook and stand. In front of the box is the Final Fantasy 14 Cookbook sitting on a stand with a Namazu design.
Pre-orders for The Ultimate FFXIV Cookbook, Vol. 2 are ongoing! ๐ฎ
Plus, we have a special gift set that includes a splendorous cookbook stand-perfect for aspiring culinarians alike!
Learn more ๐ sqex.to/eVAbC
The fact that he gets it and is fine doing it but Waffle House doesn't want to reply to him is so funny
04.08.2025 18:41 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0kill him
01.08.2025 17:35 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0#Battlefield6 will avoid crossover skins
โI donโt think it needs Nicki Minaj. Letโs keep it real, keep it grounded"
(via DBLTAP)
battlefield 6 map editor is... *checks notes*... godot
31.07.2025 20:58 โ ๐ 816 ๐ 130 ๐ฌ 25 ๐ 38I thought I saw this being handled by one of izica's mods but I guess not.
What a pickle. Finding the right way to express info in the shell and the game is such a challenge lol
I want to have hope but I am fearful
31.07.2025 18:40 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0(Yes, I know there are many technical ways to do it! I use a URL forwarder and udm14.com This is just a lot funnier)
31.07.2025 13:52 โ ๐ 109 ๐ 2 ๐ฌ 2 ๐ 0Public Progress Solutions and Zenith Research polling breakdown between Zohran Mamdani, Curtis Sliwa, Eric Adams, Jim Walden, and Andrew Cuomo
85% of young men support Zohran Mamdani.
79% of young women support Zohran Mamdani.
67% of Jewish people aged 18-44 support Zohran Mamdani.
65% of Black people aged 18-44 support Zohran Mamdani.
Would you have guessed *any* of that with how the media has covered all this?