this post was submitted on 30 Apr 2026
101 points (94.7% liked)

Technology

84281 readers
3330 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 18 comments
sorted by: hot top controversial new old
[–] phutatorius@lemmy.zip 10 points 16 hours ago

You're using the wrong model. Look up how a pump-and-dump scam works, all will be revealed.

[–] hahattpro@lemmy.world 4 points 19 hours ago

So if you overuse, you will be billed extra/or being disrupted. So to prevent disruption or incur extra charge, most people will not use all $19 monthly credits=> profit to github.

[–] MisterMoo@lemmy.world 3 points 20 hours ago
[–] Fizz@lemmy.nz 7 points 1 day ago (2 children)

The economics for daily use of inference seem to make sense. The cost of inference is highly profitable. The margins on inference around 80%. The lost money from power users is made up but the average user who doesn't user their tokens. They lose money on the free inference given away but that's marketing and getting used to people having the product there as a crutch. It's not the best business model but they can change it at any time and have vc cash to burn.

What doesn't make sense is recouping the investment cost of model training and building new data centers. Because the moat on a new model doesn't last long enough to recover its training cost.

[–] very_well_lost@lemmy.world 2 points 7 hours ago

The margins on inference around 80%.

Do you have a reliable source for this information? I've only ever heard numbers like this directly from the AI companies themselves.

[–] MangoCats@feddit.it 17 points 23 hours ago (1 children)

I don't think the current phase is about sensible business models, it's about jockeying for position to control the power of the new thing. The people doing the core investment have more money than they will ever need - this is a play to turn that money into more power than it currently represents for them - get in on the ground floor - shape the landscape - help form the regulations and relationships that will propel them up the next rung of the ladder.

It's a bit like spending $44B for a social media platform then running it into the ground financially - it's not about the money, it's about the things you get in exchange for that "waste" of money.

[–] Fizz@lemmy.nz 2 points 22 hours ago (3 children)

They're kinda past that phase and now need to show that they have sustainable revenue and user growth. From all the numbers I've seen they(open ai, Gemini, anthropic) have crazy numbers. Hundreds of millions of users paying $50 a month. It's not enough to cover training but it covers inference very nicely.

Then with agent bullshit they've managed to turn 1 prompt into 12 and bill the user for that extra so it's even more profitable than the monthly subscriptions.

[–] mlg@lemmy.world 4 points 17 hours ago

If it weren't for the massive silicon supply lockdown, I feel like we could easily see local models making it into consumer tech in the coming years and effectively replace all those casual users since you no longer have to pay a subscription to do regular/low effort tasks on whatever device you own. A lot of it has gotten really good, especially with lots of quantization techniques getting superseded by new ones each year.

Actually I guess it could probably go the same way as cable and streaming. Eventually they'll keep amping up the ante with the billing (because they always do), and people will just get turned off into a bunch of "cheaper" 3rd parties that have lower costs with some niche tricks, which will fragment the userbase too much.

Also I haven't looked into it, but do they advertise those $50 users separately from enterprise? I don't really know anyone outside of "power" users that aren't just using the $20 a month basic plans that give you enough tokens to get by (for now).

I feel like they're inflating their numbers from enterprise estimates because that's where they can bait with cheap API prices and then hook with vendor lock in.

[–] partofthevoice@lemmy.zip 2 points 22 hours ago* (last edited 22 hours ago)

If that’s true, you make a good point. Sounds like they could keep the business model around inference. They’d just need to figure out how to make up the budget for training, which I imagine could be done with good marketing on new releases. As well as finding ways to use more tokens, injecting ads, selling data, investors, … did I miss anything? It doesn’t sound impossible if the inference portion is as lucrative as you say.

[–] MangoCats@feddit.it 2 points 22 hours ago

Hundreds of millions of users paying $50 a month.

That's a hell of a lot better than .com had back in the day: hundreds of millions of "hits" on their websites, with no proof if there was even a human connected to the request.

turn 1 prompt into 12 and bill the user

I think there's a whole lot of variability of user experience out there still, and that's some of what is getting shaken out of the systems - new models are better for some, worse for others. Overall, I think they are still improving, quite dramatically for software creation in the past 12 months, but as they grow in their specialty skills, some of the users who were getting better results for other things do get hurt in the process.

If there ever is such a thing as GenAI, I suspect it will follow the medical model of your General Practitioner referring you to specialists as warranted.

[–] CombatWombat@feddit.online 11 points 1 day ago (4 children)

I hope Rimu can figure out how to make Piefed run on GPUs because we are absolutely going to have to bail out these tech companies when the AI bubble pops, and when we do we’re gonna nationalize these shitty data centers and turn them into the biggest fediverse deployment you’ve ever seen (I’m mostly joking. Mostly).

[–] crunchy@lemmy.dbzer0.com 17 points 1 day ago (1 children)

Assuming any of the proposed data centers, even the mentioned 15GW's worth that are "under construction," ever actually get built. Unfortunately it looks like we're steering more toward another 1929 than another 1999.

[–] phutatorius@lemmy.zip 2 points 16 hours ago

it looks like we’re steering more toward another 1929 than another 1999

What really needs to happen is another 1789.

I vote we give it to PeerTube, video hosting is so resource intensive but there's so much potential good content to be liberated from YouTube.

[–] cecilkorik@piefed.ca 3 points 21 hours ago

My greedy fingers are poised on ebay for the great datacenter clearance sale.

[–] EvergreenGuru@lemmy.world 5 points 1 day ago* (last edited 1 day ago) (2 children)

We should let them fail and be happy that home lab users will be getting a massive discount.

Also, crypto will be back as those people use similar resources. So crypto will be cheaper to mine. Hard to say if that will make it worth more or less. Crypto’s its own ponzi scheme waiting to collapse.

[–] greyscale@lemmy.grey.ooo 3 points 17 hours ago

They'll shred them before they let us have them

Additionally, they're hillariously power intensive. you'd need to get 6-7 full racks worth of it together to make it work and when it does it blows through 130KW a rack.

[–] XLE@piefed.social 4 points 1 day ago

I can't wait for crypto mining to be the (ear)bleeding edge once again.