this post was submitted on 06 Jul 2025
295 points (97.7% liked)

Technology

72498 readers
3939 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 30 comments
sorted by: hot top controversial new old
[–] QuadratureSurfer@lemmy.world 183 points 3 days ago (2 children)

Someone just got the AWS bill.

[–] douglasg14b@lemmy.world 2 points 17 hours ago

More like they just got their Anthropic bill.

Cloud compute is gonna be cheap compared to the API costs for LLMs they use/offer.

[–] crunchy@lemmy.dbzer0.com 82 points 2 days ago (1 children)

That's got to be it. Cloud compute is expensive when you're not being funded in Azure credits. One the dust settles from the AI bubble bursting, most of the AI we'll see will probably be specialized agents running small models locally.

[–] fmstrat@lemmy.nowsci.com 13 points 2 days ago (1 children)

I'm still running Qwen32b-coder on a Mac mini. Works great, a little slow, but fine.

[–] And009@lemmynsfw.com 2 points 2 days ago (2 children)

I'm somewhat tech savvy, how do I run llm locally. Any suggestions? How to know if my local data is safe

[–] Retro_unlimited@lemmy.world 2 points 1 day ago

I have been using a program called GPT4ALL and you can download many models and run them locally. They give you a prompt at boot if you want to share data or not. I select no and use it offline anyway.

[–] Llak@lemmy.world 5 points 2 days ago

Checkout lm studio https://lmstudio.ai/ and you can pair it with vs continue extension https://docs.continue.dev/getting-started/overview.

[–] napkin2020@sh.itjust.works 16 points 2 days ago
[–] Ulrich@feddit.org 31 points 2 days ago* (last edited 2 days ago)

Ah they're learning from the "unlimited" mobile carriers.

"Unlimited" until you meet your limit, then throttled.

[–] Jesusaurus@lemmy.world 62 points 3 days ago

Sounds like charge back territory

[–] Glitchvid@lemmy.world 16 points 2 days ago

Imagine the price hikes when they need to get that return on hundreds of billions they've poured into these models, datacenters and electricity.

[–] Admax@lemmy.world 43 points 2 days ago (2 children)

Hopefully (?) this is the start of a trend and people might begin to realize how all those products are not worth their price and AI is an overhyped mess made to hook users before exploiting them...

[–] douglasg14b@lemmy.world 2 points 17 hours ago (1 children)

Are you a software engineer who has made use of these and similar tools?

If not, this is epic level armchairing.

The tools are definitely hyped, but they are also incredibly functional. They have many problems, but they also work and achieve their intended purpose.

[–] Admax@lemmy.world 1 points 14 hours ago

I have a rough idea of their efficiency as I've used them, not in professional settings but I wager it would not be too different.

My point is more that it feels like the rugs are finally starting to get pulled. This tech is functionnal as you said, it works to a point and that point is enough for a sizeable amount of people. But I doubt that the price most people are paying now is enough to cover the cost of answering their queries. Now that some people, especially younger devs or people who never worked without those tools are dependant on it, they can go ahead and charge more.

But it's not too late, so I'm hoping it will make some people more aware of that kind of scheme and that they will stop feeding the AI hype in general.

[–] cley_faye@lemmy.world 5 points 2 days ago

The whole industry is projecting something like negative $200B for next years. They know it's not worth the price.

[–] TrumpetX@programming.dev 13 points 3 days ago (1 children)

Well shit, I've been on vacation, and I signed up with Cursor a month ago. Not allowed at work, but for side projects at home in an effort to "see what all the fuss is about".

So far, the experience was rock solid, but I assume when I get home that I'll be unpleasantly surprised.

Has anyone here had rate limiting hit them?

[–] errer@lemmy.world 17 points 2 days ago* (last edited 2 days ago) (1 children)

I’ve primarily use claude-4-sonnet in cursor and was surprised to see a message telling me it would start costing extra above and beyond my subscription. This was prolly after 100 queries or so. However, switching to “auto” instead of a specific model continues to not cost anything and that still uses claude-4-sonnet when it thinks it needs to. Main difference I’ve noticed is it’s actually faster because it’ll sometimes hit cheaper/dumber APIs to address simple code changes.

It’s a nice toy that does improve my productivity quite a bit and the $20/month is the right price for me, but I have no loyalty and will drop them without delay if it becomes unusable. That hasn’t happened yet.

[–] axEl7fB5@lemmy.cafe 2 points 2 days ago

Common People