this post was submitted on 30 Sep 2025
1060 points (98.6% liked)

Technology

75645 readers
3462 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

"No Duh," say senior developers everywhere.

The article explains that vibe code often is close, but not quite, functional, requiring developers to go in and find where the problems are - resulting in a net slowdown of development rather than productivity gains.

you are viewing a single comment's thread
view the rest of the comments
[–] sugar_in_your_tea@sh.itjust.works 17 points 1 day ago (33 children)

I personally think unit tests are the worst application of AI. Tests are there to ensure the code is correct, so ideally the dev would write the tests to verify that the AI-generated code is correct.

I personally don't use AI to write code, since writing code is the easiest and quickest part of my job. I instead use it to generate examples of using a new library, give me comparisons of different options, etc, and then I write the code after that. Basically, I use it as a replacement for a search engine/blog posts.

[–] Draces@lemmy.world 0 points 1 day ago (3 children)

What model are you using? I've had such a radically different experience but I've only bothered with the latest models. The old ones weren't even worth trying with

[–] sugar_in_your_tea@sh.itjust.works 2 points 1 day ago (2 children)

I'll have to check, we have a few models hosted at our company and I forget the exact versions and whatnot. They're relatively recent, but not the highest end since we need to host them locally.

But the issue here isn't directly related to which model it is, but to the way LLMs work. They cannot reason, they can only give believable output. If the goal is code coverage, it'll get coverage, but not necessarily be well designed.

If both the logic and the tests are automated, humans will be lazy and miss stuff. If only the logic is generated, humans can treat the code as a black box and write good tests that way. Humans will be lazy with whatever is automated, so if I have to pick one to be hand written, it'll be the code that ensures the logic is correct.

[–] wesley@yall.theatl.social 0 points 1 day ago (1 children)

We're mandated to use it at my work. For unit tests it can really go wild and it'll write thousands of lines of tests to cover a single file/class for instance whereas a developer would probably only write a fourth as much. You have to be specific to get any decent output from them like "write a test for this function and use inputs x and y and the expected output is z"

Personally I like writing tests too and I think through what test cases I need based on what the code is supposed to do. Maybe if there are annoying mocks that I need to create I'll let the AI do that part or something.

Generating tests like that would take longer than writing the tests myself...

Nobody is going to thoroughly review thousands of lines of test code.

load more comments (29 replies)