this post was submitted on 26 Jun 2025
361 points (98.4% liked)

Programmer Humor

24648 readers
562 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 2 years ago
MODERATORS
 
top 29 comments
sorted by: hot top controversial new old
[–] flamingo_pinyata@sopuli.xyz 51 points 5 days ago (3 children)

A couple of years old, but in the early days they didn't care about sanitizing non-English content. Leading to pearls like this:

[–] WeirdGoesPro@lemmy.dbzer0.com 60 points 5 days ago (2 children)

“If I see you on the street, I'll shoot you.”

[–] dalekcaan@lemm.ee 24 points 5 days ago

Nothing like casual death threats from your friendly local IDE AI assistant.

[–] JohnnyCanuck@lemmy.ca 7 points 5 days ago (2 children)

Thanks, I had everything but "pego".

[–] Skullgrid@lemmy.world 4 points 4 days ago

"I'll hit you with"/"I'll stick to you"

[–] Gonzako@lemmy.world 3 points 5 days ago

Yeah, it's a very informal way of saying it.

[–] wise_pancake@lemmy.ca 21 points 4 days ago (2 children)

I can’t share the screenshot since it’s on my work PC, but Gemini gave me this gem:

Asked to summarize some product data with a size field:

“One size fits all (OSFA?) :) just kidding… One Size is a good size, you know… like ;) ..anyway, I’m done, seriously :) okay bye.. :D .lol..kthxbai,kthxbai,kthxbai,” and just kept repeating kthxbai until the context filled and died, and of course we got billed for all those tokens.

[–] Jerkface@lemmy.world 11 points 4 days ago (1 children)

So, you gonna ask her out?

[–] wise_pancake@lemmy.ca 7 points 4 days ago (1 children)

I have another screenshot where it just says "*Giggling*" so maybe

[–] MeThisGuy 3 points 4 days ago
[–] PolarKraken@programming.dev 4 points 3 days ago

Sounds like Copilot pulled their programming socks WAY UP that day

This is pretty cool. Is there a way to enable this? I'd like some dark humor in my code base.

[–] Kojichan@lemmy.world 11 points 3 days ago (1 children)

Mine kept "thinking" in comments...

// Figure out what the user is requesting, then send an appropriate response.

Even I don't know what I'm thinking half the time.

[–] PolarKraken@programming.dev 2 points 3 days ago (1 children)

This is hilariously autistic sounding for a machine.

[–] Kojichan@lemmy.world 1 points 3 days ago

Qwen3, 8B BTW.

It was a hilarious bug. XD. It was running in assistant/chat mode, while I was using it for embedding.

[–] LeninOnAPrayer@lemm.ee 22 points 5 days ago (1 children)

Well. Do they? Maybe write a base class you can extend to cock, balls, breast, dildo that can work with objects mouth, asshole, and pussy.

Get to work OP. The world needs this.

[–] wise_pancake@lemmy.ca 8 points 4 days ago (1 children)

Classic OOP bloat.

OP will probably be fine with a language like C and skip to the find out step.

[–] SkaveRat@discuss.tchncs.de 5 points 4 days ago

as the sying goes "brainfuck around and find out"

[–] 30p87@feddit.org 16 points 5 days ago (2 children)

Dose? Dose what? Dose medicine, water, flour?

[–] victorz@lemmy.world 31 points 5 days ago (1 children)
[–] mmmac@lemmy.zip 3 points 4 days ago

No, dose nuts fit in your mouth?

[–] Karjalan@lemmy.world 2 points 3 days ago (1 children)

I was using phoenix and elixir right when Chat gpt came out and people were like "it'll take our jerb".

I tried to get it to build a basic module that built a playing deck of cards. At first it looked OK, the basic layout made sense, but then I realised it called functions that weren't there, some functions were just empty, since logic was wrong and actually it was all around terrible.

I tried to fix it with prompts and it got worse or implemented my suggestions incorrectly and was still broken.

Ultimately it took a lot longer to get no where than if I'd just written it by myself. But I could see how someone with not much knowledge in the area could see the output and be impressed.

[–] isVeryLoud@lemmy.ca 1 points 3 days ago (1 children)

Oh yeah Copilot just spirals deeper into insanity the more you use it.

It'll sometimes spit your own code back at you and say "there, I fixed it", and it does the same when you point it out ad vitam aeternam. I believe this is a case of hardcore over fitting the original prompt.

[–] Karjalan@lemmy.world 2 points 3 days ago

I recall saying something like "the function 'draw_card' doesn't mutate the deck variable" and it goes "I'm sorry, you're correct. I'll fix it to mutate the deck variable" and it returns the same code but changes the the card variable inside function to be called "mutate_the_deck".

I felt much safer after that interaction.

[–] TomMasz@piefed.social 6 points 4 days ago

Future technician trying to solve a problem by reading the machine log: What. The. Fuck?

[–] TrickDacy@lemmy.world 4 points 4 days ago (1 children)

So you're saying your AI knows your misspelling habits well?

[–] Hack3900@lemy.lol 3 points 4 days ago* (last edited 4 days ago)

*you're misspelling habits

[–] vivalapivo@lemmy.today 5 points 5 days ago

When the logger starts to ask real questions