this post was submitted on 17 Jun 2025
87 points (100.0% liked)

TechTakes

1967 readers
255 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
 

I love to show that kind of shit to AI boosters. (In case you're wondering, the numbers were chosen randomly and the answer is incorrect).

They go waaa waaa its not a calculator, and then I can point out that it got the leading 6 digits and the last digit correct, which is a lot better than it did on the "softer" parts of the test.

you are viewing a single comment's thread
view the rest of the comments
[–] mountainriver@awful.systems 7 points 10 hours ago (1 children)

I find it a bit interesting that it isn't more wrong. Has it ingested large tables and got a statistical relationship between certain large factors and certain answers? Or is there something else going on?

[–] codexarcanum@lemmy.dbzer0.com 0 points 4 hours ago (1 children)

I posted a top level comment about this also, but Anthropic has done some research on this. The section on reasoning models discusses math I believe. The short version is it has a bunch of math in its corpus so it can approximate math (kind of, seemingly, similar to how you'd do a back of the envelope calculation in your head to get the orders of magnitude right) but it can't actually do calculations which is why they often get the specifics wrong.

[–] froztbyte@awful.systems 1 points 31 minutes ago

reasoning models

that’s a shot, everyone drink up