this post was submitted on 20 Jul 2023
47 points (92.7% liked)

Technology

58031 readers
4417 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Over just a few months, ChatGPT went from correctly answering a simple math problem 98% of the time to just 2%, study finds. Researchers found wild fluctuations—called drift—in the technology’s abi...::ChatGPT went from answering a simple math correctly 98% of the time to just 2%, over the course of a few months.

all 44 comments
sorted by: hot top controversial new old
[–] blue_zephyr@lemmy.world 11 points 1 year ago* (last edited 1 year ago)

This paper is pretty unbelievable to me in the literal sense. From a quick glance:

First of all they couldn't even bother to check for simple spelling mistakes. Second, all they're doing is asking whether a number is prime or not and then extrapolating the results to be representative of solving math problems.

But most importantly I don't believe for a second that the same model with a few adjustments over a 3 month period would completely flip performance on any representative task. I suspect there's something seriously wrong with how they collect/evaluate the answers.

And finally, according to their own results, GPT3.5 did significantly better at the second evaluation. So this title is a blatant misrepresentation.

Also the study isn't peer-reviewed.

[–] DominicHillsun@lemmy.world 7 points 1 year ago (4 children)

It seems rather suspicious how much ChatGPT has deteorated. Like with all software, they can roll back the previous, better versions of it, right? Here is my list of what I personally think is happening:

  1. They are doing it on purpose to maximise profits from upcoming releases of ChatGPT.
  2. They realized that the required computational power is too immense and trying to make it more efficient at the cost of being accurate.
  3. They got actually scared of it's capabilities and decided to backtrack in order to make proper evaluations of the impact it can make.
  4. All of the above
[–] Windex007@lemmy.world 11 points 1 year ago (2 children)
  1. It isn't and has never been a truth machine, and while it may have performed worse with the question "is 10777 prime" it may have performed better on "is 526713 prime"

ChatGPT generates responses that it believes would "look like" what a response "should look like" based on other things it has seen. People still very stubbornly refuse to accept that generating responses that "look appropriate" and "are right" are two completely different and unrelated things.

[–] deweydecibel@lemmy.world 3 points 1 year ago* (last edited 1 year ago) (1 children)

In order for it to be correct, it would need humans employees to fact check it, which defeats its purpose.

[–] Windex007@lemmy.world 1 points 1 year ago

It really depends on the domain. Asking an AI to do anything that relies on a rigorous definition of correctness (math, coding, etc) then the kinds of model that chatGPT just isn't great for that kinda thing.

More "traditional" methods of language processing can handle some of these questions much better. Wolfram Alpha comes to mind. You could ask these questions plain text and you actually CAN be very certain of the correctness of the results.

I expect that an NLP that can extract and classify assertions within a text, and then feed those assertions into better "Oracle" systems like Wolfram Alpha (for math) could be used to kinda "fact check" things that systems like chatGPT spit out.

Like, it's cool fucking tech. I'm super excited about it. It solves pretty impressively and effiently a really hard problem of "how do I make something that SOUNDS good against an infinitely variable set of prompts?" What it is, is super fucking cool.

Considering how VC is flocking to anything even remotely related to chatGPT-ish things, I'm sure it won't be long before we see companies able to build "correctness" layers around systems like chatGPT using alternative techniques which actually do have the capacity to qualify assertions being made.

[–] oktoberpaard@feddit.nl 0 points 1 year ago

That’s not necessarily true: https://arstechnica.com/google/2023/06/googles-bard-ai-can-now-write-and-execute-code-to-answer-a-question/. If the question gets interpreted correctly and it manages to write working code to answer it, it could correctly answer questions that it has never seen before.

[–] spiderman@ani.social 1 points 1 year ago* (last edited 1 year ago)

I think that there is another cause. Remember the screenshots of users correcting chatgpt wrongly? I mean chatgpt takes user's inputs for it's benefit and maybe too much of these wrong and funny inputs and chatgpt's own mistake of not regulating what it should take in and what it should not might be an additional reason here.

[–] ZagTheRaccoon@reddthat.com 1 points 1 year ago

They are lobotomizing the softwares ability to provide bad PR answers which is having cascading effects via a skewed data set.

[–] james1@lemmy.world 7 points 1 year ago* (last edited 1 year ago) (5 children)

It's a machine learning chat bot, not a calculator, and especially not "AI."

Its primary focus is trying to look like something a human might say. It isn't trying to actually learn maths at all. This is like complaining that your satnav has no grasp of the cinematic impact of Alfred Hitchcock.

It doesn't need to understand the question, or give an accurate answer, it just needs to say a sentence that sounds like a human might say it.

[–] Gazumbo@lemmy.world 2 points 1 year ago

If it's trying emulate a human then it's spot on. I suck at maths.

[–] bric@lemm.ee 2 points 1 year ago (1 children)

This. It is able to tap in to plugins and call functions though, which is what it really should be doing. For math, the Wolfram alpha plugin will always be more capable than chatGPT alone, so we should be benchmarking how often it can correctly reformat your query, call Wolfram alpha, and correctly format the result, not whether the statistical model behind chatGPT happens to use predict the right token

[–] Gork@lemm.ee 2 points 1 year ago

It sounds like it's time to merge Wolfram Alpha's and ChatGPT's capabilities together to create the ultimate calculator.

[–] dbilitated@aussie.zone 2 points 1 year ago (1 children)

to be fair, fucking up maths problems is very human-like.

I wonder if it could also be trained on a great deal of mathematical axioms that are computer generated?

[–] Cabrio@lemmy.world 0 points 1 year ago* (last edited 1 year ago) (1 children)

It doesn't calculate anything though. You ask chatgpt what is 5+5, and it tells you the most statistically likely response based on training data. Now we know there's a lot of both moronic and intentionally belligerent answers on the Internet, so the statistical probability of it getting any mathematical equation correct goes down exponentially with complexity and never even approaches 100% certainty even with the simplest equations because 1+1= window.

[–] dbilitated@aussie.zone 1 points 1 year ago

i know it doesn't calculate, that's why I suggested having known correct calculations in the training data to offset noise in the signal?

[–] TimewornTraveler@lemm.ee 1 points 1 year ago (1 children)

so it confidently spews a bunch of incorrect shit, acts humble and apologetic while correcting none of its behavior, and constantly offers unsolicited advice.

I think it trained on Reddit data

[–] cxx@lemmy.world 1 points 1 year ago

acts humble and apologetic

We must be using different Reddits, my friend

[–] R00bot@lemmy.blahaj.zone 1 points 1 year ago

You're right, but at least the satnav won't gaslight you into thinking it does understand Alfred Hitchcock.

[–] solstice@lemmy.world 7 points 1 year ago

GPT was always really bad at math.

I've asked it word problems before and it fails miserably, giving me insane answers that make no sense. For example, I was curious once how many stars you would expect to find in a region of the milky way with a radius of 650 light years, assuming an average of 4 light years per star. The first answer it gave me was like a trillion stars or something, and I asked it if that makes sense to it, a trillion stars in a subset of space known to only contain about a quarter of that number, and it gave me a wildly different answer. I asked it to check again and it gave me a third wildly different number.

Sometimes it doubles down on wrong answers.

GPT is amazing but it's got a long way to go.

[–] orphiebaby@lemmy.world 5 points 1 year ago* (last edited 1 year ago) (1 children)

HMMMM. It's almost like it's not AI at all, but just a digital parrot. Who woulda thought?! /s

To it, everything is true and normal, because it understands nothing. Calling it "AI" is just for compromising with ignorant people's "knowledge" and/or for hype.

[–] Mikina@programming.dev 6 points 1 year ago (1 children)

Exactly. It should be called ML model, because that's what it is, and I'll just keep calling that. Everyone should do that.

[–] orphiebaby@lemmy.world 2 points 1 year ago (1 children)

What does that stand for? O:

You'd think I'd know that since I'm talking about AI; but actually most of my knowledge is about how things work or don't work, not current trends/news.

[–] 0x01@lemmy.ml 4 points 1 year ago (1 children)

ML stands for machine learning

[–] orphiebaby@lemmy.world 1 points 1 year ago

Ah, thanks! ^^

[–] lorcster123@lemmy.world 3 points 1 year ago

I used GPT4 the other day and it worked perfectly for calculating formulas of straight lines on linear-log plots but maybe I was the 2%

[–] RufusLoacker@feddit.it 2 points 1 year ago (2 children)

Why are people using a language model for math problems?

[–] Fixbeat@lemmy.ml 0 points 1 year ago (1 children)

Because it works, or at least it used to. Is there something more appropriate ?

[–] bassomitron@lemmy.world 1 points 1 year ago

I used Wolfram Alpha a lot in college (adult learner, but that was about ~4 years ago that I graduated, so no idea if it's still good). https://www.wolframalpha.com/

I would say that Wolfram appears to probably be a much more versatile math tool, but I also never used chatgpt for that use case, so I could be wrong.

[–] gratux@lemmy.blahaj.zone 0 points 1 year ago (1 children)

It was initially presented as the all-problem-solver, mainly by the media. And tbf, it was decently competent in certain fields.

[–] MeanEYE@lemmy.world 1 points 1 year ago

Problem was it was presented as problem solved which it never was, it was problem solution presenter. It can't come up with a solution, only come up with something that looks like a solution based on what input data had. Ask it to invert sort something and goes nuts.

[–] TheBlue22@lemmy.world 1 points 1 year ago

"AI" taking our jobs and all that huh

[–] spaduf@lemmy.blahaj.zone 0 points 1 year ago (1 children)

My personal pet theory is that a lot of people were doing work that involved getting multiple LLMs in communication. When those conversations were then used in the RL loop we start seeing degradation similar to what’s been in the news recently with regards to image generation models. I believe this is the paper that got everybody talking about it recently: https://arxiv.org/pdf/2307.01850.pdf

[–] teydam@lemmy.world 1 points 1 year ago (2 children)

This is peer-reviewed? they use a line in the discussion which seems relatively unprofessional, telling people to join a 12-step program if they like to use artificial training data.

[–] blue_zephyr@lemmy.world 1 points 1 year ago* (last edited 1 year ago) (1 children)

ArXiv papers are never peer reviewed.

[–] teydam@lemmy.world 1 points 1 year ago
[–] spaduf@lemmy.blahaj.zone 1 points 1 year ago

Not affiliated with the paper in any way. Have just been following the news around it.