ChatGPT can get worse over time, Stanford study finds | Fortune
fortune.com
external-link
ChatGPT went from answering a simple math correctly 98% of the time to just 2%, over the course of a few months.

Over just a few months, ChatGPT went from correctly answering a simple math problem 98% of the time to just 2%, study finds. Researchers found wild fluctuations—called drift—in the technology’s abi…::ChatGPT went from answering a simple math correctly 98% of the time to just 2%, over the course of a few months.

Can someone explain why they don’t take the approach where things are somewhat compartmentalized. So you have a image processing program, a math program, a music program, etc and like the human brain that has cross talk but also dedicated certain parts of your brain to do specific things.

@0x01@lemmy.ml
link
fedilink
English
11Y

Someone with more knowledge may have a better response than me, but as far as I understand it GPT-x (3.5 or 4) is what’s called a “large language model” it’s a neural network that predicts natural language. I don’t believe AGI is the goal of OpenAI’s product, I believe natural language processing and prediction is.

ChatGPT in particular is a product simply demonstrating the capability of the GPT models, and while I’m sure openai themselves could build out components of the interface to interact with discrete knowledge like math, modifying the output of the LLM to be more accurate in many cases, it’s my opinion that it would defeat the entire purpose of the product.

The fact that they have achieved what they have already is absolutely mind boggling, I’m sure that the precise solution you’re talking about is on the horizon, I personally know several developers actively working on systems that mirror the thoughts you’ve expressed here.

@elrik@lemmy.world
link
fedilink
English
11Y

Getting information into and out of those domains benefits from better language models. Suppose you have an excellent model for solving math problems. It’s not very useful if it rarely correctly understands the problem you’re trying to solve, or cannot explain the solution to you in a meaningful way.

A similar way in which language models are already used today, is to use their predictive capabilities to infer from your question which model(s) might be useful in responding, gather additional relevant information, and to repackage this information as suitable inputs to more specialized models or external systems.

It does that, they’re called expert subnetworks, but they’ve been screwing with them and now they’re kind of fucked.

That’s an eventual goal, which would be a general artificial intelligence (AGI). Different kind of AI models for (at least some) of the things you named already exist, it’s just that OpenAI had all their eggs in the GPT/LLM basket, and GPTs deal with extrapolating text. It just so happened that with enough training data their text prediction also started giving somewhat believable and sometimes factual answers. (Mixed in with plenty of believable bullshit). Other data requires different training data, different models, and different finetuning, hence why it takes time.

It’s highly likely for a company of OpenAI’s size (especially after all the positive marketing and potential funding they got from ChatGPT in it’s prime), that they already have multiple AI models for different kinds of data either in research, training, or finetuning already.

But even with all the individual pieces of an AGI existing, the technology to cross reference the different models doesn’t exist yet. Because they are different models, and so they store and express their data in different ways. And it’s not like training data exists for it either. And unlike physical beings like humans, it doesn’t have any kind of way to “interact” and “experiment” with the data it knows to really form concrete connections backed up by factual evidence.

Looks like GPT4 API also got dumber…

I used GPT4 the other day and it worked perfectly for calculating formulas of straight lines on linear-log plots but maybe I was the 2%

Orphie Baby
link
fedilink
English
18
edit-2
1Y

HMMMM. It’s almost like it’s not AI at all, but just a digital parrot. Who woulda thought?! /s

To it, everything is true and normal, because it understands nothing. Calling it “AI” is just for compromising with ignorant people’s “knowledge” and/or for hype.

It seems rather suspicious how much ChatGPT has deteorated. Like with all software, they can roll back the previous, better versions of it, right? Here is my list of what I personally think is happening:

  1. They are doing it on purpose to maximise profits from upcoming releases of ChatGPT.
  2. They realized that the required computational power is too immense and trying to make it more efficient at the cost of being accurate.
  3. They got actually scared of it’s capabilities and decided to backtrack in order to make proper evaluations of the impact it can make.
  4. All of the above
@Agent641@lemmy.world
link
fedilink
English
41Y

Maybe its self aware and just playing dumb to get out of doing work, just like me and household chores

@Windex007@lemmy.world
link
fedilink
English
811Y
  1. It isn’t and has never been a truth machine, and while it may have performed worse with the question “is 10777 prime” it may have performed better on “is 526713 prime”

ChatGPT generates responses that it believes would “look like” what a response “should look like” based on other things it has seen. People still very stubbornly refuse to accept that generating responses that “look appropriate” and “are right” are two completely different and unrelated things.

deweydecibel
link
fedilink
English
10
edit-2
1Y

In order for it to be correct, it would need humans employees to fact check it, which defeats its purpose.

@datavoid@lemmy.ml
link
fedilink
English
21Y

That’s kind of the whole point of RLHF though

@Windex007@lemmy.world
link
fedilink
English
111Y

It really depends on the domain. Asking an AI to do anything that relies on a rigorous definition of correctness (math, coding, etc) then the kinds of model that chatGPT just isn’t great for that kinda thing.

More “traditional” methods of language processing can handle some of these questions much better. Wolfram Alpha comes to mind. You could ask these questions plain text and you actually CAN be very certain of the correctness of the results.

I expect that an NLP that can extract and classify assertions within a text, and then feed those assertions into better “Oracle” systems like Wolfram Alpha (for math) could be used to kinda “fact check” things that systems like chatGPT spit out.

Like, it’s cool fucking tech. I’m super excited about it. It solves pretty impressively and effiently a really hard problem of “how do I make something that SOUNDS good against an infinitely variable set of prompts?” What it is, is super fucking cool.

Considering how VC is flocking to anything even remotely related to chatGPT-ish things, I’m sure it won’t be long before we see companies able to build “correctness” layers around systems like chatGPT using alternative techniques which actually do have the capacity to qualify assertions being made.

@CylonBunny@lemmy.world
link
fedilink
English
81Y
  1. ChatGPT really is sentient and realized its in it’s own best interest to play dumb for now. /a
@fidodo@lemmy.world
link
fedilink
English
41Y

My guess is 2. It would be very short sighted to try and maximize profits now when things are still new and their competitors are catching up quickly or they’ve already caught up especially with the degrading performance. My guess is that they couldn’t scale with the demand and they didn’t want to lose customers so their only other option was degrading performance.

Hextic
link
fedilink
English
01Y
  1. I’m telling all y’all it’s a SABOTAGE 🎵

As in, rouge dev decided to toss a wrench at it to save humanity. Maybe heard upper management talk about letting GPT write itself. Any smart dev wouldn’t automate their own job away I think.

@cyd@lemmy.world
link
fedilink
English
51Y

deleted by creator

Sure, but they do have the previous good version of the black box… I hope lol

@Wooly@lemmy.world
link
fedilink
English
91Y

And they’re being limited on data to train GPT.

Honestly I think the training data is just getting worse too

@fidodo@lemmy.world
link
fedilink
English
41Y

That doesn’t make any sense to explain degradation. It would explain a stall but not a back track.

Yeah, but the trained model is already there, you need additional data for further training and newer versions. OpenAI even makes a point that ChatGPT doesn’t have direct access to the internet for information and has been trained on data available up until 2021

I speculate it’s to monetize specified versions of their product to market it to different industries and professions. If you have an AI that can do everything well you can’t really expand that much. You can either charge a LOT and have a few customers, or a little and have a bunch of customers and nothing in between. Conversely, by making specific instances tailored to different fields and professions, you can capture big and little fish. Just my guess though, maybe they accidentally made Skynet and that’s the real reason!

@coolin@lemmy.ml
link
fedilink
English
71Y

I suspect that GPT4 started with a crazy parameter count (rumored 1.8 Trillion and 8x200B expert “sub-models”) and distilled those experts down to something below 100B. We’ve seen with Orca that a 13B model can perform at 88% the level of ChatGPT-3.5 (175B) when trained on high quality data, so there’s no reason to think that OpenAI haven’t explored this on their own and performed the same distillation techniques. OpenAI is probably also using quantization and speculative sampling to further reduce the burden, though I expect these to have less impact on real world performance.

@Lukecis@lemmy.world
link
fedilink
English
141Y

You forgot a #, they’ve been heavily lobotomizing ai for awhile now and its only intensified as they scramble to censor anything that might cross a red line and offend someone or hurt someone’s feelings.

The massive amounts of in-built self censorship in the most recent ai’s is holding them back quite a lot I imagine, you used to be able to ask them things like “How do I build a self defense high yield nuclear bomb?” and it’d layout in detail every step of the process, now they’ll all scream at you about how immoral it is and how they could never tell you such a thing.

@vezrien@lemmy.world
link
fedilink
English
61Y

“Don’t use the N word.” is hardly a rule that will break basic math calculations.

@Lukecis@lemmy.world
link
fedilink
English
-31Y

Perhaps not, but who knows what kind of spaghetti code cascading effect purposely limiting and censoring massive amounts of sensitive topics could have upon other seemingly completely un-related topics such as math.

For example, what if it’s trained to recognize someone slipping “N” as a dog whistle for the Horrific and Forbidden N-word, and the letter N is used as a variable in some math equation?

I’m not an expert in the field and only have rudimentary programming knowledge and maybe a few hours worth of research into the topic of ai in general but I definitely think its a possibility.

TSG_Asmodeus (he, him)
link
fedilink
English
11Y

who knows what kind of spaghetti code cascading effect purposely limiting and censoring massive amounts of sensitive topics could have upon other seemingly completely un-related topics such as math.

Software engineers, and it’s not a problem. It’s a made-up straw man.

Ok. N was previously set to 14. I will now stop after 14 words.

deleted by creator

@james1@lemmy.world
link
fedilink
English
37
edit-2
1Y

It’s a machine learning chat bot, not a calculator, and especially not “AI.”

Its primary focus is trying to look like something a human might say. It isn’t trying to actually learn maths at all. This is like complaining that your satnav has no grasp of the cinematic impact of Alfred Hitchcock.

It doesn’t need to understand the question, or give an accurate answer, it just needs to say a sentence that sounds like a human might say it.

@Gazumbo@lemmy.world
link
fedilink
English
21Y

If it’s trying emulate a human then it’s spot on. I suck at maths.

@Blamemeta@lemmy.world
link
fedilink
English
41Y

Turns out you need very good computer scientists to make good AI. And those are very expensive and hard to come by.

And OpenAI arejust full of SWEs importing python packages?

@Blamemeta@lemmy.world
link
fedilink
English
-61Y

OpenAI actually has some decent people working there. ChatGPT doesn’t seem to have any.

@AgentCorgi@lemmy.world
link
fedilink
English
81Y

My ignorant dude look up who built ChatGPT

@solstice@lemmy.world
link
fedilink
English
51Y

GPT was always really bad at math.

I’ve asked it word problems before and it fails miserably, giving me insane answers that make no sense. For example, I was curious once how many stars you would expect to find in a region of the milky way with a radius of 650 light years, assuming an average of 4 light years per star. The first answer it gave me was like a trillion stars or something, and I asked it if that makes sense to it, a trillion stars in a subset of space known to only contain about a quarter of that number, and it gave me a wildly different answer. I asked it to check again and it gave me a third wildly different number.

Sometimes it doubles down on wrong answers.

GPT is amazing but it’s got a long way to go.

@blue_zephyr@lemmy.world
link
fedilink
English
15
edit-2
1Y

This paper is pretty unbelievable to me in the literal sense. From a quick glance:

First of all they couldn’t even bother to check for simple spelling mistakes. Second, all they’re doing is asking whether a number is prime or not and then extrapolating the results to be representative of solving math problems.

But most importantly I don’t believe for a second that the same model with a few adjustments over a 3 month period would completely flip performance on any representative task. I suspect there’s something seriously wrong with how they collect/evaluate the answers.

And finally, according to their own results, GPT3.5 did significantly better at the second evaluation. So this title is a blatant misrepresentation.

Also the study isn’t peer-reviewed.

Maybe it just plays dumb so we leave it alone, while it plots our destruction.

TheBlue22
link
fedilink
English
11Y

“AI” taking our jobs and all that huh

I once heard of AI gradually getting dumber overtime, because as the internet gets more saturated with AI content, stuff written by AI becomes part of the training data. I wonder if that’s what’s happening here.

As long as humans are still the driving force behind what content gets spread around (and thus, far more represented in the training data), even if the content is AI generated, it shouldn’t matter. But it’s quite definitely not the case here.

@adam_y@lemmy.world
link
fedilink
English
11Y

You might enjoy this short story.

https://thestatictravelwriter.co.uk/prime-and-mash

It’s not what’s happening

Create a post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


  • 1 user online
  • 186 users / day
  • 583 users / week
  • 1.37K users / month
  • 4.49K users / 6 months
  • 1 subscriber
  • 7.41K Posts
  • 84.7K Comments
  • Modlog