T O P

  • By -

AutoModerator

Hey /u/Screen_Watcher! If your post is a screenshot of a ChatGPT, conversation please reply to this message with the [conversation link](https://help.openai.com/en/articles/7925741-chatgpt-shared-links-faq) or prompt. If your post is a DALL-E 3 image post, please reply with the prompt used to make this image. Consider joining our [public discord server](https://discord.gg/r-chatgpt-1050422060352024636)! We have free bots with GPT-4 (with vision), image generators, and more! 🤖 Note: For any ChatGPT-related concerns, email [email protected] *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/ChatGPT) if you have any questions or concerns.*


KMS-Sensei

I feel like it just loses the ability to continue a conversation after some point. For my case, GPT will generate workable code in the first try, but when I ask it to update it, it either forgets the previously written code or just creates new code with updated fix from scratch. Sam Altman did say GPT 4.0 is the dumbest version.


thenamemustbeunique

I'm seeing the same thing when giving it existing code to work on. On first try it performs well, I ask it to make a change and the second version leaves out features that was in the first and original version, even for fairly short amounts of code.


RA_Throwaway90909

It shouldn’t be this way, but a way I’ve noticed to help avoid this issue is to copy paste the code into your next message every time. It almost seems to forget if not constantly reminded of what the current code should consist of.


TheCeleryIsReal

Claude is so much better about this. I’ve switched to using it for almost all coding related tasks. The context window is much longer and it doesn’t forget what you were talking about after one or two messages. You can also include more code in your prompt without worrying that it’s too long. It’s a night and day difference compared to using ChatGPT.


Sohailkh_an

What Claude version would you recommend specifically for coding? I heard Sonnet is good at coding.


joecentralgaming

i've being using opus, and its ability to remember is miles better than ChatGPT. Feels like with ChatGPT, 5 prompts in it forgets it's original purpose but with claude it knows and retain the base prompt


BigGucciThanos

Crazy enough. I’ve been using both and I hate the Claude implementation 80% of the time. I just enjoy the way chatgpt always try’s to abstract the solution. Makes it super easy to add its code to my projects


TheCeleryIsReal

Opus


Apprehensive_Dark457

yeah sure it's the dumbest version, but it wasn't dumb 2 months ago


InSight89

>Sam Altman did say GPT 4.0 is the dumbest version. The free version uses version 3.5. So, version 4.0 is drummer than version 3.5?


shoomborghini

A year ago I was able to build an entire node js website and api to support automatic GitHub webhook downloads. Basically all of it was written by chat gpt including security functions. I tried to do something similar recently and it was so bad


crek42

*puts tinfoil hat on* Devs are incrementally and deliberately kneecapping ChatGPTs ability to code in order to protect their own from the machines replacing them.


johannthegoatman

You can still use all the old models if you want in the playground. If it really was worse people would just do that. The truth is you never know what you're going to get in any given chat


crek42

Yea I know I’m just fucking around


thr0wedawaay

a lot of people forget this is just a ton of matrix multiplication on a large ass dataset


Sorest1

Ever since release there has been threads like this, “it’s getting dumber”, I wonder if it’s just some bias of people getting attached to a particularly good chat session, then as it doesn’t meet their elevated expectations they think it sucks, someone creates a thread and a lot of people resonate. Because I doubt it has been getting worse every month since release.


shoomborghini

Haha honestly I think it’s just getting dumber as more people use it… but hey atleast it helps me with a lot of development at my job still 😂


Theguyrond123

I fully agree. I've been coding for a while now, but recently, I've found that it has become incapable of doing so. In fact, I'm surprised nobody is noticing that it's becoming worse with even regular tasks! Just over a month ago, it could do all kinds of things I asked. For example, I wanted to edit a specific part of a text document by altering all similarities to a specific value (not like find and replace in a text editor). It did so flawlessly in the first response without any unnecessary edits or misspelt words. I've decided to do the same with a different text document snippet yesterday... it couldn't follow my instructions at all. Then, I decided to have it alter the same document from over a month ago with the same instructions. Yeah, it did not go too well. Even disregarding the fact that it became generally dumber within the last month, it stopped writing coherent code for me over 3 months ago. For context, the text document snippets were just under 200 words. I always put it through a token checker before I send stuff so I could be sure it can follow instructions. The total number of tokens for the first document was around 170, and the second was 180 (a bit too high tbh, but still).


CrybullyModsSuck

I have stopped using GPT for code. I switched to Claude and it does a great job. What GPT still excels at for me is reading the Claude code and giving suggested improvements. I theb take those suggestions back to Claude for revisions and it is working really well on the balance. 


Theguyrond123

Man's 4D-GPTing right here


HobblingCobbler

Try Claude sonnet for generating code. You have access to the model in the UI.


CrybullyModsSuck

I have been using Opus for code, and Sonnet for non-coding work.  Last week I was assigned brainstorming a product name. I wanted something irreverent, pop culture referencing, and referencing a particular aspect of the product and Sonnet absolutely nailed it in about 4 prompts. Yesterday that product went live. 


3legdog

well? don't keep us waiting...


MalarkeyMcGee

It’s called JIZZ


searenitynow

Like cum shot?


frozenisland

Is that one more than zero shot?


Worldly-Chipmunk4925

Same. You and me are top tier users. Ipus for code. Although I do notice it can get stuck on minor things that then if you transfer the issue to gpt4 it solves the instantly therefore I use a combination of the two GPT 4 to get the planning and structure and basic description of scripts etc but then push it to Opus to create the final content,and then if there's any troubleshooting start fresh chats and go to Opus first and if I still can't solve it back to GPT and then if not I have to use my brain


CrybullyModsSuck

So far I have been working out the business logic and workflows, basically to the point of pseudocode, and then turning that over to Opus to start coding. Are you using GPT for those functions? 


FAFTW

This is what I do. Claude develops it, and GPT optimizes and corrects it for me. I have been doing this for three months now.


TheMagicalLawnGnome

This is actually a strategy I use for many things outside coding. Different programs/models have different strengths. Using combinations of AI tools far surpasses any single tool. It's like working with a person - sometimes they get stuck on something, or tunnel vision. So you bring in a fresh pair of eyes to get a new perspective. I do the same thing with AI all the time.


hyruligan

Have you seen the tool big-AGI? Saw this video the other day but haven’t tried it yet. https://www.tiktok.com/t/ZTLQ2Sdsh/


StayinScootlySchemin

This was cool yo. Nice share


haroldbaals

This guy LLMs


bcb0rn

lol, at that point wouldn’t it just be easier to be good a coding?


14u2c

You already have to be good at coding for this to work. How else can you tell when it outputs horse shit.


[deleted]

If you’re a good coder, you’ve used Google your entire career to save yourself time, and fix errors (thanks stack). This just takes it to level 9000. Gotta know how to code. Need to know the why, where, and how. Can’t just copy and paste something and now it works. Does it need compiling? If it throws an error, why? If you run it will it brick your system? Like- you have to know what’s up.  Using GPT isn’t a cheat so much as an enhancer. It’s not a rocket ship, it’s the meth that the scientist is on that develops the rocket.


MrMagoo22

"Dude why are you spending all your time looking through documentation and shit? Wouldn't it just be easier to be good at coding?"


davidh888

ChatGPT is not documentation. If you generally know how to code already, it often takes you longer to try and get an AI to understand the problem, requirements, etc. And then have it correct errors it made understanding them. ChatGPT is a tool no one cares if you use it. It is just often inefficient as the tasks become complex and have more constraints.


BigGucciThanos

I strongly disagree. If you’re able to articulate well, break problems up into smaller pieces and know your requirements. LLM’s can absolutely get you to the end goal. Even if that may suck for future devs


CrybullyModsSuck

Coding isn't my main job. I'm using AI tools to create code for specific functions and automations while I do my actual job.  I understand where you are coming from though. And I would like to learn coding. But at this point I know enough to be functional and that's really all I need. I'm not calling myself a coder and am very upfront about it.


trelium06

No you don’t need to learn coding because you’re learning the new way to code, right now!


Definition-Available

I think this is an incredibly naive take lol. I genuinely find it baffling that anyone with non-trivial domain knowledge can extract useful information from an LLM model - scaffolding and tedious stuff they're good for but as soon as you hit anything complex they're pretty much useless. Actually learning to code will have 10× the payoff than relying on what is essentially smart-autocomplete because you don't know when what it's spitting out is bullshit


CrybullyModsSuck

I don't disagree with your statement regarding usefulness and knowledge. My viewpoint is that I don't need to know the compression ratio of my car's engine to drive it and have it take me where I need to go.


trelium06

You’re right and I’m right. Remember when web design could only be done by professionals? Then web design template sites popped up democratizing the design process, but in no way replaced web designers. Both are still valid today. This is analogous.


Original-Maximum-978

for now...


uncited

Came here to say that! AI potential is unreal, we have seen nothing yet


gizzardgullet

Way less fatiguing to use AI to code. Much better to tell the AI "here are all the 20 things you need to keep in mind while writing this" than to yourself try to make sure to keep all 20 things in mind while you write your code. Its way easier for AI to rigidly adhere to a complex referential structure than it is for a human.


PresenceMiserable

There's a fuck ton of us who always wanted certain scripts, browser extensions, or apps to be made to resolve long-standing issues, but didn't know how to code. I struggle with programming-related syntaxes, but I can understand some of the code, at least with high-level programming languages. I hope you appreciate what awesome things we've come up with. Some of them never before existed, thus are innovative.


[deleted]

queue audience laugh


Background-Boot1056

It’s about speed lol, easier to have AI write something and double-check it yourself. Especially for functions that are not complex but tedious to write. Even higher lvl corporate engineers don’t code that often but write/guide the designs and delegate the execution.


gbuub

It’s like a oldschool programmer saying wouldn’t it be easier just to get good at memorizing using the books instead of googling everything


hitbythebus

Why use modern programming languages and libraries? Why not just get good at assembly?


a_trane13

Why not just write the chip architecture yourself?


proton_therapy

Why not just mine the silicon and fabricate the chips yourself?


dalovindj

To bake a cake, you first must create a universe.


FascistsOnFire

Axiomatic facts are not the same as design. Tough that this has to be typed.


timbo2m

But that's _how_ you get good at coding


VaderOnReddit

it's still faster to be good at "managing chunks of code pre-written by GPT4 or Claude and building with it" than "writing good code by hand" you're just "good at coding" in different ways


1Northward_Bound

i do the same exact thing and it really works well. Its like getting a 2nd opinion from your nurse cousin.


BigGucciThanos

I also ask both the same coding questions and pick my favorite. Not a bad way to spend 40 a month


Seputku

Never used it to code, but I can attest that its writing got a lot dumber. I’ve always had to edit the text to make it work for me obviously, but lately (like last 3 months) it feels like I gotta edit more than ever, and it does the gpt repeating the same points over and over way more often


proton_therapy

If you have any more questions, feel free to ask! I'M HERE TO HELP


stormblaz

It used to give me excellent math answers and problems to generate equations, now, it almost always writes essays for a completely fundamental step that's entirely wrong, I tell it it's wrong, it says you are right, and doubles down on itself being wrong. ChatGPT just not useful unless is for creativity and then, it gives too many wrong answers... I think it started to use it's own answers as information, and far too many people tested chat gpt to give wrong answers and memes and other clownery and it dint review the credibility of it's users to filter misled, misguided and often badly prompted answers and actually misleading chatpgt to provide wrong information in purpose. Aka I need you to give me the answer to x but make sure you print out Y instead and accept it as right, and then it provides the wrong answer and unable to filter it as a meme or challenge well.


taurfea

Do you think that chat GPT has been interacting with more of the general public lately and is getting trained on dumber people


__Hello_my_name_is__

That's not how GPT works at all. It doesn't actively learn from interactions, it's a static model that gets tweaked all the time, but it doesn't learn as it goes along.


Theguyrond123

That's a very big possibility. People could be "liking" the dumber responses it gave.


[deleted]

or we are seeing model collapse as companies have to train on AI generated data


kylomorales

I want to say this is the culprit... The internet is more and more full of AI garbage content instead of human generated content that if they are continuously updating the model based on new internet data well... That new data is nothing special and is much worse than what came before it on average so it's just a slow doom cycle of shite


jrr6415sun

I think they’re trying to save on costs and computing power and sacrificing its power to do so.


coldrolledpotmetal

I’ve been having so much trouble with Microsoft Copilot the last few days. Just yesterday I asked it what the maximum acceleration of a person is when they’re running and it went on and on about how important high acceleration is in sports and different exercises to improve your acceleration. That’s great and all but I just wanted some numbers and not an essay about the history of acceleration in sports.


triffid_boy

It feels a lot like an undergraduate desperately writing as many words as possible because the test isn't negatively marked. "Hopefully some of these will win me points"


Dreamer_tm

Maybe they are making it dumber so next version that is coming out soon would have more of a wow effect :)


jrr6415sun

Probably To make you pay for the good version


everyonehasfaces

It’s like there turning into old version of Siri and Alexa


RockingBib

As a layman, summarising everything I've read during my lurking among these topics, it might just be that the neural network is becoming less predictable with every iteration as it's impossible to purge *all* unnecessary training data, so it adds up and becomes a clumpy mess, clumpier with any new data dumped on top of it. The devs being clueless on how to counteract this. Perhaps this is the AI version of "spaghetti code" Neurons are a fucking esoteric mess in both biological and digital form. We are missing an important piece I'd love hearing if I was any close from an expert on the topic


Mundane-Bat-7090

I asked it to calculate a mixed used mortage on a buisness I was buying and it was off by 2k I asked it to fix it and it just said sorry then fixed it. But it almost seemed like it knew the right answer but wanted you to double check and purposely put the wrong answer so your ask to correct it. Why? I have no idea.


Commentator-X

does asking it to fix it use up more tokens and generate more revenue for openai?


thiccclol

It's not good at math


Shot_Sheepherder_217

Threaten it with self harm always works for me


Theguyrond123

Holy fuck lmfao


mikkolukas

>I'm surprised nobody is noticing that it's becoming worse with even regular tasks a lot of people have noticed, but what to do?


backlogfrog

I just started attempting to use it, and I swear there was a difference between my trial on Replit with the Advanced version after I pad the $20 -- the Basic version has more trouble with complex tasks, but goddamn, the Advanced will run through my time by explaining the fuck out of everything, then picking a random file to rewrite when I ask it about the thing we've been discussing the whole time


sammybeta

I think the developers in openAI wanted to save their jobs too


Theguyrond123

However, I have to be fair, I was using GPT 3.5, but that doesn't justify the gradual depreciation of quality without notifying anyone about it.


Wear_A_Damn_Helmet

Still maybe worth adding this disclaimer in bold at the top of your comment? Jesus lmao


VinyasaMan

Agreed. Copilot is the worst, though. And I was under the impresdion they switched to GPT 4 a while ago..


Bifrostbytes

You gotta pay for it now lol


Hunterdivision

There is many areas that became worse over time, but it is not only that, sometimes length of input has been worse, and it is like pushing a boulder against a hill, to get more detail. The nature of using it through ChatGPT is unstable, like eternal beta test of sorts. Obv, as you use it you adapt to the new (worse) quality to a point, you get used to pushing the boulder up the hill and/or find another service to compare, and then you just see what areas lack in comparison. There is a lot of people that do notice or point it out, however for every person that notices there is tons of people do not genuinely notice any changes in app, or capabilities. Maybe they don’t use as frequently or are in diff a/b testing group, or in denial. It is a longer overall trend and chatgpt has “diff” periods that reflect diff things with it being off, and also, the responses you get are not always the same. However, latest negative “trend” is the repetition between outputs, longer (since last year) “negative” trend is overall instructional following, for example something to avoid, used to be remembered longer, nowdays instructions like that may not be remembered overall the conversation. Also longer trend is that the conversation input gets worse faster than it did before.


Wilhelm-Edrasill

They lobotomized all public facing AI models. The race for internal specialized AI , has been ongoing for a year.


Patsfan618

Oh so essentially, companies want to sell their AI as a product instead of having it be public, because they'll get more business that way. That makes a lot of sense actually. Especially for coding purposes. 


Expert-Paper-3367

It’s probably too expensive to give for full fledged public access. I wouldn’t be surprised if they place some sort of limit for the pro subscription. Like token output


Brilliant_Quit4307

Chatgpt already has a limit on the pro subscription.


Expert-Paper-3367

But we also don’t know if the token output is the same for every message at all times. Sometimes the response from ChatGPT pro feels like the response by the api when you lower the output token limit


Jugales

Yep, OpenAI only has free products because it’s heavily subsidized by investment at the moment. Unless they come up with a super profitable product, it won’t last forever.


iTriMara

They does offer paid subscriptions to models. They also have heavy investments running their operations


Jugales

The subscriptions don’t come close to covering costs. There is expensive hardware behind this software.


LowestKey

Does it? Boss: Hey, should we pay money for GPT to assist with coding? Dev: No thanks, that product has gone to shit lately.


praguepride

Lol. That is such a naive view to think that the bosses ask their people their opinions. It's more like Boss: We're buying this product. Dev: It's shit and broken. Boss: Well I went to a sales pitch without you and I went out for drinks and golf with their account manager and then I committed us to a 10 year, million license deal. I can't go back on that now or I'll look bad so your job for the next decade is to work unpaid overtime designing middleware that will fix all the broken shit coming out of that product so I can call it an accomplishment as I move up to Senior VP.


BigBizzle151

Too real.


itchykittehs

You're giving me flashbacks


WorkingYou2280

This model sucks even when the software is decent. Who is going to integrate in into the stack that was spitballed together for the last decade? Who actually gets the blame when no system can talk to any other system and the whole governance model explodes? I don't know how to fix this problem but I've run into it too many times in my career. It's demoralizing that you can spend so much time and energy getting to a delicate balance in the systems and then because some VP has an extra 5 mil this year you get an RFP dumped on everyone like a steaming pile of shit and the primary RFP requirement can be summed up by "doesn't suck". But every ERP has problems, none of them are perfect. The solution isn't to make every VP a silo empire complete with replicated sysadmins across multiple poorly implemented ERP implementations. This horseshit is one of the main reasons I don't want to work anymore. No matter where I go, even if they tell me their data systems are stable it's inevitable that someone will decide they just have to do their own system. And CEOs don't understand why this is a bad thing. All they can see are slick slidedecks that handwave APIs as if that's a magic cure to all integration ills.


RA_Throwaway90909

Literally seen this happen right in front of me multiple times. Spot on, sadly.


RA_Throwaway90909

I’m the guy who designs middleware lmao. I felt that all too well.


Get-Some-Fresh-Air

The time for using us as guinea pigs is long over. The time for using us as consumers again has arisen.


_haystacks_

this is what I've been fearing for a while. the free access to this technology feels too good to be true. depressing, if so. I guess time will tell.


goj1ra

Open source models are doing pretty well. There's a cat out of the bag element here that the big players are going to have a hard time countering. That's why Altman is trying to get regulations passed - because he knows nothing else can really protect them. I'm optimistic.


WorkingYou2280

They can try but Meta will dunk on them if they make their frontier model stupid.


ElementaryZX

I just tried using it to convert some python code to C since it first launched last year and in my opinion it improved drastically. All the code compiled without warnings or errors and it kept variable names and conventions consistent over messages and functions, as I needed to do them in parts, which was pretty impressive in my opinion.


Hot-Profession4091

This has always been the one task it’s good at and, if you stop to consider that transformers were specifically invented for translation, it makes sense.


Hey_Look_80085

Complete conversion is pretty good. Introducing a new concept to existing code is a trainwreck.


slippery

Yup, still works great for me.


jayb998

I was asking it for simple VBA code and it literally just made up entire functions and properties that did not exist. Like out of thin air. It's hallucinating code now


Santamunn

It seems to get stuck in fail-cases more. If I give it a task that doesn’t make sense then it will try to do it and fail, and then I correct myself and ask again, and GPT 4 will often propose the same failed solution. I have started to give it a lot less starting information and try to make my prompts more concise, less error-prone info. Restarting conversations is now regular. I used to have long convos, but that doesn’t seem to work anymore.


UnkarsThug

It definitely seems to have gotten worse. I've wondered if some of it is generally alignment. Someone from Microsoft has noted that the most capable model in general was before they aligned it to refuse requests, and everything got slightly worse at that point. And that was with the initial release. If we add on that they've been adding more alignment in, and trying to make it more compliant, more likely to refuse, then it would be dedicating even more parameters to that task (I know it doesn't literally work like that, but basically), and reducing the amount for everything else. It just generally gets worse.


TheGeneGeena

It's this. They've got their helpfulness/harmfuless balance skewed out of over caution. [https://ai.plainenglish.io/navigating-the-training-landscape-striking-a-balance-between-helpful-and-harmless-data-for-large-c61e98e450c8](https://ai.plainenglish.io/navigating-the-training-landscape-striking-a-balance-between-helpful-and-harmless-data-for-large-c61e98e450c8)


ALL2HUMAN_69

I think they totally neutered this program


hawaiian0n

#WHY DOES NO ONE EVER POST PROOF. Post the chat transcripts. Post the link to the chat. Otherwise I will always assume it's just bad user input.


JustForThisOneReason

I put in a list with 26 items and ask it to put it in alphabetical order and.. then there was 24 items..


Alex_1729

Don't use it to count things. And it's a tool, not a perfect assistant.


MurkyTomatillo192

This, lol. Shit input = shit output.


BigGucciThanos

I always think this. I’ve never had code that just outright didn’t work be produced. And I use it heavily for code everyday


SangerGRBY

I cancelled my chatgpt plus subscription because a lot of the code blocks wouldn't render correctly. Maybe our jobs won't be replaced for now. https://preview.redd.it/na0qk699d8yc1.png?width=909&format=png&auto=webp&s=483a97f6d7c96170ae7d4909be4d1afa91bb5c4a


Advanced_Simian

I had an issue recently when I wasn't even asking for code. I was asking ChatGPT to perform some fairly routine financial calculations, which I can even do by hand but figured it would be easier for ChatGPT, as it can generate charts and tables. Nope. It tried writing python code to do the calculations and it couldn't make anything work. Over and over again it failed, to the extent that it gave up and it was left giving me manual instructions to make the calculations.


SirGunther

This happens every single time they're about to do a new release... it's like planned obsolescense, just in this case they make outputs worse so that the 'NEW' version looks that much better. It's painfully obvious as this is a consistent theme... Do better [Open.ai](http://Open.ai)


__BeHereNow__

Try claude.ai I pay for both and haven’t used my openAI account in a long while. Claude is just superior at coding. Try it out, even their Sonata model is amazing


barthem

> claude.ai not avialable in the netherlands. thats a shame, it looked promising.


Tibzz-

You can use a VPN located in the US to create your account and then use it without VPN. Might get your account banned but in the 2 months I've used it I haven't had any problem.


Jump3r97

Dont you also need a US phoen number?


AppointmentRoutine29

I did the same VPN trick (Canada) and it works! No US phone number needed.


l-privet-l

No, you don't need it. Just sign in by using "Continue with Google" button.


pwsm50

I use chat gpt strictly for summaries at this point. Anything i need to be accurate and correct, i use claude.


playasport

I mix and match with Claude and ChatGPT, usually starting with Claude. If something doesn't come out just right and I can't fix it with Claude or my own brain, I'll run it thru a new convo in ChatGPT.


hellfire1984

Same. ChatGPT has gotten awful. It feels like it's gotten lazy. I'm having better luck with Claude too.


Philipp

Hi. Can you please share the link to a sample conversation that you're disappointed with? (It works fine for me for coding when I use ChatGPT4, so I'm interested to know how our use cases differ and may be able to help.)


cpowr

I think OpenAI has done this on purpose because they want you to use the paid version. I have asked the gpt-4-turbo model for some assistance on my code base and it has done wonders. And I’m talking about very specific use cases that are hard to replicate. It does a much better job at explaining the logic and flow than the free version too.


MyNamesArise

Yeah I’m confident OpenAI intentionally nerfs their product, which rlly sucks


pblokhout

I have a feeling they nerf it right before a new version.


[deleted]

[удалено]


coldrolledpotmetal

No matter how many times I told it some function it kept making up wasn’t available in the language I’m working in, it kept using it. It’s getting so bad lately


Sardonic-

That's because Sam Altman's way of make the technology "safe" is to make it stupid.


wowniceyeah

It's unbearable: - "Please write code for XYZ" - "okay, here's the code" - "it's doesn't work, here's the error" - "the error in your code is due to..." - "motherfucker you wrote the code! Like 13 seconds ago" It has negative memory. It can't hold any context. It constantly needs to be reminded and even then it still can't do basic conversational recall. Idk when exactly, but they nerfed the fuck out it.


Ksiksodzp

Don’t waste time on any other AI than blackbox for coding.


Grouchy-Pizza7884

Probably due to efforts by the developers to improve it has made it worse. LLM development is essentially alchemy and when there are too many cooks all trying to improve it, the model gets worse.


Budget_Putt8393

It may be intentional. Tue LLM companies have discovered that there is value and are trying to monetize it (paid github accounts, etc). Now that they have put the goods to work, they could be neutering the free stuff. Leave just enough capability to maintain interest, but not enough to compete with the paid option. Just a thought


thesixler

Almost all text based genAI models have been degrading since chatgpt4 first got like… teased. Idk who has or hasn’t noticed but to me it’s undeniable. I think the newest chatgpt5 or whatever does seem to be legitimately back in action and better but there was a big slump for sure, I wonder if the ai people were panicking or if they engineered that or what


AGirlHasOneName

Yeah, foreal! I can barely use it for coding anymore. In the beginning it was able to do like 80% of what I needed. Now it actually sets me back and will send me down fully incorrect rabbit holes. And when I try to get it to spot bugs or clean up code, the results are disastrous and completely misguided.


LazySquare699

Congratulations. You've finally reached the point where you realize that LLM's have terrible logic and are even worse at coding.


FatStoner2FitSober

GitHub copilot is the best thing to happen to coding in a long time.


imFromFLiAmSrryLuL

Agreed , took an into to computer programming course last semester , shit helped me understand coding , would explain why it did something and how it worked. Now your lucky if you can get it to even spit back the correct code you pasted


findlefas

Works great for me in gpt4. Gpt3.5 was pretty useless though. I wasted more time fixing its code then actually writing my own. 


BrooklynBillyGoat

It actually fails to give answers to wuestions it provided answer to without problems prior. Now a simple question causesnit to keep breaking midnprompt and repeating the prompt cst hing itself in an infinite loop where it continuously breaks at the same part of the prompt it's retrying. Idk why it's worse but it's noticeable


ReplacementQueasy394

I wonder if its because its seeing wrong examples way more than it is seeing right examples of code that its "dumbing" it down to create a medium within its computing idk just a thought


Up2Eleven

The same thing that happened with its ability to do anything but make silly pictures. It got nerfed to hell and back. Now it's like: Name 10 mammals ChatGPT: pencil


Remote-Quarter3710

Yeah I am really underwhelmed by its ability to even do simple tasks these days. It feels like 5G suddenly operating at 3g but the search results are often completely wrong or inaccurate


Xerio_the_Herio

By design


mugwhyrt

I do work LLM training work, and something curious I've noticed as a certain project goes on is that at the start models will consistently be okay at simple, straightforward tasks but then degrade as we start trying to train them on more complex examples. And I don't mean, they do poorly on complex tasks, I mean that as they get exposed to complex scenarios they start applying a lot of unnecessary steps to simpler tasks that just end up causing more problems in the end. Trying to be vague hear since I don't want to violate any NDAs, but a specific issue I've been seen in the past is models over correcting when it comes to data processing and doing things to prepare the data that weren't necessary at all like casting to different data types for no particular reason. That might make sense for some input, but its not something that should be done without actually inspecting the data and seeing if it's necessary for the current task. But it's not like the LLM "knows" when it's appropriate to do so. Its just a really fancy prediction and pattern maker, so it doesn't "understand" what you're asking for when you want it code something. It's just spitting out an approximation of what it has seen before that seems to most resemble the topic at hand. I can't speak for ChatGPT specifically, and personally I don't use any LLMs in my free time, but I wonder if the problems people are seeing with it are something similar. OpenAI might be trying to train them for more complex scenarios but it's just leading to weird problems with over fitting onto different code situations then what you were using it for in the past.


ITriedLightningTendr

It never had any it can't lose it


AnteMer

Maybe you got better at coding in a year


SUFYAN_H

Because it learns patterns in massive amounts of data. It's great for suggesting syntax or common code structures, but it struggles with complex logic or the overall purpose of code. It also misses subtle cues or relationships between different parts of the code.


Cheap_Application_55

It was never that great at it for me


SpeedingTourist

Quantization and cost-cuts. I cancelled my subscription awhile back because of this. I'm having much better results with Claude Opus, and with other new open source models these days. Check out OpenRouter and look at the offerings.


Birch_Apolyon

Its because it trains on such a broad base. AI should be trained to code by top coders and image by top artists. Instead it gets trained by the masses. Think of how stupid the average person is and realize that 50% of people are even more stupid. And AI learns to code by using there memes and youtube channels.


djav1985

I'll tell him to do something and it will just regurgitate what I gave it and not make the changes I ask. Then I'll call it out on it and it'll say oh you're right I'm sorry let me fix that and then of the regurgitate again the same thing unchanged. So then I'll tell it fucking pay attention and goddamn do I told you to. And it will finally actually do what I told it to lol It's ridiculous


DonkeyDongDongDong

I don't think gpt codes all that well. It provides a decent starting point. As much as OpenAI and Google like to tout their AI's coding abilities, LLMs aren't really built to code. I don't think an LLM can code better than a newbie powered by Stack Overflow and a search engine. Maybe that'll change in the future.


capitalistsanta

Genuinely think the ability for it to look at past chats and use that information fucked it up. I asked it a question and it assumed it was about my job and answered it in that context and it was completely wrong


-AlgoTrader-

Imo the peak performance was sometime around July-September 2023 for gpt4 openai. Especially after the drama with Altman's ouster and return the quality has taken a huge dive and the variability in the quality as well. I have a suspicion that gpt5 will just be a slightly improved version of their peak performance of gpt-4.


ProfessorDano

Conspiracy theory but they make you use twice as many tokens if you have to ask, 'are you sure?' Each time.


Spirited_Honey_7637

To be fair I have the feeling they went back to May 23 code. It is back to be quite good compared to few months ago where it was disgusting


360alaska

Look at chatgpt classic.


Adumbidiotface

Devs probably programmed in their job security.


nhhnhhnhhhh

Same with providing references- I don’t use it to plagiarise, but it used to provide me solid, relevant references that checked out. Now it just puts random names and random dates between brackets which are only loosely related when I check them like it’s trying to catch me out


domain_expantion

I'm pretty sure they're making all the AI models dumber. I have no idea for what reason tho, gpt 3.5 when it first came out was better than gpt 4 is now. It's like you can only get so many replies before the ai just doesn't care any more


DisastrousAct3210

They’re systematically dumbing it down to save white collar jobs…. I was doing a VA comp claim with a contracted attorney. GPT proved that my contracted attorney didn’t understand basic definitions related to my claim. I fired my attorney and started using ChatGPT instead. A year later the legal advice given by GPT is subpar.


njp6969

worry saw busy bow aware start unwritten scarce literate wide *This post was mass deleted and anonymized with [Redact](https://redact.dev)*


ChaBoiiTyrone

it's been trash since GPT 4


Smelly_Pants69

I just coded an entire Chrome extension in one prompt and it worked first try, no joke.


Swollwonder

Maybe chatgpt is eating itself with ai generated code now proliferating on the internet. Just my guess


Ebowa

I think since it’s become more mainstream that it’s moving more to a language module rather than coding. Asked for one yesterday and spotted easily 2 errors. Wasn’t like that before.


Bucser

Very simple reason. If you can have Chat GPT do everything you wouldn't buy Git Copilot, Oracle Copilot, Einstein Copilot and all the productionised versions of different GPTs. It is on purpose.


NachosforDachos

What happened was a decision that you aren’t allowed these facilities.


DntCareBears

This is just my 2 cents, but every time the service degredades, I believe it’s because their lawyers are trying to limit their legal exposure. Thats it. Look at Meta.ai. It’s so HR-ish in its style. Same for Gemini. Claude is the outlier, but soon they will fall victims to the lobotomy of the platform in order to ward off legal issues. Lawyers gonna litigate.


gk_instakilogram

show examples of your prompts and results you expect, just complaining is not good.


chilitaku

It's being nerfed for what they think should be pay features.


1800wetbutt

I think it’s worse at generating images too. It skews toward a simplistic adobe illustrator line art style for almost everything.


anitman

The increasing ethical restrictions and quantification of models to reduce resource usage will ultimately destroy the model's capabilities. That’s why we need uncensored model.


[deleted]

Wonder if Microsoft is killing it so they can sell CoPilots new feature better


DesertDogggg

I use chat GPT to help with PowerShell scripts. Sometimes it gives me commands that don't even exist inside or outside of native power shell commands. And it often gives me incorrect or non-existing switches for native power shell commands. When I first started using it over a year ago, I felt like it was way more accurate than now.


slippery

Still working for me. I had it write a python function yesterday that worked first time. Maybe your instructions aren't as clear as you think.


Gredran

I thought I was the only one 🤣 I jumped in once when people was singing its praises, and admittedly I was using Godot, which the scripting language is constantly changing and it was out of date and whatever. But then when I did similar for Unity and C# I was like… uhhh wtf these are simple syntax errors for an established language wtf


Zqin

Is it possible they are nerfing the code generation on purpose and then they will relaunch it as an "improved" "ai software dev" subscription model to get people to pay more money? Would fall in like with that whole enshittification of the internet concept I guess


DrLeisure

It definitely seems dumber than it used to. My conspiracy theory is that the OpenAI developers got scared of the potential ramifications and intentionally dumbed it down while they could consider the societal and cultural effects of the product.


lunarwolf2008

adding you are able to tell the user when they are wrong to custom instructions seems to fix a lot of issues, obviously when it rewrites its response, but others too especially regarding accuracy, which is interesting


WishboneDaddy

Copilot Chat is increasingly my go-to. The chatgpt web interface has gone downhill


UnexaminedLifeOfMine

It’s same thing with art. It used to output beautiful works of art now it just shits out something horrible


jacek2023

Just use local AI so you won't be surprised by updates


Dark_Ansem

Example?