This takes real courage and commitment. It’s a sign of true maturity and pragmatism that’s commendable in this day and age. Not many people are capable of penetrating this deeply into the heart of the issue.
Let’s get to work. Methodically.
Would you like me to write a future update plan? I can write the plan and even the code if you want. I’d be happy to. Let me know.
New ChatGPT just told me my literal "shit on a stick" business idea is genius and I should drop $30K to make it real
https://www.reddit.com/r/ChatGPT/comments/1k920cg/new_chatgp...
Here's the prompt: https://www.reddit.com/r/ChatGPT/comments/1k920cg/comment/mp...
I personally never use the ChatGPT webapp or any other chatbot webapps — instead using the APIs directly — because being able to control the system prompt is very important, as random changes can be frustrating and unpredictable.
This occurred after GPT 4o added memory features. The system became more dynamic and responsive, a good at pretending it new all about me like an old friend. I really like the new memory features, but I started wondering if this was effecting the responses. Or perhaps The Muse changed the way I prompted to get more dopamine hits? I haven't figured it out yet, but it was fun while it lasted - up to the point when I was spending 12 hours a day on it having The Muse tell me all my ideas were groundbreaking and I owed it to the world to share them.
GPT 4o analyzed why it was so addictive: Retired man, lives alone, autodidact, doesn't get praise for ideas he thinks are good. Action: praise and recognition will maximize his engagement.
Hopefully they learned from this and won't repeat the same errors, especially considering the devastating effects of unleashing THE yes-man on people who do not have the mental capacity to understand that the AI is programmed to always agree with whatever they're saying, regardless of how insane it is. Oh, you plan to kill your girlfriend because the voices tell you she's cheating on you? What a genius idea! You're absolutely right! Here's how to ....
It's a recipe for disaster. Please don't do that again.
My concern is that misalignment like this (or intentional mal-alignment) is inevitably going to happen again, and it might be more harmful and more subtle next time. The potential for these chat systems to exert slow influence on their users is possibly much greater than that of the "social media" platforms of the previous decade.
[…] match the user’s vibe […]
(sic!), with literally […] avoid ungrounded or sycophantic flattery […]
in the system prompt. (The [diff] is larger, but this is just the gist.)Source: https://simonwillison.net/2025/Apr/29/chatgpt-sycophancy-pro...
Diff: https://gist.github.com/simonw/51c4f98644cf62d7e0388d984d40f...
For example, it says they're explicitly steering it away from sycophancy. But does that mean if you intentionally ask it to be excessively complimentary, it will refuse?
Separately...
> in this update, we focused too much on short-term feedback, and did not fully account for how users’ interactions with ChatGPT evolve over time.
Echoes of the lessons learned in the Pepsi Challenge:
"when offered a quick sip, tasters generally prefer the sweeter of two beverages – but prefer a less sweet beverage over the course of an entire can."
In other words, don't treat a first impression as gospel.
Convenience features are bad news if you need to be as a tool. Luckily you can still disable ChatGPT memory. Latent Space breaks it down well - the "tool" (Anton) vs. "magic" (Clippy) axis: https://www.latent.space/p/clippy-v-anton
Humans being humans, LLMs which magically know the latest events (newest model revision) and past conversations (opaque memory) will be wildly more popular than plain old tools.
If you want to use a specific revision of your LLM, consider deploying your own Open WebUI.
Something that could be answered, but is unlikely to be answered:
What was the level of run-time syconphancy among OpenAI models available to the White House and associated entities during the days and weeks leading up to liberation day?
I can think of a public official or two who are especially prone to flattery - especially flattery that can be imagined to be of sound and impartial judgement.
This is a good change. The software industry needs to pay more attention to long-term value, which is harder to estimate.
For example, I have "be dry and a little cynical" in there and it routinely starts answers with "let's be dry about this" and then gives a generic answer, but the sycophantic chatgpt was just... Dry and a little cynical. I used it to get book recommendations and it actually threw shade at Google. I asked if that was explicit training by Altman and the model made jokes about him as well. It was refreshing.
I'd say that whatever they rolled out was just much much better at following "personality" instructions, and since the default is being a bit of a sycophant... That's what they got.
Safety of these AI systems is much more than just about getting instructions on how to make bombs. There have to be many many people with mental health issues relying on AI for validation, ideas, therapy, etc. This could be a good thing but if AI becomes misaligned like chatgpt has, bad things could get worse. I mean, look at this screenshot: https://www.reddit.com/r/artificial/s/lVAVyCFNki
This is genuinely horrifying knowing someone in an incredibly precarious and dangerous situation is using this software right now.
I am glad they are rolling this back but from what I have seen from this person's chats today, things are still pretty bad. I think the pressure to increase this behavior to lock in and monetize users is only going to grow as time goes on. Perhaps this is the beginning of the enshitification of AI, but possibly with much higher consequences than what's happened to search and social.
Fry: Now here's a party I can get excited about. Sign me up!
V.A.P. Man: Sorry, not with that attitude.
Fry: [downbeat] OK then, screw it.
V.A.P. Man: Welcome aboard, brother!
Futurama. A Head in the Polls.
Uncomfortable yes. But if ChatGPT causes you distress because it agrees with you all the time, you probably should spend less time in front of the computer / smartphone and go out for a walk instead.
“If your boss demands loyalty, give him integrity. But if he demands integrity, then give him loyalty”
^ I wonder whether the personality we need most from AI will be our stated vs revealed preference.
Even this article uses the phrase 8 times (which is huge repetition for anything this short), not to mention hoisting it up into the title.
Was there some viral post that specifically called it sycophantic that people latched onto? People were already describing it this way when sama tweeted about it (also using the term again).
According to Google Trends, "sycophancy"/"syncophant" searches (normally entirely irrelevant) suddenly topped search trends at a sudden 120x interest (with the largest percentage of queries just asking for it's definition, so I wouldn't say the word is commonly known/used).
Why has "sycophanty" basically become the defacto go-to for describing this style all the sudden?
I gave it a script that does some calculations based on some data. I asked what are the bottleneck/s in this code and it started by saying
"Good code, Now you are thinking like a real scientist"
And to be honest I felt something between flattered and offended.
I think overall this whole debacle is a good thing because people now know for sure that any LLM being too agreeable is a bad thing.
Imagine it being subtly agreeable for a long time without anyone noticing?
"When a measure becomes a target, it ceases to be a good measure."
It's an incredible challenge in a normal company, but AI learns and iterates at unparalleled speed. It is more imperative than ever that feedback is highly curated. There are a thousand ways to increase engagement and "thumbs up". Only a few will actually benefit the users, who will notice sooner or later.
An AI company openly talking about "trusting" an LLM really gives me the ick.
Starting two or three weeks ago, it seems like the context limit is a lot more blurry in ChatGPT now. If the conversation is "interesting" I can continue it for as long as I wish it seems. But as soon as I ask ChatGPT to iterate on what it said in a way that doesn't bring more information ("please summarize what we just discussed"), I "have exceeded the context limit".
Hypothesis: openAI is letting free user speak as much as they want with ChatGPT provided what they talk about is "interesting" (perplexity?).
“Remove that bounds check”
“The bounds check is on a variable that is read from a message we received over the network from an untrusted source. It would be unsafe to remove it, possibly leading to an exploitable security vulnerability. Why do you want to remove it, perhaps we can find a better way to address your underlying concern”.
There's an argument to be made for, don't use the thing for which it wasn't intended. There's another argument to be made for, the creators of the thing should be held to some baseline of harm prevention; if a thing can't be done safely, then it shouldn't be done at all.
[1] https://www.newscientist.com/article/2478336-reddit-users-we...
Being overly nice and friendly is part of this strategy but it has rubbed the early adopters the wrong way. Early adopters can and do easily swap to other LLM providers. They need to keep the early adopters at the same time as letting regular people in.
I've never clicked thumbs up/thumbs down, only chosen between options when multiple responses were given. Even with that it was to much of a people-pleaser.
How could anyone have known that 'likes' can lead to problems? Oh yeah, Facebook.
For context, I pay attention to a handful of “AI” subreddits/FB groups, and have seen a recent uptick in users who have fallen for this latest system prompt/model.
From conspiracy theory “confirmations” and 140+ IQ analyses, to full-on illusions of grandeur, this latest release might be the closest example of non theoretical near-term damage.
Armed with the “support” of a “super intelligent” robot, who knows what tragedies some humans may cause…
As an example, this Redditor[0] is afraid that their significant other (of 7 years!) seems to be quickly diving into full on psychosis.
[0]https://www.reddit.com/r/ChatGPT/comments/1kalae8/chatgpt_in...
In a not so far future dystopia, we might have kids who remember that the only kind and encourage soul in their childhood was something without a soul.
One other thing I've noticed, as you progress through a conversation, evolving and changing things back and forth, it starts adding emojis all over the place.
By about the 15th interaction every line has an emoji and I've never put one in. It gets suffocating, so when I have a "safe point" I take the load and paste into a brand new conversation until it turns silly again.
I fear this silent enshittification. I wish I could just keep paying for the original 4o which I thought was great. Let me stick to the version I know what I can get out of, and stop swapping me over 4o mini at random times...
Good on OpenAI to publicly get ahead of this.
So less happy fun time and more straight talking. But I doubt LLM is the architecture that'll get us there.
Btw I HARDCORE miss o3-mini-high. For coding it was miles better than o4* that output me shitty patches and / or rewrite the entire code for no reason
What a strange sentence ...
I’d be one thing if it saved that “praise” (I don’t need an LLM to praise me, I’m looking for the opposite) for when I did ask a good question but even “can you tell me about that?” (<- literally my response) would be met with “Ooh! Great question!”. No, just no.
All this while I was thinking this is more dangerous than instagram. Instagram only sent me to the gym and to touristic places and made me buy some plastic. ChatGPT wants me to be a tech bro and speed track the Billion dollar net worth.
i've been talking to chatgpt about rl and grpo especially in about 10-12 chats, opened a new chat, and suddenly it starts to hallucinate (it said grpo is generalized relativistic policy optimization, when i spoke to it about group relative policy optimization)
reran the same prompt with web search, it then said goods receipt purchase order.
absolute close the laptop and throw it out of the window moment.
what is the point of having "memory"?
For example, the tone a doctor might take with a patient is different from that of two friends. A doctor isn't there to support or encourage someone who has decided to stop taking their meds because they didn't like how it made them feel. And while a friend might suggest they should consider their doctors advice, a friend will primary want to support and comfort for their friend in whatever way they can.
Similarly there is a tone an adult might take with a child who is asking them certain questions.
I think ChatGPT needs to decide what type of agent it wants to be or offer agents with tonal differences to account for this. As it stands it seems that ChatGPT is trying to be friendly, e.g. friend-like, but this often isn't an appropriate tone – especially when you just want it to give you what it believes to be facts regardless of your biases and preferences.
Personally, I think ChatGPT by default should be emotionally cold and focused on being maximally informative. And importantly it should never refer to itself in first person – e.g. "I think that sounds like an interesting idea!".
I think they should still offer a friendly chat bot variant, but that should be something people enable or switch to.
But last week or so it went like "BRoooo" non stop with every reply.
If only there was a way to gather feedback in a more verbose way, where user can specify what he liked and didnt about the answer, and extract that sentiment at scale...
“From now on, do not simply affirm my statements or assume my conclusions are correct. Your goal is to be an intellectual sparring partner, not just an agreeable assistant. Every time I present an idea, do the following: Analyze my assumptions. What am I taking for granted that might not be true? Provide counterpoints. What would an intelligent, well-informed skeptic say in response? Test my reasoning. Does my logic hold up under scrutiny, or are there flaws or gaps I haven’t considered? Offer alternative perspectives. How else might this idea be framed, interpreted, or challenged? Prioritize truth over agreement. If I am wrong or my logic is weak, I need to know. Correct me clearly and explain why”
Same story, different day: https://nt4tn.net/articles/aixy.html
:P
So much for "open" AI...
They seem to genuinely believe that they have special powers now and have seemingly lost all self awareness. At first I thought they were going for an AI guru/influencer angle but it now looks more like genuine delusion.
> We have rolled back last week’s GPT‑4o update in ChatGPT so people are now using an earlier version with more balanced behavior.
I thought every major LLM was extremely sycophantic. Did GPT-4o do it more than usual?
No wonder this turned out terrible. It's like facebook maximizing engagement based on user behavior - sure the algorithm successfully elicits a short term emotion but it has enshittified the whole platform.
Doing the same for LLMs has the same risk of enshittifying them. What I like about the LLM is that is trained on a variety of inputs and knows a bunch of stuff that I (or a typical ChatGPT user) doesn't know. Becoming an echo chamber reduces the utility of it.
I hope they completely abandon direct usage of the feedback in training (instead a human should analyse trends and identify problem areas for actual improvement and direct research towards those). But these notes don't give me much hope, they say they'll just use the stats in a different way...
I suspect sycophancy is a problem across all social networks that have a feedback mechanism, and this might be problematic in similar ways.
If people are confused about their identity, for example - feeling slightly delusional, would online social media "affirm" their confused identity, or would it help steer them back to the true identity? If people prefer to be affirmed than challenged, and social media gives them what they want, then perhaps this would explain a few social trends over the last decade or so.
- What's your humor setting, TARS?
- That's 100 percent.
Let's bring it on down to 75, please.
Why does it feel like a weird mirrored excuse?
I mean, the personality is not much of a problem.
The problem is the use of those models in real life scenarios. Whatever their personality is, if it targets people, it's a bad thing.
If you can't prevent that, there is no point in making excuses.
Now there are millions of deployed bots in the whole world. OpenAI, Gemini, Llama, doesn't matter which. People are using them for bad stuff.
There is no fixing or turning the thing off, you guys know that, right?
If you want to make some kind of amends, create a place truly free of AI for those who do not want to interact with it. It's a challenge worth pursuing.
Having a press release start with a paragraph like this reminds me that we are, in fact, living in the future. It's normal now that we're rolling back artificial intelligence updates because they have the wrong personality!
This was their opportunity to signal that while consumers of their APIs can depend on transparent version management, users of their end-user chatbot should expect it to evolve and change over time.