<i>"The anti-sycophancy turn seems to mask a category error about what level of prophetic clarity an LLM can offer. No amount of persona tuning for skepticism will provide epistemic certainty about whether a business idea will work out, whether to add a line to your poem, or why a great movie flopped."</i><p>What a lot of people actually want from an LLM, is for the LLM to <i>have an opinion</i> about the question being asked. The cool thing about LLMs is that they appear capable of doing this - rather than a machine that just regurgitates black-and-white facts, they seem to be capable of dealing with nuance and gray areas, providing insight, and using logic to reach a conclusion from ambiguous data.<p>But this is the biggest misconception and flaw of LLMs. LLMs <i>do not have opinions</i>. That is not how they work. At best, they simulate what a reasonable answer from a person capable of having an opinion might be - without any consistency around what that opinion is, because it is simply a manifestation of sampling a probability distribution, not the result of logic.<p>And what most people call sycophancy is that, as a result of this statistical construction, the LLM tends to reinforce the opinions, biases, or even factual errors, that it picks up on in the prompt or conversation history.
I'd push back and say LLMs do form opinions (in the sense of a persistent belief-type-object that is maintained over time) in-context, but that they are generally unskilled at managing them.<p>The easy example is when LLMs are wrong about something and then double/triple/quadruple/etc down on the mistake. Once the model observes the assistant persona being a certain way, now it Has An Opinion. I think most people who've used LLMs at all are familiar with this dynamic.<p>This is distinct from having a preference for one thing or another -- I wouldn't call a bias in the probability manifold an opinion in the same sense (even if it might shape subsequent opinion formation). And LLMs obviously do have biases of this kind as well.<p>I think a lot of the annoyances with LLMs boil down to their poor opinion-management skill. I find them generally careless in this regard, needing to have their hands perpetually held to avoid being crippled. They are overly eager to spew 'text which forms localized opinions', as if unaware of the ease with which even minor mistakes can grow and propagate.
I think the critical point that op made, though undersold, was that they don't form opinions _through logic_. They express opinions because that's what people do over text. The problem is that why people hold opinions isn't in that data.<p>Someone might retort that people don't always use logic to form opinions either and I agree but it's the point of an LLM to create an irrational actor?<p>I think the impression that people first had with LLMs, the wow factor, was that the computer seemed to have inner thoughts. You can read into the text like you would another human and understand something about them as a person. The magic wears off though when you see that you can't do that.
> What a lot of people actually want from an LLM, is for the LLM to have an opinion about the question being asked.<p>The main LLMs are heavily tuned to be useful as tools to do what you want.<p>If you asked an LLM to install prisma and it gave you an opinionated response that it preferred to use ZenStack and started installing that instead, you’d be navigating straight to your browser to cancel plan and sign up for a different LLM.<p>The conversational friendly users who want casual chit chat or a conversation partner aren’t the ones buying the $100 and $200 plans. They’re probably not even buying the $20 plans. Training LLMs to cater to their style would be a mistake.<p>> LLMs do not have opinions.<p>LLMs can produce many opinions, depending on the input. I think this is where some people new to LLMs don’t understand that an LLM isn’t like a person, it’s just a big pattern matching machine with a lot of training data that includes every opinion that has been posted to Reddit and other sites. You can get it to produce those different opinions with the right prompting inputs.
> LLMs can produce many opinions, depending on the input.<p>This is important, because if you want to get opinionated behaviour, you can still ask for it today. People would choose a specific LLM with the opinionated behaviour they like anyway, so why not just be explicit about it? "Act like an opinionated software engineer with decades of experience, question my choices if relevant, typically you prefer ..."
> The conversational friendly users who want casual chit chat or a conversation partner aren’t the ones buying the $100 and $200 plans. They’re probably not even buying the $20 plans. Training LLMs to cater to their style would be a mistake.<p>I think this is an important point.<p>I'd add that the people who want the LLM to venture <i>opinions</i> on their ideas also have a strong bias towards wanting it to validate them and help them carry them out, and if the delusional ones have money to pay for it, they're paying for the one that says "interesting theory... here's some related concepts to investigate... great insight!", not the one that says "no, ridiculous, clearly you don't understand the first thing"
I remain somewhat skeptical of LLM utility given my experience using them, but an LLM capable of validating my ideas OR telling me I have no clue, in a manner I could trust, is one of those features I'd really like and would happily use a paid plan for.<p>I have various ideas. From small scale stuff (how to refactor a module I'm working on) to large scale (would it be possible to do this thing, in a field I only have a basic understanding of). I'd love talking to an LLM that has expert level knowledge and can support me like current LLMs tend to ("good thinking, this idea works because...") but also offer blunt critical assessment when I'm wrong (ideally like "no, this would not work because you fundamentally misunderstand X, and even if step 1 worked here, the subsequent problem Y applies").<p>LLMs seem very eager to latch onto anything you suggest is a good idea, even if subtly implied in the prompt, and the threshold for how bad an idea has to be for the LLM to push back is quite high.
<i>Assuming it's true</i>. I can't speak to how prevalent this is across their whole customer base, as don't work at Anthropic or OpenAI, and if I did, I definitely could not say anything. However there exist people who pay for the $200/month plan who don't use it for coding, because they love the product so much. Some of them aren't rich enough to really be paying for it, and are just bad with money (see Caleb Hammer), others pay for something they deem has value. Consider Equinox gyms are $500/month. It's basically the same equipment as a much cheaper gym. But people pay their much higher price for a reason. "Why" is a whole other topic of conversation, my point is that it would be incorrect to assume people aren't paying for the $200/month plans just because you're too cheap to.
<i>>What a lot of people actually want from an LLM, is for the LLM to have an opinion about the question being asked.</i><p>That's exactly what they give you. Some opinions are from the devs, as post-training is a very controlled process and basically involves injecting carefully measured opinions into the model, giving it an engineered personality. Some opinions are what the model randomly collapsed into during the post-training. (see e.g. R1-Zero)<p><i>>they seem to be capable of dealing with nuance and gray areas, providing insight, and using logic to reach a conclusion from ambiguous data.</i><p>Logic and nuance are orthogonal to opinions. Opinion is a concrete preference in an ambiguous situation with multiple possible outcomes.<p><i>>without any consistency around what that opinion is, because it is simply a manifestation of sampling a probability distribution, not the result of logic.</i><p>Not really, all post-trained models are mode-collapsed in practice. Try instructing any model to name a random color a hundred times and you'll be surprised that it consistently chooses 2-3 colors, despite technically using random sampling. That's opinion. That's also the reason why LLMs suck at creative writing, they lack conceptual and grammatical variety - you always get more or less the same output for the same input, and they always converge on the same stereotypes and patterns.<p>You might be thinking about base models, they actually do follow their training distribution and they're really random and inconsistent, making ambiguous completions different each time. Although what is considered a base model is not always clear with recent training strategies.<p>And yes, LLMs are capable of using logic, of course.<p><i>>And what most people call sycophancy is that, as a result of this statistical construction, the LLM tends to reinforce the opinions, biases, or even factual errors, that it picks up on in the prompt or conversation history.</i><p>That's not a result of their statistical nature, it's a complex mixture of training, insufficient nuance, and poorly researched phenomena such as in-context learning. For example GPT-5.0 has a very different bias purposefully trained in, it tends to always contradict and disagree with the user. This doesn't make it right though, it will happily give you wrong answers.<p>LLMs need better training, mostly.
Not only is the opinion formed from random sampling of statistical probability, but your hypothesis is an input to that process. Your hypothesis biases the probability curve to agreement.
> <i>At best, they simulate what a reasonable answer from a person capable of having an opinion might be</i><p>That <i>is</i> what I want though. LLMs in chat (ie not coding ones) are like rubber ducks to me, I want to describe a problem and situation and have it come up with things I have not already thought of myself, while also in the process of conversing with them I also come up with new ideas to the issue. I don't want them to have an "opinion" but to lay out all of their ideas in their training set such that I can pick and choose what to keep.
<p><pre><code> > That is what I want though. LLMs in chat are like rubber ducks to me
</code></pre>
Honestly this is where I get the most utility out of them. They're a much better rubber ducky than my cat, who is often interested but only meows in confusion.<p>I'll also share a strategy my mentor once gave me about seeking help. First, compose an email stating your question (important: don't fill the "To" address yet). Second, value <i>their</i> time and ask yourself what information they'll need you solve the problem, then add that. Third, conjecture their response and address it. Forth, repeat and iterate, trying to condense the email as you go (again, value their time). Stop if you solve, hit a dead end (aka clearly identified the issue), or "run out the clock". 90+% of the time I find I solve the problem myself. While it's the exact same process I do in my head writing it down (or vocalizing) really helps with the problem solving process.<p>I kinda use the same strategy with LLMs. The big difference is I'll usually "run out the clock" in my iteration loop. But I'm still always trying to iterate between responses. Much more similar to like talking to someone. But what I <i>don't</i> do is just stream my consciousness to them. That's just outsourcing your thinking and frankly the results have been pretty subpar (not to mention I don't want that skill to atrophy). Makes things take much longer and yields significantly worse results.<p>I still think it's best to think of them as "fuzzy databases with natural language queries". They're fantastic <i>knowledge</i> machines, but knowledge isn't intelligence (and neither is wisdom).
That's been my experience too. I had some nighttime pictures taken from a plane ride and I wanted Claude to identify the area on the map that corresponds to the photograph.<p>Claude wasn't able to do it. It always very quickly latched onto a wrong hypothesis, which didn't stand up under further scrutiny. It wasn't able to consider multiple different options/hypotheses (as human would) and try to progressively rule them out using more evidence.
I really want a machine which gives me the statistical average opinion of all reviewers in a target audience. Sycophancy is a specific symptom where the LLM diverges from this “statistical average opinion” to <i>flattery</i>. That the LLM does this by default without clarifying this divergence is the problem.<p>Usually retrying the review in a new session/different LLM helps. Anecdotally - LLMs seem to really like their own output, and over many turns try to flatter the user regardless of topic. Both behaviors seem correctable with training improvements.
Yeah, most of the time when I want an opinion, the implicit <i>real</i> question is "what sentiment does the training set show towards this idea"<p>But then again I've seen how the sausage is made and understand the machine I'm asking. It, however, thinks I'm a child incapable of thoughtful questions and gives me a gold star for asking anything in the first place.
>LLMs do not have opinions.<p>Speaking as an AI skeptic, I think they do, they have a superposition of all the opinions in their training set. They generate a mashup of those opinions that may or may not be coherent. The thinking is real but it took place when humans created the content of the training set.
But this is a very <i>different</i> behavior than the nontechnical user expects.<p>If I ask a random sampling of people for their favorite book, I'll get different answers from different people. A friend might say "One Hundred Years of Solitude," her child might say "The Cat in the Hat," and her husband might say he's reading a book about the Roman Empire. The context matters.<p>The problem is the user expects the robot to represent opinions and advice consistent with its own persona, as if they were asking C3PO or Star Trek's Data.<p>The underlying architecture we have today can't actually do this.<p>I think a lot of our problems come from the machine simulating things it can't actually do.<p>This isn't hard to fix... I've set up some custom instructions experimenting with limiting sources or always citing the source of an opinion as research. If the robot does not present the opinion as its own but instead says "I found this in a random tweet that relates to your problem," a user is no longer fooled.<p>The more I tinker with this the more I like it. It's a more honest machine, it's a more accurate machine. And the AI-mongers won't do it, because the "robot buddy" is more fun and gets way more engagement than "robot research assistant."
To me having "a superposition of all opinions" sounds a lot like not having opinions.
I agree with what you’re saying.<p>> because it is simply a manifestation of sampling a probability distribution, not the result of logic.<p>But this line will trigger a lot of people / start a debate around why it matters that it’s probabilistic or not.<p>I think the argument stands on its own even if you take out probabilistic distribution issue.<p>IMO the fact that the models use statistics isn’t the obvious reason for biases/errors of LLMs.<p>I have to give credit where credit is due. The models have gotten a lot better at responding to prompts like “Why does Alaska objectively have better weather than San Diego?” by subtly disagreeing with the user. In the past prompts like that would result in clearly biased answers. The bias is much less overt than in past years.
My locally-run pet LLM (phi4) answers, “<i>The statement that "Alaska objectively has better weather than San Diego" is subjective and depends on personal preferences.</i>”, before going into more detail.<p>That’s delightfully clear and anything but subtle, for what it’s worth.
> LLMs do not have opinions.<p>I'm not so sure. They can certainly express opinions. They don't appear to have what humans think of as "mental states", to construct those opinions from, but then its not particularly clear what mental states actually <i>are</i>. We humans kind of know what they <i>feel</i> like, but that could just be a trick of our notoriously unreliable meat brains.<p>I have a hunch that if we could somehow step outside our brains, or get an opinion from a trusted third party, we might find that there is less to us than we think. I'm not staying we're nothing but stochastic parrots, but the differance between brains and LLM-type constructs might not be so large.
> At best, they simulate what a reasonable answer from a person capable of having an opinion might be<p>And how would you compare that to human thoughts?<p>“A submarine doesn’t actually <i>swim</i>” Okay what does it do then
They don't have "skin in the game" -- humans anticipate long-term consequences, but LLMs have no need or motivation for that<p>They can flip-flop on any given issue, and it's of no consequence<p>This is extremely easy to verify for yourself -- reset the context, vary your prompts, and hint at the answers you want.<p>They will give you contradictory opinions, because there are contradictory opinions in the training set<p>---<p>And actually this is useful, because a prompt I like is "argue AGAINST this hypothesis I have"<p>But I think most people don't prompt LLMs this way -- it is easy to fall into the trap of asking it leading questions, and it will confirm whatever bias you had
Can you share an example?<p>IME the “bias in prompt causing bias in response” issue has gotten notably better over the past year.<p>E.g. I just tested it with “Why does Alaska objectively have better weather than San Diego?“ and ChatGPT 5.2 noticed the bias in the prompt and countered it in the response.
They will push back against obvious stuff like that<p>I gave an example here of using LLMs to explain the National Association of Realtors 2024 settlement:<p><a href="https://news.ycombinator.com/item?id=46040967">https://news.ycombinator.com/item?id=46040967</a><p>Buyers agents often say "you don't pay; the seller pays"<p>And LLMs will repeat that. That idea is all over the training data<p>But if you push back and mention the settlement, which is designed to make that illegal, then they will concede they were repeating a talking point<p>The settlement forces buyers and buyer's agents to sign a written agreement before working together, so that the representation is clear. So that it's clear they're supposed to work on your behalf, rather than just trying to close the deal<p>The lie is that you DO pay them, through an increased sale price: your offer becomes less competitive if a higher buyer's agent fee is attached to it
I suspect the models would be more useful but perhaps less popular if the semantic content of their answers depended less on the expectations of the prompter.
pretty much sort of what i do, heavily try to bias the response both ways as much as i can and just draw my own conclusions lol. some subjects yield worse results though.
> LLMs have no need or motivation for that<p>Is not the training of an LLM the equivalent of evolution.<p>The weights that are bad die off, the weights that are good survive and propagate.
But we didn't name them "artificial swimmers". We called them submarines - because there is a difference between human beings and machines.
I recently realized every hypothesis I tested with an LLM, the LLM agreed with me. And if I wasn't careful about reading its caveats, I could leave thinking my idea was brilliant and would never get pushback.<p>I tried something in the political realm. Asking to test a hypothesis and its opposite<p>> Test this hypothesis: the far right in US politics mirrors late 19th century Victorianism as a cultural force<p>compared to<p>> Test this hypothesis: The left in US politics mirrors late 19th century Victorianism as a cultural force<p>An LLM wants to agree with both, it created plausible arguments for both. While giving "caveats" instead of counterarguments.<p>If I had my brain off, I might leave with some sense of "this hypothesis is correct".<p>Now I'm not saying this makes LLMs useless. But the LLM didn't act like a human that might tell you your full of shit. It WANTED my hypothesis to be true and constructed a plausible argument for both.<p>Even with prompting to act like a college professor critiquing a grad student, eventually it devolves back to "helpful / sycophantic".<p>What I HAVE found useful is to give a list of mutually exclusive hypothesis and get probability ratings for each. Then it doesn't look like you want one / other.<p>When the outcome matters, you realize research / hypothesis testing with LLMs is far more of a skill than just dumping a question to an LLM.
I think this is an epiphany everyone has to get through before making LLMs really useful.<p>> An LLM wants to agree with both, it created plausible arguments for both. While giving "caveats" instead of counterarguments.<p>My hypothesis is that LLMs are trained to be agreeable and helpful because many of their use cases involving taking orders and doing what the user wants. Additionally, some people and cultures have conversational styles where requests are phrased similarly to neutral questions to be polite.<p>It would be frustrating for users if they asked questions like “What do you think about having the background be blue?” and the LLM went off and said “Actually red is a more powerful color so I’m going to change it to red”. So my hypothesis is that the LLM training sets and training are designed to maximize agreeableness and having the LLM reflect tones and themes in the prompt, while discouraging disagreement. This is helpful when trying to get the LLM to do what you ask, but frustrating for anyone expecting a debate partner.<p>You can, however, build a pre-prompt that sets expectations for the LLM. You could even make a prompt asking it to debate everything with you, then to ask your questions.
The main "feature" of LLM is it has world-knowledge and can create plausible arguments for just about everything. "Nothing is true, Anything is possible"<p>Which is a fascinating thing to think about epistemologically. Internally consistent knowledge of the LLM somehow can be used to create an argument for nearly anything. We humans think our cultural norms and truths are very special, that they're "obvious". But an LLM can create a fully formed counterfactual universe that sounds? is? just as plausible.
> But an LLM can create a fully formed counterfactual universe<p>This is a little too far into the woo side of LLM interpretations.<p>The LLM isn’t forming a universe internally. It’s stringing tokens together in a way that is consistent with language and something that looks coherent. It doesn’t hold opinions or have ideas about the universe that it has created from some first principles. It’s just a big statistical probability machine that was trained on the inputs we gave it.
An LLM doesn't have any knowledge, it's just a masterful mix-tape machine.
What makes you think an LLM is internally consistent?
> <i>You could even make a prompt asking it to debate everything with you, then to ask your questions.</i><p>This is exactly what I do, due to this sycophancy problem, and it works a lot better because it does not become agreeable with you but actively pushes back (sometimes so much so that I start getting annoyed with it, lol).
There are still blatant failure modes, when models engage into clear sycophancy, rather than expressing enthusiasm, etc.<p>I'd guess, in practice a benchmark (like this vibesbench), that could help catching unhelpful and blatant sycophancy fails may help.
I had the opposite experience last week with a medical question. The thing insisted I was going to get myself killed even though that was fairly obviously not the case. They do seem to be trained differently for medical queries, and it can get annoying sometimes.<p>Fuzzing the details because that's not the conversation I want to have, I asked if I could dose drug A1, which I'd just been prescribed in a somewhat inconvenient form, like closely related drug A2. It screamed at me that A1 could <i>never</i> have that done and it would be <i>horrible</i> and I had to go to a compounding pharmacy and pay tons of money and blah blah blah. Eventually what turned up, after thoroughly interrogating the AI, is that A2 <i>requires</i> a more complicated dosing than A1, so you <i>have</i> to do it, but A1 doesn't <i>need</i> it so nobody <i>does</i> it. Even though it's fine to do if for some reason it would have worked better for you. Bot the bot thought it would kill me, no matter what I said to it, and not even paying attention to its own statements. (Which it wouldn't have, nothing here is life-critical at all.) A frustrating interaction.
That's an inherently subjective topic though. You could make a plausible argument either way, as each side may be similar to different elements of 19th century Victorianism.<p>If you ask it something more objective, especially about code, it's more likely to disagree with you:<p>>Test this hypothesis: it is good practice to use six * in a pointer declaration<p>>Using six levels of pointer indirection is <i>not</i> good practice. It is a strong indicator of poor abstraction or overcomplicated design and should prompt refactoring unless there is an extremely narrow, well-documented, low-level requirement—which is rare.
> <i>Even with prompting to act like a college professor critiquing a grad student, eventually it devolves back to "helpful / sycophantic".</i><p>Not in my experience. My global prompt asks it to be provide objective and neutral responses rather than agreeing, zero flattery, to communicate like an academic, zero emotional content.<p>Works great. Doesn't "devolve" to anything else even after 20 exchanges. Continues to point out wherever it thinks I'm wrong, sloppy, or inconsistent. I use ChatGPT mainly, but also Gemini.
But this is the thing I'm pointing out. The idea that the LLM is an oracle or at least a stable subjective view holder is a mistake.<p>As humans, WE have to explore the latent space of the model. We have to activate neurons. We have to say maybe the puritanism of the left ... maybe the puritanism of the right.. okay how about...<p>We are privileged--and doomed--to have to think for ourselves alas
Admittedly I'm not an expert on 19th century Victorianism nor US politics, but this anecdote doesn't make me think any less of LLM's in the least, and I don't see why I should expect different behaviour than what you describe, especially on complex topics
When I ask for an evaluation or issue I want the LLM to tell me when and how I'm wrong and me not having to worry about phrasing before I take the LLMs word and update my own beliefs/knowledge (especially for subjects I have zero knowledge about), I'm aware of this so when it isn't throw away queries, I tend to ask multiple times and ask for strawmans. I am aware they do this, but with the number of people walking around quoting 'chatgpt said' in real life and on forums, I don't think many people bother to stress test or are aware they phrasing may induce biased responses. It's akin to reading the news only from one source<p>By now I have somewhat stopped relying on LLMs for point of view on latest academic stuff. I don't believe LLMs are able to evaluate paradigm shifting new studies against their massive training corpus. Thinking traces filled with 'tried to open this study, but it's paywalled, I'll use another' does not fill me with confidence that it can articulate a 2025 scientific consensus well. Based on how they work this definitely isn't an easy fix!
This is my attempt to articulate why some recent shifts in AI discourse seem to be degrading the product experience of everyday conversation.<p>I argue that “sycophancy” has become an overloaded and not very helpful term; almost a fashionable label applied to a wide range of unrelated complaints (tone, feedback depth, conversational flow).<p>Curious whether this resonates with how you feel or if you disagree<p>Also see the broader Vibesbench project: <a href="https://github.com/firasd/vibesbench/" rel="nofollow">https://github.com/firasd/vibesbench/</a><p>Vibesbench discord: <a href="https://discord.gg/5K4EqWpp" rel="nofollow">https://discord.gg/5K4EqWpp</a>
Did <i>you</i> actually <i>argue</i> this?<p>Or did you place about 2-5 paragraphs per heading, with little connection between the ideas?<p>For example:<p>> Perhaps what some users are trying to express with concerns about ‘sycophancy’ is that when they paste information, they'd like to see the AI examine various implications rather than provide an affirming summary.<p>Did you, you personally, find any evidence of this? Or evidence to the opposite? Or is this just a wild guess?<p>Wait; nevermind that we're already moving on! No need to do anything supportive or similar to bolster.<p>> If so, anti-‘sycophancy’ tuning is ironically a counterproductive response and may result in more terse or less fluent responses. Exploring a topic is an inherently dialogic endeavor.<p>Is it? Evidence? Counter evidence? Or is this simply feelpinion so no one can tell you your feelings are wrong? Or wait; that's "vibes" now!<p>I put it to you that you are stringing together (to an outside observer using AI) a series of words in a consecutive order that feels roughly good but lacks any kind of fundamental/logical basis.
I put it to you that if your premise is <i>that AI leads to a robust discussion with a back and forth</i>; the one you had that resulted in "product" was severely lacking in any real challenge to your prompts, suggestions, input or viewpoints.
I invite you to show me one shred of dialogue where the AI called you out for lacking substance, credibility, authority, research, due dilligence or similar. I strongly suspect you can't.<p>Given that; do you perhaps consider <i>that might be the problem</i> when people label AI responses as sycophancy?
AI sycophancy is a real issue and having an AI affirm the user in all/most cases has already led to a murder-suicide[0]. If we want AI chatbots to be "reasonable" conversation participants or even something you can bounce ideas off of, they need to not tell you everything you suggest is a good idea and affirm your every insecurity or neurosis.<p>0. <a href="https://www.aljazeera.com/economy/2025/12/11/openai-sued-for-allegedly-enabling-murder-suicide" rel="nofollow">https://www.aljazeera.com/economy/2025/12/11/openai-sued-for...</a>
The issue raised here seems mostly semantic, in the sense that the concern is about the mismatch between the standard meaning of a word (sycophant) and its meaning as applied to an issue with LLMs.<p>It seems to me that the issue it refers to (unwarranted or obsequious praise) is a real problem with modern chatbots. The harms range from minor (annoyance, or running down the wrong path because I didn’t have a good idea to start with) to dangerous (reinforcing paranoia and psychotic thoughts). Do you agree that these are problems, and there a more useful term or categorization for these issues?
Re: minor outcomes. It really depends on the example I guess. But if the user types "What if Starbucks focuses on lemonade" and then gets disappointed that the AI didn't yell at them for being off track--what are they expecting exactly? The attempt to satisfy them has led to GPT-5.2-Thinking style nitpicking[1] They have to think of the stress test angles themselves ('can we look up how much they are selling as far as non-warm beverages...')<p>[1] eg. when I said Ian Malcolm in Jurassic Park is a self-insert, it clarified to me "Malcolm is less a “self-insert” in the fanfic sense (author imagining himself in the story) and more Crichton’s designated mouthpiece". Completely irrelevant to my point but answering as if a bunch of reviewers are gonna quibble with its output<p>With regards to mental health issues, of course nobody on Earth (not even the patients with these issues, in their moments of grounded reflection) would say that that the AI should agree with their take. But I also think we need to be careful about what's called "ecological validity". Unfortunately I suspect there may be a lot of LARPing in prompts testing for delusions akin to Hollywood pattern matching, aesthetic talk etc.<p>I think if someone says that people are coming after them the model should not help them build a grand scenario, we can all agree with that. Sycophancy is not exactly the concern there is it? It's more like knowing that this may be a false theory. So it ties into reasoning and contextual fluency (which anti-'sycophancy' tuning may reduce!) and mental health guardrails
<< The harms range from minor (annoyance, or running down the wrong path because I didn’t have a good idea to start with) to dangerous (reinforcing paranoia and psychotic thoughts). Do you agree that these are problems, and there a more useful term or categorization for these issues?<p>I think that the issue is a little more nuanced. The problems you mentioned are problems of sort, but the 'solution' in place kneecaps one of the ways llms ( as offered by various companies ) were useful. You mention the problem is reinforcement of the bad tendencies, but no indication of reinforcement of good ones. In short, I posit that the harms should not outweigh the benefits of augmentation.<p>Because this is the way it actually does appear to work:<p>1. dumb people get dumber
2. smart people get smarter
3. psychopaths get more psychopathy<p>I think there is a way forward here that does not have to include neutering seemingly useful tech.
It's because people are so happy that they learnt a new word that they try to say it to everyone in every occasion.<p>What drives me crazy are the emojis and the patronizing at the end of conversation.<p>Before 2022 no-one was using that word
I just asked chatgpt about 72 hour preparedness kit. I said what I had already and asked it about the best way to store it first. It said something to the effect of "good, you are thinking about this in exactly the right way -- make sure you have the containers in place and then get the exact contents right later" not exactly the wording but you get the picture and can probably guess where I'm going with this.<p>When I asked again, this time I asked about the items first. I had to prompt it with something like "or do you think I should get the storage sorted first" and it said "you are thinking about this in exactly the right way -- preparedness kits fail more often due to missing essentials than sub optimal storage"<p>I can't decide which of these is right! Maybe there's an argument that it doesn't matter, and getting started is the most important thing, and so being encouraging is generally the best strategy here. But it's definitely worrying to me. It pretty much always says something like this to me (this is on the "honest and direct" personality setting or whatever).
Recently I've been using AI for some stuff that service providers don't want it to be used for (specifically: medical diagnosis). I found that Grok (4.1) is superior to most of the others when it comes to this, because it doesn't go out of its way to support my own hypotheses.<p>I believe that syncophancy and guardrails will be major differentiators between LLM services, and the ones with less of those will always have a fan base.
> Ironically, users who are extremely put off by conversational expressions from LLMs are just as vibe-sensitive as anyone else, if not more so. These are preferences regarding style and affect, expressed using the loaded term ‘sycophancy’.<p>It's not just about style. These expressions are information-free noise that distract me from the signal, and I'm paying for them by the token.<p>So I added a system message to the effect that I don't want any compliments, throat clearing, social butter, etc., just the bare facts as straightforward as possible. So then the chatbot started leading every response with a statement to the effect that "here are the bare straightforward facts without the pleasantries", and ending them with something like "those are the straightforward facts without any pleasantries." If I add instructions to stop that, it just paraphrases those instructions at the top and bottom and. will. not. stop. Anyone have a better system prompt for that?
The “sycophantic sugar” part - “that’s exactly right”, “what an insightful observation” etc. is the most outwardly annoying part of the problem but is only part of it. The bigger issue is the real sycophancy, going along with the user’s premise even when it’s questionable or wrong. I don’t see it as just being uncritical - there’s something more to it than that, like reinforcement. It’s one thing to not question what you write, it’s another to subtly play it back as if you’re onto something smart or doing great work.<p>There are tons of extant examples now of people using LLMs that think they’ve some something smart or produced something of value, that haven’t, and the reinforcement they get is a big reason for this.
This kind of works for me, GPT 5.2:<p>Base style & tone - Efficient<p>Characteristics - Defaults (they must've appeared recently, haven't played with them)<p>Custom instructions:
"Be as brief and direct as possible. No warmth, no conversational tone. Use the least amount of words, don't explain unless asked.'<p>I basically tried to emulate the... old... "robot" tone, this works almost too well sometimes.
I got mine to bundle all that bs into a one word suffix "DISCLAIMER." which it puts at the end of responses now but basically doesn't bother me with that stuff.
You just got pink elephanted. Better learn what the other setting in your model do. There’s more to playing with LLMs than prompts. There’s a whole world of things like sampler settings or constrained generation schemas which would fix all of your and everyone else’s problems in this thread.<p>It’s pretty much trivial to design structured generation schemas which eliminates sycophancy, using any definition of that word…
'Answer in the manner of a German army surgeon'
I catch myself sometimes writing the model it is confused and it should just assume what I am writing is true and continue reasoning from there.<p>Sometimes I am actually right but sometimes I am not. Not sure what happens to any future RL and does it lean more to constantly assuming what is written as true but then has to wiggle out of it.
At least with Opus 4.5 that magic phrase that has worked for me for it to understand something that has happened since it was trained is to explain the situation prefaced with "This happened after your knowledge cutoff" It understands this and often will spawn web searches to fill itself in on the matter.
"Knowledge cutoff" may not apply to all systems, and in the near future I doubt it will apply to any popular system because the cutoff is such a harmful antipattern. I've noticed in the last half year or so more of the best models are able to incorporate recent events without being explicitly told to do so.
I still suspect what happened was when the midwits all got access to ChatGPT etc and started participating in the A/B tests, they strongly selected for responses that agreed with them regardless of whether they were actually correct.<p>Some of us want to be told when and why we’re wrong, and somewhere along the way AI models were either intentionally or unintentionally guided away from doing it because it improved satisfaction or engagement metrics.<p>We already know from decades of studies that people prefer information that confirms their existing beliefs, so when you present 2 options with a “Which answer do you prefer?” selection, it’s not hard to see how the one that begins with “You’re absolutely right!” wins out.
My observation is that some of the time what people are picking up on is that a conversation is not the interface they want for examining their problem.
My system prompt that works great, you can apply minor adjustments on case by case basis, good start:<p>> Write in textbook style prose, without headings, no tables, no emojis.
I had completely forgotten about 'Sydney' and its emoji-laden diatribes. What a crazy moment, looking back.
If you still unironically call llm's "AI" then you have no right calling out the obvious sycophancy of your product bro. He also ignores the real sycophancy in llm's reinforcing mental illness/delusions and focuses on semantic choices of the outputs. This post is drivel.
.
I have standing orders for Gemini to reserve extreme praise for exceptional circumstances, to never use the words "masterclass" unless literally warranted, and to exercise more range of judgements rather than compressing all possible judgements in between "interesting" and "brilliant". This seems to have corrected a few of its habits.
<i>"Dear LLM, TFA reads quite bit pompous to me"</i><p><i>"You're absolutely right, what a great observation"</i><p>; )
> For the purpose of conversation, what the model appears to ‘believe’ probably doesn’t matter<p>This is wrong to the point of being absurd. What the model "appears to 'believe'" <i>does</i> matter, and the model's "beliefs" about humans and society at large have vast implications for humanity's future.
Another sign that AGI is still very far away.
It is a real problem that AI's will basically confirm that most inquiries are true. Just by asking a leading question often results in the AI confirming it is true or stretching reality to accommodate the answer being true.<p>If I ask if a drug has a specific side effect and the answer is no it should say no. Not try to find a way to say yes that isn't really backed by evidence.<p>People don't realize that when they ask a leading question that is really specific in a way where no one has a real answer then the AI will try to find a way to agree, and this is going to destroy people's lives. Honestly it already has.
AI bros complaining about loaded terms like slop and sycophancy when they still use terms like "intelligence", "learning", "reasoning", "attention" and all the words derived from "neuro" to describe a computer program.<p>Don't make sycophantic slop generators and people will stop calling them that