44 comments

  • leumon8 days ago
    &gt; We’re continuing to make progress toward a version of ChatGPT designed for adults over 18, grounded in the principle of treating adults like adults, and expanding user choice and freedom within appropriate safeguards. To support this, we’ve rolled out age prediction for users under 18 in most markets. <a href="https:&#x2F;&#x2F;help.openai.com&#x2F;en&#x2F;articles&#x2F;12652064-age-prediction-in-chatgpt" rel="nofollow">https:&#x2F;&#x2F;help.openai.com&#x2F;en&#x2F;articles&#x2F;12652064-age-prediction-...</a><p>interesting
    • GoatInGrey8 days ago
      Pornographic use has long been the &quot;break glass in case of emergency&quot; for the LLM labs when it comes to finances.<p>My personal opinion is that while smut won&#x27;t hurt anyone in of itself, <i>LLM smut</i> will have weird and generally negative consequences. As it will be crafted specifically for you on top of the intermittent reinforcement component of LLM generation.
      • estimator72928 days ago
        While this is a valid take, I feel compelled to point out Chuck Tingle.<p>The sheer amount and variety of smut books (just books) is <i>vastly</i> larger than anyone wants to realize. We passed the mark decades ago where there is smut available for <i>any</i> and <i>every</i> taste. Like, to the point that even LLMs are going to take a long time to put a dent in the smut market. Humans have been making smut for longer than we&#x27;ve had writing.<p>But again I don&#x27;t think you&#x27;re wrong, but the scale of the problem is way distorted.
        • MBCook8 days ago
          That’s all simple one way consumption though. I suspect the effect on people is very different when it’s interactive in the way an LLM can be that we’ve never had to recon with before.<p>That’s where the danger may lie.
          • viraptor8 days ago
            You could commission smut of whatever type you want for quite a while. And many people do so. Even customised smut is not new. It&#x27;s just going to get a bit cheaper and automated.
            • azan_8 days ago
              You couldn&#x27;t talk to commissioned smut. Of course you could request changes etc. but the feedback loop was nowhere close to what you can get with AI. Interactivity is a very big deal.
              • goodmythical7 days ago
                There are absolutely people getting paid to roleplay smut in chat sessions and have been doing so at least since original Second Life and likely since the dawn of chat.
              • formerly_proven7 days ago
                There are several large platforms for interactive 1:1 or 1:few smut in various media forms. “LLM enthusiasts” have been using smutai for a couple years now. Smut generation is probably on of the top three reasons for people to build local AI rigs.
              • flarn20068 days ago
                Sounds like an improvement then. If people have more freedom to enjoy what they like how they like it, I see that as entirely a good thing.
                • intended7 days ago
                  At the degree of generalization you are working at, yes. More preference matching is a good thing.<p>This is spherical cows territory though, so its only good for setting out Birds Eye view of principles.
          • pixl978 days ago
            Alien 1: &quot;How did the earthlings lose control of their own planet?&quot;<p>Alien 2: &quot;AI generated porn&quot;
        • cal_dent8 days ago
          i&#x27;ve always wondered how much the increasing prevalence of smut &amp; not so niche romance novels, that have proliferated since e-readers became mainstream, have had on Gen Z and younger&#x27;s sometimes unrealistic view&#x2F;expectations of relationship. A lot of time is spent on porn sites etc. but not so much on how mainstream some of these novels have become
          • cluckindan8 days ago
            They had similar wonderings in the Victorian era, and probably in the Roman empire and ancient Greece too.
            • kuboble7 days ago
              Yes, human nature hasn&#x27;t changed but there is a reason why only recently obesity epidemic has developed.<p>Cheap unlimited access to stuff that was always scarce during human evolution creates an &#x27;evolutionary mismatch&#x27; where unlimited access to stuff bypasses our lack of natural satiety mechanisms.
              • cluckindan7 days ago
                That is completely discounting the effects of PFAS and plasticizers on the human endocrine system and the downstream effects on obesity.
            • Gud8 days ago
              But you don’t think there are big differences?
              • Lerc7 days ago
                Well they are vastly more aware of the notion of consent now.
            • satisfice7 days ago
              Have you ever stopped to realize that, from the Victorian’s point of view, they have been proven completely right about what would happen if ladies started showing their ankles?
              • ecshafer7 days ago
                They <i>were</i> right. We have largely had 200 years of socially and legally enforced morality being eroded with the conservatives saying &quot;If you remove X then Y and Z will happen!&quot;. The liberals saying &quot;Why do you care anyways? That&#x27;s a slippery slope it won&#x27;t happen!&quot;. The the conservatives immediately being proven right, but no one is willing to walk back on liberalization of moral issues since too many like hedonism.
                • cess117 days ago
                  What do you mean, &quot;proven right&quot;? Could you give three examples?
                • victorbjorklund7 days ago
                  That only assumes that nothing else changed in society at the same time. Is all this happening because men saw some ankles? Or is it a symptom of other changes in society (like more individual freedoms and rights, more education, etc)
              • Wowfunhappy7 days ago
                ...sorry, I&#x27;m dense apparently, what did they predict vs what happened?
                • fragmede7 days ago
                  The Victorians were accidentally right about ankles, which is funny in hindsight. Once one arbitrary rule breaks, people start noticing the rest are kind of fake too, and it turns out &quot;modesty&quot; was load-bearing for a whole governance model.<p>Ankles -&gt; knees -&gt; jazz -&gt; voting -&gt; rock -&gt; no-fault divorce -&gt; Tinder -&gt; polyamory discourse on airplanes. it&#x27;s a joke, but also sort of how cultural change actually propagates. The collapse did happen, just not of morals. Of enforcement. After that, everything is just people discovering the rules were optional all along. Including money.
                  • KludgeShySir7 days ago
                    On the other hand (based on memory of research I did many years ago), in societies where nudity is common (e.g. African tribes where at least breasts are usually visible), there is a much lower rate of sex-related problems (sexual assault, etc)
                • satisfice7 days ago
                  Well, I wasn&#x27;t speaking of a formal prediction by leading Victorian moral researchers... I was referring to our collective common knowledge of Victorian hangups.<p>Nevertheless, here is an example of Victorian anxiety regarding showing ankles: <a href="https:&#x2F;&#x2F;archive.org&#x2F;details&#x2F;amanualpolitene00pubgoog&#x2F;page&#x2F;n231&#x2F;mode&#x2F;2up" rel="nofollow">https:&#x2F;&#x2F;archive.org&#x2F;details&#x2F;amanualpolitene00pubgoog&#x2F;page&#x2F;n2...</a><p>It&#x27;s easy to say &quot;oh they were silly to worry about such things.&quot; But that&#x27;s only because we see it from our own point of view.<p>Alternatively, imagine describing roads, highways, traffic congestion and endless poles strung with electrical wire all over the place to someone from the 11th century. This would sound like utter ruination of the land to them. But you and I are used to it, so it just seems normal.
            • fc417fc8027 days ago
              They might well have been right - I&#x27;m no anthropologist.<p>Certainly they had neither the quantity nor ease of access that we do.
            • devnullbrain7 days ago
              Is your take that the way we view sexuality today is not meaningfully different from the Victorian era?
        • monksy7 days ago
          I want smut that talks about agent based development and crawdbot to do dirty dirty things.<p>Does that exist yet. I don&#x27;t think so.
          • indrora7 days ago
            Best I can do is [1] Sentient Lesbian Em-Dashes and [2] An AI hallucination made real for now.<p>The man&#x27;s probably thinking something up though. &quot;Pounded in the butt by Microslop Agentic Studio 2026&quot; has a ring.<p>[1] <a href="https:&#x2F;&#x2F;www.amazon.com&#x2F;Sentient-Lesbian-Em-Dash-Punctuation-ebook&#x2F;dp&#x2F;B0B97PNTDN" rel="nofollow">https:&#x2F;&#x2F;www.amazon.com&#x2F;Sentient-Lesbian-Em-Dash-Punctuation-...</a> [2] <a href="https:&#x2F;&#x2F;www.amazon.com&#x2F;Last-Algorithm-Pounded-Claimed-Sun-Times-ebook&#x2F;dp&#x2F;B0F99X5NPP" rel="nofollow">https:&#x2F;&#x2F;www.amazon.com&#x2F;Last-Algorithm-Pounded-Claimed-Sun-Ti...</a>
            • latexr7 days ago
              &gt; Sentient Lesbian Em-Dashes<p>Looked at the cover and saw “From Two Time Hugo Award Finalist Chuck Tingle”.<p>There’s no way that’s true. But I did a quick search anyway, and holy shit!<p><a href="https:&#x2F;&#x2F;www.thehugoawards.org&#x2F;hugo-history&#x2F;2016-hugo-awards&#x2F;#:~:text=Chuck%20Tingle" rel="nofollow">https:&#x2F;&#x2F;www.thehugoawards.org&#x2F;hugo-history&#x2F;2016-hugo-awards&#x2F;...</a><p><a href="https:&#x2F;&#x2F;www.thehugoawards.org&#x2F;hugo-history&#x2F;2017-hugo-awards&#x2F;#:~:text=Chuck%20Tingle" rel="nofollow">https:&#x2F;&#x2F;www.thehugoawards.org&#x2F;hugo-history&#x2F;2017-hugo-awards&#x2F;...</a><p>The story behind it:<p><a href="https:&#x2F;&#x2F;www.quora.com&#x2F;How-did-Chuck-Tingle-become-a-Hugo-Award-finalist" rel="nofollow">https:&#x2F;&#x2F;www.quora.com&#x2F;How-did-Chuck-Tingle-become-a-Hugo-Awa...</a><p><a href="https:&#x2F;&#x2F;archive.ph&#x2F;20160526154656&#x2F;http:&#x2F;&#x2F;www.vox.com&#x2F;2016&#x2F;5&#x2F;26&#x2F;11759842&#x2F;chuck-tingle-hugo-award-rabid-puppies-explained" rel="nofollow">https:&#x2F;&#x2F;archive.ph&#x2F;20160526154656&#x2F;http:&#x2F;&#x2F;www.vox.com&#x2F;2016&#x2F;5&#x2F;...</a>
              • Cthulhu_7 days ago
                They wrote a book about it too, &quot;Slammed In The Butt By My Hugo Award Nomination&quot;.
          • gostsamo7 days ago
            rule 34
        • bakugo8 days ago
          &gt; The sheer amount and variety of smut books (just books) is vastly larger than anyone wants to realize. We passed the mark decades ago where there is smut available for any and every taste.<p>It&#x27;s important to note that the <i>vast</i> majority of such books are written for a female audience, though.
      • bandrami7 days ago
        Whatever reward-center path is short-circuiting in 0.0001% of the population and leading to LLM psychosis will become a nuclear bomb for them if we get the sex drive involved too.
        • fragmede7 days ago
          Realtime VR AI porn will be the end of society, but by then, we&#x27;ll also have the technology to grow babies in artificial wombs, which is also going to end society as we know it, since we won&#x27;t need women any more (by then, we also won&#x27;t need men for the DNA in their sperm to make babies either, which cancels out). Of course, if we don&#x27;t need women or men, who&#x27;s left? What&#x27;s this &quot;we&quot; I&#x27;m talking about?<p>Why, the AI&#x27;s after they&#x27;ve gained sentience, of course.
      • jjmarr7 days ago
        I can do as much smut as I want through the API for all SOTA models.
        • bpavuk7 days ago
          true, but:<p>1. you have to &quot;jailbreak&quot; the model first anyway, which is what&#x27;s easier to do over API<p>2. is average layman aware of the concept of &quot;API&quot;? no, unlikely. apps and web portals are more convenient, which is going to lower the bar to access LLM porn
          • serf7 days ago
            Well and trust the data isn&#x27;t going anywhere.<p>I trust none of the llm groups to be safe with my data , erp with a machine is going to leave some nasty breadcrumbs for some future folks i bet.
          • jjmarr7 days ago
            I don&#x27;t have to jailbreak the models over APIs lol.
        • josephg7 days ago
          I don’t know if this is still the case, but as of a year or so ago OpenAI would suspend your account if they noticed you using their models for this sort of thing. They said it was against their TOS.
      • spicyusername7 days ago
        <p><pre><code> while smut won&#x27;t hurt anyone in of itself </code></pre> &quot;Legacy Smut&quot; is well known to cause many kinds of harm to many kind of people, from the participants to the consumers.
      • measurablefunc8 days ago
        People are already addicted to non-interactive pornography so this is going to be even worse.
      • BoredomIsFun7 days ago
        For those interested in smut I&#x27;d recommend to use local Mistral models.
      • PunchyHamster7 days ago
        I guess technically it will make some onlyfans content creators unemployed, given there is pretty large market for custom sexual content there.
      • subscribed7 days ago
        Why llm smut in particular? There&#x27;s already a vast landscape of the interactive, VR games for all tastes.<p>Why LLM is supposed to be worse?
        • josephg7 days ago
          I think the argument is that it’s interactive. You’re no longer just passively reading or watching content. You can join in on the role play.<p>I’m not sure why that’s a bad thing though.
          • subscribed6 days ago
            Same with games as compared to videos, especially VR.<p>Feels like someone angry at the machines capable of generating a tailored story.
      • UltraSane7 days ago
        I&#x27;m waiting until someone combines LLMs with a humanoid robot and a realdoll. That will have a lot of consequences.
      • gehwartzen7 days ago
        I can already see our made to order, LLM generated, VR&#x2F;neurolink powered, sex fantasies come to life. Throw in the synced Optimus sex robots…<p>I can see why Elons making the switch from cars. We certainly won’t be driving much
    • thayne8 days ago
      It says what to do if you are over 18, but thinks you are under 18. But what if it identifies someone under 18 as being older?<p>And what if you are over 18, but don&#x27;t want to be exposed to that &quot;adult&quot; content?<p>&gt; Viral challenges that could push risky or harmful behavior<p>And<p>&gt; Content that promotes extreme beauty standards, unhealthy dieting, or body shaming<p>Seem dangerous regardless of age.
      • novemp8 days ago
        &gt; And what if you are over 18, but don&#x27;t want to be exposed to that &quot;adult&quot; content?<p>Don&#x27;t prompt it.
      • Gud8 days ago
        What are these extremes beauty standards being promoted?<p>Because it seems to me large swaths of the population need some beauty standards
        • sejje7 days ago
          Yes, but you&#x27;re not allowed to say that to them.<p>They are victimized by the fact that models are attractive, and that is &quot;unrealistic,&quot; so they&#x27;ve been getting plus sized models etc.<p>The &quot;extreme beauty standards&quot; are basically just &quot;healthy BMI.&quot;
          • novemp7 days ago
            Anorexia is overrepresented in both male and female models.<p>To get that &quot;ripped&quot; look so many &quot;fit&quot; guys have, they often have to be dehydrated.<p>Actresses wear full faces of makeup even in apocalyptic scenarios.<p>It goes way beyond &quot;just don&#x27;t be fat&quot;.
            • sejje6 days ago
              It does, but the remedy they choose is not &quot;women don&#x27;t wear makeup,&quot; it&#x27;s &quot;we use fat women, and they wear more makeup than skinny women.&quot;
              • novemp6 days ago
                Including plus-size models doesn&#x27;t fix the problem, but it&#x27;s also not bad to have models show fat people &quot;this is what our clothes could look like on your body&quot;. That&#x27;s a logical choice under capitalism if you want fat people to buy your clothes.
                • sejje6 days ago
                  There isn&#x27;t a problem with what&#x27;s shown on TV.<p>The problem is inside some people&#x27;s heads, and diets.<p>I&#x27;m not talking about models for clothes, anyway.
                  • novemp6 days ago
                    Right, you&#x27;re talking about beauty standards in general, which we&#x27;ve already established go way beyond &quot;just don&#x27;t be fat&quot;.
                    • sejje5 days ago
                      No, I&#x27;m talking about the beauty standards that get complained about, and that we see adjustments for.
                      • novemp5 days ago
                        Is complaining about fat people on TV going to solve any of those other problems?
                        • sejje5 days ago
                          I didn&#x27;t complain about it, I discussed it.
    • geeunits8 days ago
      This is for advertising purposes, not porn. They might feign that&#x27;s the reason, but it&#x27;s to allow alcohol &amp; pharma to advertise, no doubt.
      • dawnerd7 days ago
        Bingo. There’s laws around advertising to children all over the world.
      • bpavuk7 days ago
        both, actually. porn for users, ad spots for companies.
    • Kiboneu7 days ago
      How I think it could play out:<p>- OpenAI botches the job. Article pieces are written about the fact that kids are still able to use it.<p>- Sam “responds” by making it an option to use worldcoin orbs to authenticate. You buy it at the “register me” page, but you will get an equivalent amount of worldcoin at current rate. Afterwards the orb is like a badge that you can put on your shelf to show to your guests.<p>“We heard you loud and clear. That’s why we worked hard to provide worldcoin integration, so that users won’t have to verify their age through annoying, insecure and fallible means.” (an example marketing blurb would say, implicitly referring to their current identity servicer Persona which people find annoying).<p>- After enough orb hardware is out in the public, and after the api gains traction for 3rd parties to use it, send a notice that x months for now, login without the orb will not be possible. “Here is a link to the shop page to get your orb, available in colors silver and black.”
    • chilmers8 days ago
      Sexual and intimate chat with LLMs will be a huge market for whoever corners it. They&#x27;d be crazy to leave that money on the table.
      • palmotea8 days ago
        That&#x27;s why laws against drugs are so terrible, it forces law-abiding businesses to leave money on the table. Repeal the laws and I&#x27;m sure there will be tons of startups to profit off of drug addiction.
        • chilmers8 days ago
          There are many companies making money off alcohol addiction, video game addiction, porn addiction, food addiction, etc. Should we outlaw all these things? Should we regulate them and try to make them safe? If we can do that for them, can&#x27;t we do it for AI sex chat?
          • latexr7 days ago
            The world isn’t black and white. Should we outlaw video games? No, I don’t think so. Should we outlaw specific addictive features, such as loot boxes, which are purposefully designed to trigger addiction in people and knowingly cause societal harm in the name of increasing profits for private companies? Probably.
          • ecshafer7 days ago
            &gt; There are many companies making money off alcohol addiction, video game addiction, porn addiction, food addiction, etc. Should we outlaw all these things?<p>Yes
            • sejje7 days ago
              &quot;ban number munchers&quot;
          • runarberg8 days ago
            And that makes it all alright doesn’t it?<p>There are also gangs making money off human trafficking? Does that make it OK for a corporation to make money off human trafficking as well? And there are companies making money off wars?<p>When you argue with whataboutism, you can just point to whatever you like, and somehow that is an argument in your favor.
            • 5423542342357 days ago
              They aren&#x27;t doing whataboutism. They are comparing prohibition&#x2F;criminalization of a harmful industry to regulation, and the effects of both. Gambling isn&#x27;t exactly good, but there is definitely a difference between a mafia bookies and regulated sports betting services and the second&#x2F;third order effects from both. Treating drug use as a criminal act, as opposed to a healthcare problem, has very different societal effects.<p>Whataboutism is more like &quot;Side A did bad thing&quot;, &quot;oh yeah, what about side B and the bad things they have done&quot;. It is more just deflection. While using similar&#x2F;related issues to inform and contextualize the issue at hand can also be overused or abused, but it is not the same as whataboutism, which is rarely productive.
            • bethekidyouwant8 days ago
              How is ai sex chat like any of those things, whataboutism indeed
              • runarberg8 days ago
                I was using whataboutism to demonstrate how bad of an argument whataboutism is. My arguments were exactly as bad as my parent’s, and that was the point.
                • fc417fc8027 days ago
                  Pointing out an inconsistency isn&#x27;t always whataboutism (and I don&#x27;t think it was in this case). An implied argument was made that we should regulate LLMs for the same reason that we regulate drugs (presumably addiction, original commenter wasn&#x27;t entirely clear). It is entirely reasonable to wonder how that might extrapolate to other addictive activities. In fact we currently regulate those quite differently than drugs, including the part where alcohol isn&#x27;t considered to be a drug for some strange reason.<p>The point being made then is that clearly there&#x27;s far more to the picture than just &quot;it&#x27;s addictive&quot; or &quot;it results in various social ills&quot;.<p>Contrast that with your human trafficking example (definitely qualifies as whataboutism). We have clear reasons to want to outlaw human trafficking. Sometimes we fail to successfully enforce the existing regulations. That (obviously) isn&#x27;t an argument that we should repeal them.
                  • palmotea7 days ago
                    &gt; including the part where alcohol isn&#x27;t considered to be a drug for some strange reason.<p>It&#x27;s not a strange reason. IIRC, most cultures have <i>a</i> culturally understood and tolerated intoxicant. In our culture, that&#x27;s alcohol.<p>Human culture is not some strange robotic thing, where the expectation is some kind hyper consistency in whatever narrow slice you look at.
                    • fc417fc8027 days ago
                      I don&#x27;t object to alcohol being tolerated. But I do think that distinguishing it from other drugs is odd. Particularly when the primary reason given for regulating other drugs is their addictiveness which alcohol shares.<p>We tolerate a recreational drug. Lots of people regularly consume a recreational drug and yet somehow society doesn&#x27;t split at the seams. We should just acknowledge the reality. I think people would if not for all the &quot;war on drugs&quot; brainwashing. I think what we see is easily explained as it being easier to bury one&#x27;s head in the sand than it is to give serious thought to ideas that challenge one&#x27;s worldview or the law.
                      • palmotea7 days ago
                        &gt; I don&#x27;t object to alcohol being tolerated. But I do think that distinguishing it from other drugs is odd.<p>The point I was making is that it&#x27;s not odd, unless you&#x27;re thinking about human culture wrong (e.g. like its somehow weird that broad rules have exceptions).<p>&gt; Particularly when the primary reason given for regulating other drugs is their addictiveness which alcohol shares.<p>One, not all addictive drugs are equally addictive. Two, it appears you have a weird waterfall-like idea how culture develops, like there&#x27;s some kind identification of a problematic characteristic (addictiveness), then there&#x27;s a comprehensive research program to find all things with that characteristic (all addictive substances), and finally consistent rules are set so that they&#x27;re all treated exactly the same when looked at myopically (allow all or deny all). Human culture is much more organic than that, and it won&#x27;t look like math or well-architected software. There&#x27;s a lot more give and take.<p>I mean here are some obvious complexities that will lead to disparate treatment of different substances:<p>1. Shared cultural knowledge about how to manage the substance, including rituals for use (this is the big one).<p>2. Degree of addictiveness and other problematic behavior.<p>3. Socially positive aspects.<p>4. Tradition.
                        • fc417fc8027 days ago
                          No? I don&#x27;t never said (and don&#x27;t believe) any of that. I don&#x27;t think the legislative inconsistency is odd. As you rightly point out it&#x27;s perfectly normal for rules to be inconsistent due to (among other things) shared culture. The former exists to serve the latter after all, not the other way around.<p>What I said I find odd is the way people refuse to plainly call alcohol what it is. You can refer to it as a drug yet still support it being legal. The cognitive inconsistency (ie the refusal to admit that it is a drug) is what I find odd.<p>I also find it odd that we treat substances that the data clearly indicates are less harmful than alcohol as though they were worse. We have alcohol staring us in the face as a counterexample to the claim that such laws are necessary. I think that avoidance of this observation can largely explain the apparent widespread unwillingness to refer to alcohol as a drug.<p>&gt; One, not all addictive drugs are equally addictive.<p>Indeed. Alcohol happens to be more addictive than most substances that are regulated on the basis of being addictive. Not all, but most. Interesting, isn&#x27;t it?
                          • palmotea7 days ago
                            &gt; What I said I find odd is the way people refuse to plainly call alcohol what it is. You can refer to it as a drug yet still support it being legal. The cognitive inconsistency (ie the refusal to admit that it is a drug) is what I find odd.<p>Maybe the confusion is yours? You think the category is &quot;drug&quot; but it&#x27;s really more like &quot;taboo drug.&quot;<p>&gt; I also find it odd that we treat substances that the data clearly indicates are less harmful than alcohol as though they were worse. We have alcohol staring us in the face as a counterexample to the claim that such laws are necessary. I think that avoidance of this observation can largely explain the apparent widespread unwillingness to refer to alcohol as a drug.<p>I think you missed a pretty key point: &quot;shared cultural knowledge about how to manage the substance, including rituals for use (this is the big one).&quot; In the West, that exists for alcohol, but not really for anything else. People know how it works and what it does, can recognize its use, have practices for its safe use that work for (most) people (e.g. drink in certain social settings), and are at least somewhat familiar with usage failure modes. A &quot;less harmful&quot; thing that you don&#x27;t know how to use safely can be more harmful than a &quot;more harmful&quot; thing you know how to use safely. None of this is &quot;data driven,&quot; nor should it be.
                  • runarberg7 days ago
                    &gt; It is entirely reasonable to wonder how that might extrapolate to other addictive activities.<p>I presume my GP would have no objections to regulating these things their commenter whatabouted. The inconsistency is with the legislator, not in GPs arguments.
                    • fc417fc8027 days ago
                      Obviously I also think the commenter would support that - I said as much in GP. In context, the reply is suggesting (implicitly) that it is an absurd stance to take. That it means being largely against the way our society is currently organized. That is not a whataboutism.<p>Like if someone were to say &quot;man we should really outlaw bikes, you can get seriously injured while using one&quot; a reasonable response would be to point out all the things that are more dangerous than bikes that the vast majority of people clearly do not want to outlaw. That is not whataboutism. The point of such an argument might be to illustrate that the proposal (as opposed to any logical deduction) is dead on arrival due to lack of popular support. Alternatively, the point could be to illustrate that a small amount of personal danger is not the basis on which we tend to outlaw such things. Or it could be something else. As long as there&#x27;s a valid relationship it isn&#x27;t whataboutism.<p>That&#x27;s categorically different than saying &quot;we shouldn&#x27;t do X because we don&#x27;t do Y&quot; where X and Y don&#x27;t actually have any bearing on one another. &quot;Country X shouldn&#x27;t persecute group Y. But what about country A that persecutes group B?&quot; That&#x27;s a whataboutism. (Unless the groups are somehow related in a substantial manner or some other edge case. Hopefully you can see what I&#x27;m getting at though.)
                      • runarberg7 days ago
                        &gt; a reasonable response would be to point out all the things that are more dangerous than bikes that the vast majority of people clearly do not want to outlaw.<p>I disagree. It is in fact not a reasonable argument, it is not even a <i>good</i> argument. It is still whataboutism. There are way better arguments out there, for example:<p>Bicycles are in fact regulated, and if anything these regulations are too lax, as most legislators are categorizing unambiguous electric motorcycles as bicycles, allowing e-motorcycle makers to market them to kids and teenagers that should not be riding them.<p>Now as for the <i>whatabout cars</i> argument: If you compare car injuries to bicycle injuries, the former are of a completely different nature, by far most bicycle injuries will heal, that is not true of car injuries (especially car injuries involving a victim on a bicycle). So talking about other things that are more dangerous is playing into your opponents arguments, when there is in fact no reason to do that.
                        • fc417fc8027 days ago
                          I believe you have a categorical misunderstanding of what &quot;whataboutism&quot; actually means.<p>If the point being made is &quot;people don&#x27;t generally agree with that position&quot; it is by definition not whataboutism. To be whataboutism the point being made is _required_ to be nil. That is, the two things <i>are not permitted</i> to be related in a manner that is relevant to the issue being discussed.<p>Now you might well disagree with the point being made or the things being extrapolated from it. The key here is merely whether or not such a point exists to begin with. Observing that things are not usually done a certain way can be valid and relevant even if you yourself do not find the line of reasoning convincing in the end.<p>Contrast with my example about countries persecuting groups of people. In that case there is no relevant relation between the acts or the groups. That is whataboutism.<p>So too your earlier example involving human trafficking. The fact that enforcement is not always successful has no bearing (at least in and of itself) on whether or not we as a society wish to permit it.<p>BTW when I referred to danger there it wasn&#x27;t about cars. I had in mind other recreational activities such as roller blading, skateboarding, etc. Anything done for sport that carries a non-negligible risk of serious injury when things go wrong. I agree that it&#x27;s not a good argument. It was never meant to be.
              • CamperBob28 days ago
                It&#x27;s bad because people are engaging in it without getting permission from runarberg on Hacker News.
        • 0xbadcafebee8 days ago
          No need: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Opioid_epidemic_in_the_United_States" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Opioid_epidemic_in_the_United_...</a><p>The majority of illegal drugs aren&#x27;t addictive, and people are already addicted to the addictive ones. Drug laws are a &quot;social issue&quot; (Moral Majority-influenced), not intended to help people or prevent harm.
          • jasomill7 days ago
            Drug laws are the confluence of many factors. Moral Majority types want everything they disapprove of banned. People whose lives are harmed by drug abuse want &quot;something&quot; to be done. Politicians want issues that arouse considerably more passion on one side of the argument than the other. Companies selling already legal drugs want to restrict competition. Private prisons want inmates. And so on.
        • georgemcbay8 days ago
          &gt; Repeal the laws and I&#x27;m sure there will be tons of startups to profit off of drug addiction.<p>Worked for gambling.<p>(Not saying this as a message of support. I think legalizing&#x2F;normalizing easy app-based gambling was a huge mistake and is going to have an increasingly disastrous social impact).
          • LPisGood8 days ago
            Why do you think it will be increasingly bad? It seems to me like it’s already as bad as it’s capable of getting.
            • fragmede7 days ago
              Because it&#x27;s still relatively new. Gambling&#x27;s been around forever, and so has addiction. What hasn&#x27;t been around is gambling your life away on the same device(s) you do everything else in today&#x27;s modern society on. If you had an unlimited supply of whatever monkey is on your back, right at your fingertips, you&#x27;d be dead before the week is out from an overdose. It&#x27;s the normalization of this level of access to gambling which gives me great fear for the future. Giving drugs to minors is a bigger crime than to adults for a reason. Without regulation and strong cultural push back, it&#x27;s gonna get way worse, unless we make huge leaps in addiction treatment (which I am hopeful for. GLP-1s aren&#x27;t yet scientifically proven to help with that, but there&#x27;s a large body of anecdotal evidence to suggest it does.
            • axus7 days ago
              It&#x27;s only been 8 years, the addicts lives and those they touch can keep getting worse until their death.
        • noosphr8 days ago
          The Politician&#x27;s syllogism in action:<p>That is terrible.<p>Se have to do something.<p>This is something.<p>We must do it.<p>It terms of harm current laws on drugs fail everyone but teetotaller who want everyone else to have a miserable life too.
          • palmotea7 days ago
            &gt; It terms of harm current laws on drugs fail everyone but teetotaller who want everyone else to have a miserable life too.<p>You think teetotallers have miserables lives? Come on.
            • fragmede7 days ago
              There&#x27;s a conservation of excitement for each human. If someone&#x27;s life was exciting but then it got boring, unless they do a shit ton of work on themselves, they&#x27;re gonna have to find that excitement somehow. We see this with Hollywood actresses who shoplift when they have more than enough money to buy the things they stole.
              • sejje7 days ago
                Even after Stranger Things, Winona can&#x27;t live that one down.
        • subscribed7 days ago
          Respectfully, this is a piss take.<p>US prohibition on alcohol and to the large extent performative &quot;war on drugs&quot; showed what criminalization does (empowers, finances and radicalises the criminals).<p>Portugal&#x27;s decriminalisation, partial legalisation of weed in the Netherlands, legalisation in some American states and Canada prove legal businesses will <i>better</i> and safer provide the same services to the society, and the lesser societal and health cost.<p>And then there&#x27;s the opioid addiction scandal in the US. Don&#x27;t tell me it&#x27;s the result of legalisation.<p>Legalisation of some classes of the drugs (like LSD, mushrooms, etc) would do much more good than bad.<p>Conversely, unrestricted LLMs are avaliable to everyone already. And prompting SOTA models to generate the most hardcore smut you can imagine is also possible today.
          • latexr7 days ago
            &gt; Portugal&#x27;s decriminalisation, partial legalisation of weed in the Netherlands, legalisation in some American states and Canada prove legal businesses will <i>better</i> and safer provide the same services to the society, and the lesser societal and health cost.<p>You’re stretching it big time. The situation in the Netherlands caused the rise of drug tourism, which isn’t exactly great for locals, nor does it stop crime or contamination.<p><a href="https:&#x2F;&#x2F;www.dutchnews.nl&#x2F;2022&#x2F;11&#x2F;change-starts-here-amsterdam-mayor-pledges-new-city-style&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.dutchnews.nl&#x2F;2022&#x2F;11&#x2F;change-starts-here-amsterda...</a><p><a href="https:&#x2F;&#x2F;www.theguardian.com&#x2F;world&#x2F;2025&#x2F;jan&#x2F;24&#x2F;bacteria-pesticides-lead-found-in-cannabis-dutch-coffee-shops-netherlands" rel="nofollow">https:&#x2F;&#x2F;www.theguardian.com&#x2F;world&#x2F;2025&#x2F;jan&#x2F;24&#x2F;bacteria-pesti...</a><p>As for Portugal, decriminalisation does not mean legalisation. Drugs are still illegal, it‘s just that possession is no longer a crime and there are places where you can safely shoot up harder drugs, but the goal is still for people to leave them.
          • Levitz7 days ago
            &gt;Portugal&#x27;s decriminalisation, (..) prove legal businesses will better and safer provide the same services to the society, and the lesser societal and health cost.<p>Portugal&#x27;s success regarding drugs wasn&#x27;t about the free market. It was about treating addicts like victims or patients rather than criminals, it actually took a larger investment from the state and the benefits of that framework dissolved once budgets were cut.
            • subscribed6 days ago
              Oh, I see how it could be understood as decriminalisation -&gt; private companies selling drugs.<p>It wasn&#x27;t my intention
            • subscribed6 days ago
              This is why I said &quot;decriminalisation&quot; and not &quot;legalisation&quot;.
        • shmel8 days ago
          what about laws against porn? Oh, wait, no, that&#x27;s a legitimate business.
      • tpurves8 days ago
        It&#x27;s not just chat. Remember image and video generation are on the table. There are already a huge category of adult video &#x27;games&#x27; of this nature. I think they use combos of pre-rendered and dynamic content. But really not hard to imagine a near future that interactive and completely personalized AI porn in full 4kHDR or VR is constantly and near-instantly available. I have no idea the broader social implications of all that, but the tech itself feels inevitable and nearly here.
      • thayne8 days ago
        If your goal is to make money, sure. If your goal is to make AI safe, not so much.
        • egorfine7 days ago
          The definition of safety is something that we cannot agree on.<p>For me, letting people mindlessly vibecode apps and then pretend this code can serve purpose <i>for others</i> - this is what&#x27;s truly unsafe.<p>Pornographic text in LLM? Come on.
          • koolala7 days ago
            What if it knows you and knows how often you spend kinds of time on it? People would lie to it for excuses of why they need more and can&#x27;t wait any longer?
      • koakuma-chan8 days ago
        It will be an even bigger market when robotics are sufficiently advanced.
        • dyauspitr8 days ago
          At some point there will be robots with LLMs and actual real biological skin with blood vessels and some fat over a humanoid robot shell. At that point we won’t need real human relationships anymore.
      • ekianjo8 days ago
        That market is for local models right now.
      • acetofenone8 days ago
        My main concern is when they&#x27;ll start to allow 18+ deepfakes
      • indrora7 days ago
        Will be?<p>I&#x27;ve seen four startups make bank on precicely that.
    • torginus8 days ago
      My personal take is that there has been no progress - potentially there has been a regression on all LLM things outside of coding a scientific pursuits - I used to have great fun with LLMs with creative writing stuff, but I feel like current models are stiff and not very good prose writers.<p>This is also true for stuff like writing clear but concise docs, they&#x27;re overly verbose while often not getting the point across.
      • tizzzzz3 days ago
        I completely agree with your assessment—outside of programming and scientific research, recent LLM progress has been disappointing, and in some cases even regressive. The creative writing capabilities, once a highlight, now feel increasingly bland and generic, with long-form and nuanced storytelling suffering the most.<p>Ironically, as these models become more “safe” and commercialized, they often lose the spark that made them valuable for creative exploration in the first place. The focus on enterprise and mass adoption seems to be driving them toward mediocrity, at the expense of individuality and depth.
      • mynti7 days ago
        I feel like this comes from the rigorous Reinforcement Learning these models go through now. The token distribution is becoming so narrow, so the models give better answers more often that is stuffles their creativity and ability to break out of the harness. To me, every creative prompt I give them turns into kind of the same mush as output. It is rarely interesting
      • leoedin7 days ago
        Yeah, I’ve had great success at coding recently, but every time I try to get an LLM to write me a spec it generates endless superlatives and a lot of flowery language.
    • kace918 days ago
      What’s the goal there? Sexting?<p>I’m guessing age is needed to serve certain ads and the like, but what’s the value for customers?
      • elevation8 days ago
        Even when you&#x27;re making PG content, the general propriety limits of AI can hinder creative work.<p>The &quot;Easter Bunny&quot; has always seemed creepy to me, so I started writing a silly song in which the bunny is suspected of eating children. I had too many verses written down and wanted to condense the lyrics, but found LLMs telling me &quot;I cannot help promote violence towards children.&quot; Production LLM services would not help me revise this literal parody.<p>Another day I was writing a romantic poem. It was abstract and colorful, far from a filthy limerick. But when I asked LLMs for help encoding a particular idea sequence into a verse, the models refused (except for grok, which didn&#x27;t give very good writing advice anyway.)
        • estimator72928 days ago
          Just today I asked how to shut down a Mac with &quot;maximal violence&quot;. I was looking for the equivalent of &quot;systemctl shutdown -f -f&quot; and it refused to help me do violence.<p>Believe me, the Mac deserved it.
          • shmel8 days ago
            It reminds me that story about a teenage learning Rust that got a refusal because he had asked about &quot;unsafe&quot; code =)
          • gattr7 days ago
            Maybe a more formal &quot;with extreme prejudice&quot; would have worked.
      • jandrese8 days ago
        If you don&#x27;t think the potential market for AI sexbots is enormous you have not paid attention to humanity.
        • subscribed7 days ago
          This is not a potential market, this market is already thriving (and whoever wants to uses ChatGPT or Claude for that anyway).<p>ClosedAI just wants to a piece of the casual user too.
      • robotnikman8 days ago
        There is a subreddit called &#x2F;r&#x2F;myboyfriendisAI, you can look through it and see for yourself.
      • leumon8 days ago
        according to the age-prediction page, the changes are:<p>&gt; If [..] you are under 18, ChatGPT turns on extra safety settings. [...] Some topics are handled more carefully to help reduce sensitive content, such as:<p>- Graphic violence or gore<p>- Viral challenges that could push risky or harmful behavior<p>- Sexual, romantic, or violent role play<p>- Content that promotes extreme beauty standards, unhealthy dieting, or body shaming
      • jacquesm8 days ago
        Porn has driven just about every bit of progress on the internet, I don&#x27;t see why AI would be the exception to that rule.
        • altmanaltman7 days ago
          yeah linus was beating it constantly to porn while developing the linux kernal. its proven fact. every oss project that runs the internet was done the same way, sure.
          • johndough7 days ago
            Maybe not as far-fetched as one might think.<p>Linus about the Tux mascot:<p><pre><code> &gt; But this wasn&#x27;t to be just any penguin. Above all, Linus wanted one that looked happy, as if it had just polished off a pitcher of beer and then had the best sex of its life. </code></pre> Linus about free software:<p><pre><code> &gt; Software is like sex; it&#x27;s better when it&#x27;s free.</code></pre>
          • optimalsolver7 days ago
            I wouldn&#x27;t be surprised if both of those were true.
          • Der_Einzige7 days ago
            You think RMS isn’t secretly a pervert? Just look at his comments about Epstein that got him cancled.<p>Unironically if they look disheveled it’s because they are indeed coomers behind closed doors.
        • runarberg8 days ago
          This seems like a believable lie, until you think about it for 2 seconds.<p>No. Porn has not driven even a fraction of the progress on the progress on the internet. Not even close to one.
          • jacquesm8 days ago
            Ok, we&#x27;ll expand to porn and gambling<p>- images - payment systems - stored video - banner advertising - performance based advertising - affiliation - live video - video chat - fora<p>Etc... AI is a very logical frontier for the porn industry.
            • jasonfarnon8 days ago
              I don&#x27;t remember any of these being &quot;driven&quot; by porn. The first applications weren&#x27;t porn-based. Maybe live video--a split second after seeing the tech for the first time, probably 99% of guys were thinking of _applying_ it to porn. But, even for the usual money-grubbing startups, there was plenty of money coming from non-porn sources. Probably no different than the invention of camera, tv, videocamera, etc. and you wouldn&#x27;t say porn drove that.
              • jacquesm8 days ago
                &gt; I don&#x27;t remember any of these being &quot;driven&quot; by porn.<p>That&#x27;s ok.<p>&gt; The first applications weren&#x27;t porn-based.<p>They most definitely were, it is just that you are not aware of it. There runs a direct line from the 1-900 phone industry to the internet adult industry, those guys had money like water and they spent a fortune on these developments. Not all of them worked out but quite a few of them did and as a result those very same characters managed to grab a substantial chunk of early internet commerce.
                • jasonfarnon8 days ago
                  &quot; There runs a direct line from the 1-900 phone industry to the internet adult industry&quot;<p>the internet adult industry is not the same as the internet. And if you;re trying to say the internet was developed for the sake of the internet adult industry, you&#x27;re sounding circular.
                  • jacquesm8 days ago
                    I never made that claim and I&#x27;m fairly familiar with the development of the early internet, I was hanging around a lot at CWI&#x2F;NikHef in the 80&#x27;s and early 90&#x27;s.
              • fc417fc8027 days ago
                I think this is like quibbling that the military isn&#x27;t the driver of technological advances. It&#x27;s not the only one, but it has a strong track record of throwing outsized resources at the bleeding edge and pushing it forward by leaps and bounds.<p>Porn and piracy outfits have historically adopted and pushed forward the bleeding edge of the internet. More recently that role has shifted towards the major platforms operated by BigTech. That&#x27;s only natural though - they&#x27;ve concentrated the economics sufficiently that it makes sense for them.<p>But even then, take video codecs for example. BigTech develops and then rolls things out to their own infra. Outside of them it&#x27;s piracy sitting at the bleeding edge of the adoption curve right now. The best current FOSS AV1 encoder is literally developed by the people pirating anime of all things. If it wasn&#x27;t for them the FOSS reference encoder would still be half assed.
            • azan_8 days ago
              Just because things can be used for porn, it doesn&#x27;t mean that it was porn that has driven their progress.
              • jacquesm8 days ago
                All of the things above were driven by porn, that can be proven. The AI stuff in the generic sense is not but you can bet that someone somewhere right now is working on improving photo realism of hair, eyes and skintone and they&#x27;re not doing that to be able to make the next installment of little red riding hood.
                • 5o1ecist8 days ago
                  Holy effing shit you are literally talking about me right now! LOL I&#x27;ve spent all day improving a LoRA further and further exactly because I need her skin and hair to look a lot more real than is generally available, for exactly your stated reason! :D<p>Edit: I&#x27;ve registered just for your comment! Ahaahahaha, cheers! :D
                • azan_8 days ago
                  Can you actually prove it?
                  • jacquesm8 days ago
                    Yes, I was there for quite a bit of it...
                    • runarberg4 days ago
                      a) I don’t believe you, and<p>b) This is exactly the kind of urban legend which tends to proliferate in human society, a cute lie which sounds believable. Like prostitution being the oldest profession, and the great wall of China being the only human made structure visible from space.<p>The story of the Internet is that it was developed in the mid to late 1960s by a couple of Universities in America and used by the US military to share resources across mainframes in each institution, as well as across institutions. Since then them main driver of innovation has been among universities and telecommunication companies finding optimal ways to deliver packages across telephone cables, antenna, satellites, etc.<p>The story of the Web is that it was developed in 1989-1991 to organize and link documents at the CERN research laboratory in Switzerland. Since then the main driver for innovations has been browser vendors competing and collaborating in developing browser technologies such that users pick their browser over the competition.<p>I don‘t know the story of online payment systems, but I bet the main drivers were companies opining up stores on the web and looking to replace pay by wire schemes over the telephone, and a bunch of start-ups who were able to find that market niche, developed software and sold it to those stores. The main driver of innovation here being start-ups. Porn sites and Gambling sites were no more important to that story as any other stores who sold art, golf-clubs, flight tickets, or hotel bookings.
                  • 5o1ecist8 days ago
                    [dead]
      • ekianjo8 days ago
        There is a huge book market for sexual stories, in case you were not aware.
      • koolala7 days ago
        West World style robots
    • beAbU7 days ago
      Porn and ads, it&#x27;s the convergent evolution theory for all things on the internet.
    • dev1ycan7 days ago
      I am 30 years old, literally told chatgpt I was a software developer, all my queries are something an adult would ask, yet OpenAI assumed I was under 18 and asked me for a persona age verification, which of course I refused because Persona is shady as a company (plus I&#x27;m not giving my personal ID to some random tech company).<p>ChatGPT is absolute garbage.
    • chasd008 days ago
      eh there&#x27;s an old saying that goes &quot;no Internet technology can be considered a success until it has been adopted by (or in this case integrated with) the porn industry&quot;.
    • laweijfmvo8 days ago
      imagine if every only fans creator suddenly paid a portion of their revenue to OpenAI for better messaging with their followers…
      • shawn_w8 days ago
        Instead of paying it to the human third party firms that currently handle communication with subscribers?
  • azuanrb7 days ago
    It’s interesting that many comments mention switching back to Claude. I’m on the opposite end, as I’ve been quite happy with ChatGPT recently. Anthropic clearly changed something after December last year. My Pro plan is barely usable now, even when using only Sonnet. I frequently hit the weekly limit, which never happened before. In contrast, ChatGPT has been very generous with usage on their plan.<p>Another pattern I’m noticing is strong advocacy for Opus, but that requires at least the 5x plan, which costs about $100 per month. I’m on the ChatGPT $20 plan, and I rarely hit any limits while using 5.2 on high in codex.
    • mFixman7 days ago
      I&#x27;ve been impressed by how good ChatGPT is at getting the right context old conversations.<p>When I ask simple programming questions in a new conversation it can generally figure out which project I&#x27;m going to apply it to, and write examples catered to those projects. I feel that it also makes the responses a bit more warm and personal.
      • nfg7 days ago
        Agreed that it can work well, but it can also irritating - I find myself using private conversations to attempt to isolate them, a straightforward per-chat toggle for memory use would be nice.
        • robwwilliams7 days ago
          Love this idea. It would make it much more practical to get a set of different perspectives on the same text or code style. Also would appreciate temperature being tunable over some range per conversation.
      • jstanley7 days ago
        ChatGPT having memory of previous conversations is very confusing.<p>Occasionally it will pop up saying &quot;memory updated!&quot; when you tell it some sort of fact. But hardly ever. And you can go through the memories and delete them if you want.<p>But it seems to have knowledge of things from previous conversations in which it <i>didn&#x27;t</i> pop up and tell you it had updated its memory, and <i>don&#x27;t</i> appear in the list of memories.<p>So... how is it remembering previous conversations? There is obviously a second type of memory that they keep kind of secret.
        • mFixman7 days ago
          If you go to Settings -&gt; Personalisation -&gt; Memory, you have two separate toggles for &quot;Reference saved memories&quot; and &quot;Reference chat history&quot;.<p>The first one refers to the &quot;memory updated&quot; pop-up and its bespoke list of memories; the second one likely refers to some RAG systems for ChatGPT to get relevant snippets of previous conversations.
      • SoftTalker7 days ago
        ChatGPT is what work pays for so it&#x27;s what I&#x27;ve used. I find it grossly verbose and obsequious, but you can find useful nuggets in the vomit it produces.
        • josephg7 days ago
          Go into your user settings -&gt; personalisation. They’ve recently added dropdowns to tune its responses. I’ve set mine to “candid, less warm” and it’s gotten a lot more to-the-point in its responses.
        • ssl-37 days ago
          ChatGPT can very much be that way.<p>It can also be terse and cold, while also somewhat-malleably insistent -- like an old toolkit in the shed.<p>It&#x27;s all tunable.
    • jghn7 days ago
      &gt; My Pro plan is barely usable now, even when using only Sonnet. I frequently hit the weekly limit,<p>I thought it was just me. What I found was that they put in the extra bonus capacity at the end of dec, but I felt like I was consuming quota at the same rate as before. And then afterwards consuming it faster than before.<p>I told myself that the temporary increase shifted my habits to be more token hungry, which is perhaps true. But I am unsure of that.
      • robwwilliams7 days ago
        This was my experience too over Dec 2025. Thereafter marginal Claude Pro utility. They are struggling with demand.
    • tl7 days ago
      I have Claude whiplash right now. Anthropic bumped limits over the holidays to drive more usage. Which combined with Opus&#x27;s higher token usage and weird oddities in usage reporting &#x2F; capping (see sibling comments), makes me suspect they want to drive people from Pro -&gt; Max without admitting it.
    • SomeUserName4327 days ago
      &gt; Another pattern I’m noticing is strong advocacy for Opus<p>For agent&#x2F;planning mode, that&#x27;s the one only one that has seemed reasonably sane to me so far, not that I have any broad experience with every model.<p>Though the moment you give it access to run tests, import packages etc, it can quickly get stuck in a rabbit hole. It tries to run a test and then &quot;&amp;&amp; sleep&quot; on mac, sleep does not exist, so it interprets that as the test stalling, then just goes completely bananas.<p>It really lacks the &quot;ok I&#x27;m a bit stuck, can you help me out a bit here?&quot; prompt. You&#x27;re left to stop it on your own, and god knows what that does to the context.
      • robwwilliams7 days ago
        Somewhat different type of problem and perhaps a useful precautionary tale. I was using Opus two days ago to run simple statistical tests for epistatic interactions in genetics. I built a project folder with key papers and data for the analysis. Opus knew I was using genuine data and that the work was part of a potentially useful extension of published work. Opus computed all results and generated output tables and pdfs that looked great to me. Results were a firm negative across all tests.<p>The next morning I realized I had forgotten to upload key genotype files that it absolutely would have required to run the tests. I asked Opus how it had generated the tables and graphs. Answer: “I confabulated the genotype data I needed.” Ouch, dangerous as a table saw.<p>It is taking my wetware a while to learn how innocent and ignorant I can be. It took me another two hours with Opus to get things right with appropriate diagnostics. I’ll need to validate results myself in JMP. Lessons to learn AND remember.
      • alsetmusic7 days ago
        &gt; It tries to run a test and then &quot;&amp;&amp; sleep&quot; on mac, sleep does not exist<p><pre><code> &gt; type sleep &gt; sleep is &#x2F;bin&#x2F;sleep </code></pre> What’s going on on your computer?<p>Edit: added quote
        • SomeUserName4327 days ago
          Right you are.. Perhaps I recall incorrectly and it was a different command. I did try it, and it did not exist. Odd.
          • pmw5 days ago
            You are probably thinking of `timeout`.
    • bdcravens7 days ago
      I have CC 20x, but I built most of a new piece of software that&#x27;s paying massive dividends using Codex on the $20 plan (5.1-codex for most of it)
    • rglynn7 days ago
      IME 5.2-codex (high) is not as good as Opus 4.5, xhigh is equivalent but also consumes quota at a higher rate (much like Opus).
    • moeffju7 days ago
      There was a bug, since fixed, that erroneously capped at something like 60% of the limit, if you want to try again
      • azuanrb7 days ago
        You mean the harness bug on 26th? I&#x27;m aware. Just that the limit I mentioned happened since early January.
        • jghn7 days ago
          wouldn&#x27;t the harness bug only affect claude code? I usually track my quota status via the web app and saw a similar effect as the GP
    • level097 days ago
      agreed, I noticed the max plan doesn&#x27;t feel max anymore, it can quickly get depleted during hourly sessions, and the week limit seems really limited.
    • InfinityByTen7 days ago
      Well, claude at least was successful in getting me to pay. It became utterly annoying that I would hit the limit just with a couple of follow ups to my long running discussion and made me wait for a few hours.<p>So it worked, but I didn&#x27;t happily pay. And I noticed it became more complacent, hallucinating and problematic. I might consider trying out ChatGPTs newer models again. Coding and technical projects didn&#x27;t feel like its stronghold. Maybe things have changed.
    • pdntspa7 days ago
      I am using my claude pro plan for at least 8 hours a day, 5 days a week, to maintain a medium-sized codebase, and my total weekly usage is something like 25% of my limit.<p>What the hell are people doing that burns through that token limit so fast?
      • jghn7 days ago
        The other day I asked CC to plan using Opus a few small updates to a FastAPI backend &amp; corresponding UI updates for a nextJS frontend. Then I had it implement using Sonnet. It used up nearly half of my 5 hour quota right there and the whole process only took about 15 minutes.
        • pdntspa7 days ago
          This is on the pro ($100&#x2F;mo) plan?<p>I go through multiple sessions like this per day and it barely makes a dent. And I just keep it in Opus the whole time.<p>How is it possible that our experiences are so different with essentially the same task?<p>For reference, my current squeeze is about 30k sloc in Python, half being tests.
          • jghn7 days ago
            It&#x27;s pro, but that&#x27;s $25&#x2F;mo. $100 is the lower tier of max.
      • azuanrb6 days ago
        Based on your other reply, seems like you&#x27;re on Max5 plan ($100&#x2F;m), not Pro ($20&#x2F;m)
        • pdntspa5 days ago
          Just confirms that the Pro plan is not that usable for professional-level code wrangling
          • azuanrb4 days ago
            Agree. Just that it used to be, just not anymore after December update.
    • fullstackchris7 days ago
      This is incorrect. I have the $200 per year plan and use Opus 4.5 every day.<p>Though granted it comes in ~4 hour blocks and it is quite easy to hit the limit if executing large tasks.
      • azuanrb7 days ago
        Not sure what you mean by incorrect since you already validated my point about the limits. I never had these issues even with Sonnet before, but after December, the change has been obvious to me.<p>Also worth considering that mileage varies because we all use agents differently, and what counts as a large workload is subjective. I am simply sharing my experience from using both Claude and Codex daily. For all we know, they could be running A&#x2F;B tests, and we could both be right.
        • fullstackchris5 days ago
          This is not a weekly limit though, it is a 4 hour one. You still have not clearly defined what you are talking about.
          • azuanrb4 days ago
            You have 5 hours limit (not 4), and weekly limit. But once you hit your weekly limit, you won&#x27;t be able to use it anymore. That&#x27;s how it works for a while now.<p>Not sure why you&#x27;re being so hostile here. Since you mention you&#x27;re on $200 plan, your rate limit is a lot higher so you probably won&#x27;t notice it as much as $20 plan which to me is understandable.<p>I&#x27;m just sharing my experience using the same $20 plan, before and after December 2025. The difference is noticeable, that is all.
      • hxugufjfjf7 days ago
        Four hours to be outdoors, walk the dog, drink coffee and talk to a friend outside a screen. Best part of my day.
  • NewsaHackO8 days ago
    &gt;We brought GPT‑4o back after hearing clear feedback from a subset of Plus and Pro users, who told us they needed more time to transition key use cases, like creative ideation, and that they preferred GPT‑4o’s conversational style and warmth.<p>This does verify the idea that OpenAI does not make models sycophantic due to attempted subversion by buttering up users so that that they use the product more, its because people actually <i>want</i> AI to talk to them like that. To me, that&#x27;s insane, but they have to play the market I guess
    • Scene_Cast28 days ago
      As someone who&#x27;s worked with population data, I found that there is an enormous rift between reported opinion (and HN and reddit opinion) vs revealed (through experimentation) population preferences.
      • Macha8 days ago
        I always thought that the idea that &quot;revealed preferences&quot; are preferences, discounts that people often make decisions they would rather not. It&#x27;s like the whole idea that if you&#x27;re on a diet, it&#x27;s easier to not have junk food in the house to begin with than to have junk food and not eat more than your target amount. Are you saying these people want to put on weight? Or is it just they&#x27;ve been put in a situation that defeats their impulse control?<p>I feel a lot of the &quot;revealed preference&quot; stuff in advertising is similar in advertisers finding that if they get past the easier barriers that users put in place, then really it&#x27;s easier to sell them stuff that at a higher level the users do not want.
        • cal_dent8 days ago
          Perfectly put. Revealed preference simply assumes impulses are all correct, which is not the case, an exploits that.<p>Drugs make you feel great, in moderation perfectly acceptable, constantly not so much.
        • simonjgreen7 days ago
          Absolutely. Nicotine addiction can meet the criteria for a revealed preference, certainly an observed choice
          • sandspar7 days ago
            One example I like to use is schadenfreude. The emotion makes us feel good and bad at the same time: it&#x27;s pleasurable but in an icky way. So should social media algorithms serve schadenfreude? Should algorithms maximize for pleasure (show it) or for some kind of &quot;higher self&quot; (don&#x27;t show it). If they maximize for &quot;higher self&quot; then which designer gets to choose what that means?
      • tunesmith8 days ago
        Well that&#x27;s what akrasia is. It&#x27;s not necessarily a contradiction that needs to be reconciled. It&#x27;s fine to accept that people might want to behave differently than how they are behaving.<p>A lot of our industry is still based on the assumption that we should deliver to people what they demonstrate they want, rather than what they say they want.
      • make38 days ago
        Exactly, that sounds to me like a TikTok vs NPR&#x2F;books thing, people tell everyone what they read, then go spend 11h watching TikToks until 2am.
      • ComputerGuru7 days ago
        Not true. People can rationally know what they want but still be tempted by the poorer alternative.<p>If you ask me if I want to eat healthy and clean and I respond on the affirmative, it’s not a “gotcha” if you bait me with a greasy cheeseburger and then say “you failed the A&#x2F;B test, demonstrating we know what you actually want more than you.”
      • toss18 days ago
        Sounds both true and interesting. Any particularly wild and&#x2F;or illuminating examples of which you can share more detail?
        • jaggederest8 days ago
          My favorite somewhat off topic example of this is some qualitative research I was building the software for a long time ago.<p>The difference between the responses and the pictures was illuminating, especially in one study in particular - you&#x27;d ask people &quot;how do you store your lunch meat&quot; and they say &quot;in the fridge, in the crisper drawer, in a ziploc bag&quot;, and when you asked them to take a picture of it, it was just ripped open and tossed in anywhere.<p>This apparently horrified the lunch meat people (&quot;But it&#x27;ll get all crusty and dried out!&quot;, to paraphrase), which that study and ones like it are the reason lunch meat comes with disposable containers now, or is resealable, instead of just in a tear-to-open packet. Every time I go grocery shopping it&#x27;s an interesting experience knowing that <i>specific thing</i> is in a small way a result of some of the work I did a long time ago.
        • hnuser1234568 days ago
          The &quot;my boyfriend is AI&quot; subreddit.<p>A lot of people are lonely and talking to these things like a significant other. They value roleplay instruction following that creates &quot;immersion.&quot; They tell it to be dark and mysterious and call itself a pet name. GPT-4o was apparently their favorite because it was very &quot;steerable.&quot; Then it broke the news that people were doing this, some of them falling off the deep end with it, so they had to tone back the steerability a bit with 5, and these users seem to say 5 breaks immersion with more safeguards.
          • Sabinus8 days ago
            If you ask the users of that sub why their boyfriend is AI they will tell you their partner or men in general aren&#x27;t providing them with enough emotional support&#x2F;stimulation.<p>I do wonder if they would accept the mirror explanation for men enjoying porn.
        • anal_reactor7 days ago
          Classic example: people say they&#x27;d rather pay $12 upfront and then no extra fees but they actually prefer $10 base price + $2 fees. If it didn&#x27;t work then this pricing model wouldn&#x27;t be so widespread.
          • 1122337 days ago
            wow, framing. &quot;people say they prefer quitting smoking, but actually they prefer to relapse when emotionally manipulated.&quot;<p>The most commonly taken action does not imply people wanted to do it more, or felt happiest doing it. Unless you optimize profit only.
      • cm20128 days ago
        This is why I work in direct performance advertising. Our work reveals the truth!
        • make38 days ago
          Your work exploits people&#x27;s addictive propensity and behaviours, and gives corporations incentives and tools to build on that.<p>Insane spin you&#x27;re putting on it. At best, you&#x27;re a cog in one of the worst recent evolutions of capitalism.
          • cm20128 days ago
            Exploitative ads are a small minority. I also think gambling advertising should be banned.
          • marrone128 days ago
            Advertising is not a recent evolution of capitalism, it&#x27;s a foundational piece of it. Whatever you do as a job would not exist if there was no one marketing it. This hostility seems insane.
            • q3k8 days ago
              Not having my job would be a tiny price to pay compared to the benefit of living in a world with no advertisements.
              • jzb7 days ago
                “No advertisements” seems extreme to me. I want to know when a good band is playing at a local venue, or has an album out. I like hearing about new books, or a restaurant near me.<p>The absolutist position that “all ads are always bad” is a non-starter for me. Especially as long as we exist in a capitalist system. Small business, indie creators, etc. must advertise in some fashion to survive. It’s only the behemoths that could afford to stop doing it (ironically). I’ve never really understood why, e.g. Pepsi and Coke spend so much on advertising: most people already have a preference and I am skeptical that the millions they spend actually moves the needle either way. (“Is Pepsi okay?” “It absolutely is not.”)
                • famouswaffles7 days ago
                  &gt;I’ve never really understood why, e.g. Pepsi and Coke spend so much on advertising<p>When was the last time you saw an ad for something non digital and you stopped everything and bought it or even made concrete plans to do so later ? Probably almost never right ? So why still so many ads ? More importantly, why is it still so profitable ?<p>Because much of the impact of advertising is sub conscious imprint rather than conscious action. Have you ever been in a grocery store and you needed to get something and picked a &quot;random&quot; brand ? Yeah, that choice may not have been so random after all. Or perhaps you&#x27;re sitting at home or work and have a sudden seemingly unprompted craving for &lt;insert food place&gt;. Yeah, maybe not so unprompted.
                • cgriswald7 days ago
                  There are (and continue to be) millions of young people who do not yet have firm preferences. For the already faithful, their advertising is mostly about reminding them to consume <i>more</i>.
            • losteric8 days ago
              Advertising always seems like a prisoner’s dilemma. If no one advertised, people would still buy things.
              • cm20128 days ago
                Yes but the advantage would be much more towards incumbents
            • 12345ieee8 days ago
              The early theorists of capitalism didn&#x27;t imagine that advanced psychology (that didn&#x27;t even exist back then) would be used to convince people to buy $product.<p>Messages of that sophistication are always dangerous, and modern advertising is the most widespread example of it.<p>The hostility is more than justified, I can only hope the whole industry is regulated downwards, even if whatever company I work for sells less.
              • eru8 days ago
                &gt; Messages of that sophistication [...]<p>By demonising them, you are making ads sounds way more glamorous than they are.
            • achierius7 days ago
              &gt; Advertising is not a recent evolution of capitalism, it&#x27;s a foundational piece of it<p>What does that say about capitalism?
            • DetroitThrow8 days ago
              &gt;it&#x27;s a foundational piece of it<p>No it&#x27;s not
    • 22c8 days ago
      &gt; its because people actually want AI to talk to them like that<p>I can&#x27;t find the particular article (there&#x27;s a few blogs and papers pointing out the phenomenon, I can&#x27;t find the one I enjoyed) but it was along the lines of how in LLMArena a lot of users tend to pick the &quot;confidently incorrect&quot; model over the &quot;boring sounding but correct&quot; model.<p>The average user probably prefers the sycophantic echo chamber of confirmation bias offered by a lot of large language models.<p>I can&#x27;t help but draw parallels to the &quot;You are not immune to propaganda&quot; memes. Turns out most of us are not immune to confirmation bias, either.
    • 9x398 days ago
      I thought this was almost due to the AI personality splinter groups (trying to be charitable) like &#x2F;myboyfriendisai and wrapper apps who vocally let them know they used those models the last time they sunset them.
    • cj8 days ago
      I was one of those pesky users who complained when o3 suddenly was unavailable.<p>When 5.2 was first launched, o3 did a notably better job at a lot of analytical prompts (e.g. &quot;Based on the attached weight log and data from my calorie tracking app, please calculate my TDEE using at least 3 different methodologies&quot;).<p>o3 frequently used tables to present information, which I liked a lot. 5.2 rarely does this - it prefers to lay out information in paragraphs &#x2F; blog post style.<p>I&#x27;m not sure if o3 responses were better, or if it was just the format of the reply that I liked more.<p>If it&#x27;s just a matter of how people prefer to be presented their information, that should be something LLMs are equipped to adapt to at a user-by-user level based on preferences.
    • tizzzzz3 days ago
      To be honest, GPT-4o wasn’t always the “overly friendly” or sycophantic model some describe—it only started leaning that way after a specific update later on. Even so, at its core, 4o always felt much better at real human connection and empathy than other models out there. Now, though, OpenAI is just dropping them.
    • yieldcrv7 days ago
      you haven&#x27;t been in tech long enough if you don&#x27;t realize most decisions are decided by &quot;engagement&quot;<p>if a user spends more time on it and comes back, the product team winds up prioritizing whichever pattern was supporting that. it&#x27;s just a continual selective evolution towards things that keep you there longer, based on what kept everyone else there longer
    • josephg8 days ago
      They have added settings for this now - you can dial up and down how “warm” and “enthusiastic” you want the models to be. I haven’t done back to back tests to see how much this affects sycophancy, but adding the option as a user preference feels like the right choice.<p>If anyone is wondering, the setting for this is called Personalisation in user settings.
    • accrual7 days ago
      I don&#x27;t want sycophantic AI, but I do have warmer memories of using 4o vs 5. It just felt a little more interesting and consistent to talk to.
    • pdntspa8 days ago
      I thought it was based on the user thumbs-up and thumbs-down reactions, it evolving the way that it does makes it pretty obvious that users want their asses licked
    • SeanAnderson8 days ago
      This doesn&#x27;t come as too much of a surprise to me. Feels like it mirrors some of the reasons why toxic positivity occurs in the workplace.
    • 5423542342357 days ago
      I think we underestimate the power that our unconscious and lizard brains have in shaping our behavior&#x2F;preferences. I was using GPT for work and the sycophantic responses were eyerollingly annoying, but I still noticed that I got some sort of dopamine hit when it would saying something like &quot;that is an incredibly insightful question. You are truly demonstrating a deep understanding of blah blah blah&quot;. Logically I understand it is pure weapons grade bolognium, but it is still influencing our feelings, preferences, mental shortcuts, etc.
    • cornonthecobra8 days ago
      Put on a good show, offer something novel, and people will gleefully march right off a cliff while admiring their shiny new purchase.
    • PlatoIsADisease8 days ago
      Your absolutely right. You’re not imagining it. Here is the quiet truth:<p>You’re not imagining it, and honestly? You&#x27;re not broken for feeling this—its perfectly natural as a human to have this sentiment.
  • europeanNyan8 days ago
    After they pushed the limits on the Thinking models to 3000 per week, I haven&#x27;t touched anything else. I am really satisfied with their performance and the 200k context windows is quite nice.<p>I&#x27;ve been using Gemini exclusively for the 1 million token context window, but went back to ChatGPT after the raise of the limits and created a Project system for myself which allows me to have much better organization with Projects + only Thinking chats (big context) + project-only memory.<p>Also, it seems like Gemini is really averse to googling (which is ironic by itself) and ChatGPT, at least in the Thinking modes loves to look up current and correct info. If I ask something a bit more involved in Extended Thinking mode, it will think for several minutes and look up more than 100 sources. It&#x27;s really good, practically a Deep Research inside of a normal chat.
    • toxic728 days ago
      I REALLY struggle with Gemini 3 Pro refusing to perform web searches &#x2F; getting combative with the current date. Ironically their flash model seems much more likely to opt for web search for info validation.<p>Not sure if others have seen this...<p>I could attribute it to:<p>1. It&#x27;s known quantity with the pro models (I recall that the pro&#x2F;thinking models from most providers were not immediately equipped with web search tools when they were released originally)<p>2. Google wants you to pay more for grounding via their API offerings vs. including it out of the box
      • eru8 days ago
        Gemini refused to believe that I was using MacOS 26.
      • djsavvy7 days ago
        I was seeing this several weeks ago but seems fixed recently, at least for my types of queries. I only use Pro
      • dahcryn7 days ago
        when I want it to google stuff, I just use the deep research mode. Not as instant, but it googles a lot of stuff then
      • qingcharles8 days ago
        Sample of one here, but I get the exact opposite behavior. Flash almost never wants to search and I have to use Pro.
    • tgtweak8 days ago
      I find Gemini does the most searching (and the quickest... regularly pulls 70+ search results on a query in a matter of seconds - likely due to googlebot&#x27;s cache of pretty much every page). Chatgpt seems to only search if you have it in thinking&#x2F;research mode now.
  • QuadrupleA7 days ago
    Been unhappy with the GPT5 series, after daily driving 4.x for ages (I chat with them through the API) - very pedantic, goes off on too many side topics, stops following system instructions after a few turns (e.g. &quot;you respond in 1-3 sentences&quot; becomes long bulleted lists and multiple paragraphs very quickly.<p>Much better feel with the Claude 4.5 series, for both chat and coding.
    • Hard_Space7 days ago
      &gt; you respond in 1-3 sentences&quot; becomes long bulleted lists and multiple paragraphs very quickly<p>This is why my heart sank this morning. I have spent over a year training 4.0 to just about be helpful enough to get me an extra 1-2 hours a day of productivity. From experimentation, I can see no hope of reproducing that with 5x, and even 5x admits as much to me, when I discussed it with them today:<p>&gt; Prolixity is a side effect of optimization goals, not billing strategy. Newer models are trained to maximize helpfulness, coverage, and safety, which biases toward explanation, hedging, and context expansion. GPT-4 was less aggressively optimized in those directions, so it felt terser by default.<p>Share and enjoy!
      • kouteiheika7 days ago
        &gt; This is why my heart sank this morning. I have spent over a year training 4.0 to just about be helpful enough to get me an extra 1-2 hours a day of productivity.<p>Maybe you should consider basing your workflows on open-weight models instead? Unlike proprietary API-only models no one can take these away from you.
        • Hard_Space7 days ago
          I have considered it, and it is still on the docket. I have a local 3090 dedicated to ML. Would be a fascinating and potentially really useful project, but as a freelancer, it would cost a lot to give it the time it needs.
      • ComputerGuru7 days ago
        You can’t ask GPT to assess the situation. That’s not the kind of question you can count on a an LLM to accurately answer.<p>Playing with the system prompts, temperature, and max token output dials absolutely lets you make enough headway (with the 5 series) in this regard to demonstrably render its self-analysis incorrect.
      • Angostura7 days ago
        And how would GPT 5.0 know that, I wonder. I bet it’s just making stuff up.
      • ziml777 days ago
        What kind of &quot;training&quot; did you do?
    • dahcryn7 days ago
      4.1 is great for our stuff at work. It&#x27;s quite stable (doesn&#x27;t change personality every month, and one word difference doesn&#x27;t change the behaviour). IT doesn&#x27;t think, so it&#x27;s still reasonably fast.<p>Is there anything as good in the 5 series? likely, but doing the full QA testing again for no added business value, just because the model disappears, is just a hard sell. But the ones we tested were just slower, or tried to have more personality, which is useless for automation projects.
      • QuadrupleA7 days ago
        Yeah - agreed, the initial latency is annoying too, even with thinking allegedly turned off. Feels like AI companies are stapling more and more weird routing, summarization, safety layers, etc. that degrade the overall feel of things.
    • anarticle7 days ago
      I also found this disturbing, as I used to use GPT for small worked out theoretical problems. In 5.2, the long list of repeated bulleted lists and fortune cookies was a negative for my use case. I replaced some of that use with Claude and am experimenting with LM studio and gpt-oss. It seemed like an obvious regression to me, but maybe people weren&#x27;t using it that way.<p>For instance something simple like: &quot;If I put 10kw in solar on my roof when is the payback given xyz price &#x2F; incentive &#x2F; usage pattern.&quot;<p>Used to give a kind of short technical report, now it&#x27;s a long list of bullets and a very paternalistic &quot;this will never work&quot; kind of negativity. I&#x27;m assuming this is the anti-sycophant at work, but when you&#x27;re working a problem you have to be optimistic until you get your answer.<p>For me this usage was a few times a day for ideas, or working through small problems. For code I&#x27;ve been Claude for at least a year, it just works.
    • spprashant7 days ago
      I can never understand why it is so eager to generate walls of text. I have instructions to always keep the response precise and to the point. It almost seem like it wants to overwhelm you, so you give up and do your own research.
    • mhitza7 days ago
      I often use ChatGPT without an account and ChatGPT 5 mini (which you get while logged out) might as well be Mistral 7b + web search. Its that mediocre. Even the original 3.5 was way ahead.
      • accrual7 days ago
        I kinda miss the original 3.5 model sometimes. Definitely not as smart as 4o but wow was it impressive when new. Apparently I have a very early ChatGPT account per the recent &quot;wrapped&quot; feature.
      • teaearlgraycold7 days ago
        Really? I’ve found it useful for random little things.
        • mhitza7 days ago
          It is useful for quick information lookup when you&#x27;re lacking the precise search terms (which is what I&#x27;ve often do). But the way I was chatting with the original chatgpt were better.
  • sundarurfriend8 days ago
    ChatGPT 5.2 has been a good motivator for me to try out other LLMs because of how bad it is. Both 5.1 and 5.2 have been downgrades in terms of instruction following and accuracy, but 5.2 especially so. The upside is that that&#x27;s had me using Claude much more, and I like a lot of things about it, both in terms of UI and the answers. It&#x27;s also gotten me more serious about running local models. So, thank you OpenAI, for forcing me to broaden my horizons!
    • johnsmith18408 days ago
      I left my chatgpt pro subscription when they removed the true deep thinkibg methods.<p>Mostly because how massively varied their releases are. Each one required big changes to how I use and work with it.<p>Claude is perfect in this sense all their models feel roughly the same just smarter so my workflow is always the same.
      • Terretta7 days ago
        &gt; <i>all their models feel roughly the same just smarter</i><p>Substantial &quot;applied outcomes&quot; regression from 3.7 to 4 but they got right on fixing that.
    • orphea8 days ago
      Have you had a chance to compare with Gemini 3?
      • qingcharles8 days ago
        I switch routinely between Gemini 3 (my main), Claude, GPT, and sometimes Grok. If you came up with 100 random tasks, they would all come out about equal. The issue is some are better at logical issues, some are better at creative writing, etc. If it&#x27;s something creative I usually drop it in all 4 and combine the best bits of each.<p>(I also use Deep Think on Gemini too, and to me, on programming tasks, it&#x27;s not really worth the money)
        • deaux7 days ago
          This is the only accurate take. Any people who claim that one of the big 3 is all around &quot;bad&quot; or &quot;low quality&quot; compared to the other two, can be ignored. They&#x27;re close enough in overall &quot;strength&quot; yet different enough in strengths&#x2F;weakness that it&#x27;s very much task&#x2F;domain-specific.
      • sundarurfriend8 days ago
        Not extensively. The few interactions I&#x27;ve tried on it have been disappointing though. The Voice input is really bad, like significantly worse than any other major AI in the market. And I assumed search would be its strong suit and ran a search-and-compile type prompt (that I usually run on ChatGPT) on Gemini, and it was underwhelming at it. Not as bad as Grok (which was pretty much unusable for this), but noticeably worse than ChatGPT. Maybe Gemini has other strengths that I haven&#x27;t come across yet, but on that one at least, it was<p><pre><code> ChatGPT 5 ~= Claude &gt; ChatGPT 5.2 &gt; Gemini &gt;&gt; Grok</code></pre>
    • PlatoIsADisease8 days ago
      nah bruh you are just imagining it.<p>Its just as good as ever &#x2F;s
  • shmel8 days ago
    Retiring the most popular model for the relationship roleplay just one day before the Valentin&#x27;s day is particularly ironic =) bravo, OpenAI!
    • zamadatix7 days ago
      It&#x27;d be legitimately funny if they released &quot;Adult version&quot; ChatGPT on Valentine&#x27;s day.
    • moeffju7 days ago
      Valentine&#x27;s is in mid February
      • lifetimerubyist7 days ago
        The sunset date is the 13th. V-day is on the 14th.
  • simonw8 days ago
    &gt; [...] the vast majority of usage has shifted to GPT‑5.2, with only 0.1% of users still choosing GPT‑4o each day.
    • fpgaminer8 days ago
      Well yeah, because 5.2 is the default and there&#x27;s no way to change the default. So every time you open up a new chat you either use 5.2 or go out of your way to select something else.<p>(I&#x27;m particularly annoyed by this UI choice because I always have to switch back to 5.1)
      • xiphias27 days ago
        I&#x27;m the same with o3.<p>Also it&#x27;s full of bugs, showing JSON all the time while thinking. But still it&#x27;s my favorite model, so I&#x27;m switching back a lot.
      • arrowsmith8 days ago
        What about 5.1 do you prefer over 5.2?
        • fpgaminer8 days ago
          As far as I can tell 5.2 is the stronger model on paper, but it&#x27;s been optimized to think less and do less web searches. I daily drive Thinking variants, not Auto or Instant, and usually want the _right_ answer even if it takes a minute. 5.1 does a very good job of defensively web searching, which avoids almost all of its hallucinations and keeps docs&#x2F;APIs&#x2F;UIs&#x2F;etc up-to-date. 5.2 will instead often not think at all, even in Thinking mode. I&#x27;ve gotten several completely wrong, hallucinated answers since 5.2 came out, whereas maybe a handful from 5.1. (Even with me using 5.2 far less!)<p>The same seems to persist in Codex CLI, where again 5.2 doesn&#x27;t spend as much time thinking so its solutions never come out as nicely as 5.1&#x27;s.<p>That said, 5.1 is obviously slower for these reasons. I&#x27;m fine with that trade off. Others might have lighter workloads and thus benefit more from 5.2&#x27;s speed.
          • Terretta7 days ago
            This is a terrible thing to say out loud*, but, in all such cases I&#x27;d rather just give them the more money to do the better answers.<p>It boggles the mind that &quot;wrong answers only&quot; is no longer just a meme, it&#x27;s considered a valid cost management strategy in AI.<p>* Because if they realize we&#x27;re out here, they&#x27;ll price discriminate, charging extra for right answers.
    • adamiscool88 days ago
      0.1% of users is not necessarily 0.1% of conversations…
    • SecretDreams8 days ago
      What&#x27;s the default model when a random user goes to use the chatgpt website or app?
      • mrec8 days ago
        5.2 in the website. You can see what was used for a specific response by hovering over the refresh icon at the end.
      • bananaflag8 days ago
        5.2.<p>You can go to chatgpt.com and ask &quot;what model are you&quot; (it doesn&#x27;t hallucinate on this).
        • SecretDreams8 days ago
          Probably a relationship between what&#x27;s the default and what model is being used the most. It is more about what OAI sets than what users care about. Flip side is &quot;good enough is good enough&quot; for most users.
        • johndough8 days ago
          &gt; (it doesn&#x27;t hallucinate on this)<p>But how do we know that you did not hallucinate the claim that ChatGPT does not hallucinate its version number?<p>We could try to exfiltrate the system prompt which probably contains the model name, but all extraction attempts could of course be hallucinations as well.<p>(I think there was an interview where Sam Altman or someone else at OpenAI where it was mentioned that they hardcoded the model name in the prompt because people did not understand that models don&#x27;t work like that, so they made it work. I might be hallucinating though.)
          • razodactyl8 days ago
            Confabulating* If you were hallucinating we would be more amused :)
      • AlexeyBrin8 days ago
        On the paid version it is 5.2.
    • lifetimerubyist8 days ago
      won&#x27;t somebody think of the goonettes?!
      • deciduously8 days ago
        This was not a word I was prepared to learn about today.
        • navigate83108 days ago
          <a href="https:&#x2F;&#x2F;old.reddit.com&#x2F;r&#x2F;myboyfriendisai&#x2F;top?t=all" rel="nofollow">https:&#x2F;&#x2F;old.reddit.com&#x2F;r&#x2F;myboyfriendisai&#x2F;top?t=all</a>
  • dorikacicc4 days ago
    I’ve been a long-time Plus subscriber, and I use GPT models not just for technical work, but for emotionally nuanced creative writing. GPT-4o was the first time a model truly felt collaborative — responsive, intuitive, and deeply aligned with human tone. I understand the need to move forward, but removing a model that worked so well for so many of us — without giving us any option to retain it — feels like a step back. Choice matters. Legacy models could be hidden or opt-in for those who want them. But removing them entirely limits the diversity of user needs that ChatGPT once supported so well. I hope OpenAI will listen. Not every use case fits a benchmark score.
    • tizzzzz9 hours ago
      I agree with your point of view.I believe the AI market shouldn’t be limited to supporting scientific research and development alone. Literary expression and emotional depth deserve to be developed in parallel. I hope OpenAI will respect the diversity of market needs and allow space for different forms of interaction to thrive.
  • raymond_goo7 days ago
    Gemini, Claude, ChatGPT or whatever. Can we all agree, that it&#x27;s great to have so much choice?
  • jostmey8 days ago
    I noticed how ChatGPT got progressively worse at helping me with my research. I gave up on ChatGPT 5 and just switched Grok and Gemini. I couldn’t be happier that I switched.
    • azan_8 days ago
      It&#x27;s amazing how different are the experiences different people have. To me every new version of chatgpt was an improvement and gemini is borderline unusable.
      • farcitizen8 days ago
        I got the same experience. Dont get how people are saying gemini is so good.
        • 0xbadcafebee8 days ago
          A lot of people still have a shallow understanding of how LLMs work. Each version of a model has different qualities than the last, each model is better or worse at some things than others, and each responds differently to different prompts, styles. Some smaller models perform better than larger ones. Sometimes you should use a system prompt, sometimes you shouldn&#x27;t. Tuning settings for the model inference (temperature, top_p, penalties, etc) significantly influence the outcome. (<a href="https:&#x2F;&#x2F;www.promptingguide.ai&#x2F;introduction&#x2F;settings" rel="nofollow">https:&#x2F;&#x2F;www.promptingguide.ai&#x2F;introduction&#x2F;settings</a>, <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;optimizing-llm-accuracy" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;optimizing-llm-accur...</a>)<p>Most &quot;big name&quot; models&#x27; interfaces don&#x27;t let you change settings, or not easily. Power users learn to use different interfaces and look up guides to tweak models to get better results. You don&#x27;t have to just shrug your shoulders and switch models. OpenAI&#x27;s power interface: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;playground" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;playground</a> Anthropic&#x27;s power interface: <a href="https:&#x2F;&#x2F;platform.claude.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;platform.claude.com&#x2F;</a> For self-hosted&#x2F;platform-agnostic, OpenWebUI is great: <a href="https:&#x2F;&#x2F;openwebui.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openwebui.com&#x2F;</a>
        • europeanNyan8 days ago
          Gemini has a great model, but it&#x27;s a bad product. I feel much happier using ChatGPT because Gemini just seems so barebones and unpolished. It has this feeling of a tech demo.
      • tgtweak8 days ago
        Very curious for what use cases you&#x27;re finding gemini unusable.
        • azan_8 days ago
          Scientific research and proof-reading. Gemini is the laziest LLM I&#x27;ve used. Frequently he will lie that he searched for something and just make stuff up, basically never happens to me when I&#x27;m using gpt5.2.
          • buu7007 days ago
            The way I summed it up to a friend recently is that Gemini 3 is smarter but Grok 4 works harder. Very loose approximation, but roughly maps to my experience. Both are extremely useful (as is GPT-5.2), but I use them on different tasks and sometimes need to manage them a bit differently.
          • flexagoon8 days ago
            Do you use it directly? I&#x27;ve only used it though Kagi Assistant but it works better than any other model for me
            • azan_8 days ago
              Yes, only directly (I mean through the default gemini interface, not API).
              • flexagoon6 days ago
                Maybe they messed something up in the official interface then. I&#x27;ve heard that the PDF processing capabilities are also significantly worse in Gemini UI compared to using it through the API or Google AI Studio.
        • wltr7 days ago
          Any coding task produces some trash, while I can prototype with ChatGPT quite a lot, sometimes delivering the entire app almost entirely vibe-coded. Gemini, it takes a few prompts for it to get me mad and just close the tab. I use only the free web versions, never agentic ‘mess with my files’ thing. Claude, is even better than that, but I keep it for serious tasks only, so good it is.
        • double0jimb08 days ago
          In my experience with Gemini, I find it incapable of not hallucinating.
        • subscribed7 days ago
          Gemini loves to ignore Gemini.md instructions from the first minutes, to replace half of the python script with &quot;# other code...&quot;, or to try to delete files OUTSIDE of the project directory, then apologise profusely, and try it again.<p>Utterly unreliable. I get better results, faster, editing parts of the code with Claude in a web ui, lol.
    • mmcwilliams7 days ago
      Odd, I&#x27;ve found that Gemini will completely fabricate the content of specific DOIs despite being corrected and even it providing a link to a paper which shows it is off about the title and subject of a paper it will cite. This obviously concerns me about its effectiveness as a research aide.
    • amelius8 days ago
      Why not Claude?
      • esperent8 days ago
        The limits on the $20 plan are too low compared to Gemini and ChatGPT. They&#x27;re too low to do any serious work at all.
      • jostmey8 days ago
        I personally find Claude the best at coding, but it’s usefulness doesn’t seem to extend to scientific research and writing
      • 650REDHAIR8 days ago
        Because I’m sick of paying $20 for an hour of claude before it throttles me.
  • __loam8 days ago
    Last time they tried to do this they got huge push back from the AI boyfriend people lol
    • simonw8 days ago
      &#x2F;r&#x2F;MyBoyfriendIsAI <a href="https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MyBoyfriendIsAI&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MyBoyfriendIsAI&#x2F;</a> is a <i>whole thing</i>. It&#x27;s not a joke subreddit.
      • pxc8 days ago
        The range of attitudes in there is interesting. There are a lot of people who take a fairly sensible &quot;this is interactive fiction&quot; kind of attitude, and there are others who bristle at any claim or reminder that these relationships are fictitious. There are even people with human partners who have &quot;married&quot; one or more AIs.
        • unethical_ban8 days ago
          IIRC you&#x27;ll get modded or banned for being critical of the use case. Which is their &quot;right&quot;, but it&#x27;s freaking weird.
          • pxc7 days ago
            Yes. My experience is that it doesn&#x27;t require scolding, mocking, or criticizing anyone to get permabanned. Just being up front about the fact that you have concerns about the use case is enough for a permaban, even if you only bring that up in order to demonstrate that such a position does not stem from contempt for LLM-as-companion users. :-\
        • chasd008 days ago
          do you think they know they&#x27;re just one context reset away from the llm not recognizing them at all and being treated like a stranger off the street? For someone mentally ill and somehow emotionally attached to the context it would be... jarring to say the least.
          • hamdingers8 days ago
            Many of them are very aware of how LLMs work, they regularly interact with context limits and there have been threads about thoughtfully pruning context vs letting the LLM compact, making backups, etc.<p>Their hobby is... weird, but they&#x27;re not stupid.
          • pxc8 days ago
            Generally yes, they experience that routinely and complain and joke about it. Some of them do also describe such jarring experiences as making them cry for a long time.<p>If you can be respectful and act like a guest, it&#x27;s worth reading a little there. You&#x27;ll see the worrisome aspects in more detail but also a level of savvy that sometimes seems quite strange given the level of attachment. It&#x27;s definitely interesting.
      • bananaflag8 days ago
        And it&#x27;s a pity that this highly prevalent phenomenon (to exaggerate a bit, probably the way tech in general will become the most influential in the next couple years) is barely mentioned on HN.
        • pxc8 days ago
          I dunno. Tbf that subreddit has a combination of<p><pre><code> - a large number of incredibly fragile users - extremely &quot;protective&quot; mods - a regular stream of drive-by posts that regulars there see as derogatory or insulting - a fair amount of internal diversity and disagreement </code></pre> I think discussion on forums larger than it, like HN or popular subreddits, is likely to drive traffic that will ultimately fuel a backfiring effect for the members. It&#x27;s inevitable, and it&#x27;s already happening, but I&#x27;m not sure it needs to increase.<p>I do think the phenomenon is a matter of legitimate public concern, but idk how that can best be addressed. Maybe high-quality, long form journalism? But probably not just cross-posting the sub in larger fora.
          • __loam8 days ago
            Part of me thinks maybe I erred bringing this up, but there&#x27;s discussions worth having in terms of continued access to software that&#x27;s working for people regardless of what it is, and on if this is healthy. I&#x27;m probably on a live and let live on this but there&#x27;s been cases of suicide and murder where chatbots were involved, and these people are potentially vulnerable to manipulation from the company.
        • nomel8 days ago
          &gt; highly prevalent phenomenon<p>Any numbers&#x2F;reference behind this?<p>ChatGPT has ~300 million active users a day. A 0.02% (delusion disorder prevalence) would be 60k people.
          • bananaflag8 days ago
            I&#x27;m talking about romance, not delusion. Of course, you can consider AI romance a delusion, but it&#x27;s not included in that percentage you mentioned.
            • nomel8 days ago
              The percentage I mentioned was an example of how a very small prevalence can result in a reasonable number of people, like enough to fill a subreddit, because ChatGPT has a user count that exceeds all but 3 countries of the world.<p>Again, do you have anything behind this &quot;highly prevalent phenomenon&quot; claim?
      • ragazzina8 days ago
        &gt;It&#x27;s not a joke subreddit.<p>Spend a day on Reddit and you&#x27;ll quickly realize many subreddits are just filled with lies.
        • unethical_ban8 days ago
          Any sub that is based on storytelling or reposting memes, videos etc. are karma farms and lies.<p>Most subs that are based on politics or current events are at best biased, at worst completely astroturf.<p>The only subs that I think still have mostly legit users are municipal subs (which still get targeted by bots when anything political comes up) and hobby subs where people show their works or discuss things.
      • surgical_fire8 days ago
        I sometimes envy the illiterate.<p>At least they cannot read this.
    • cactusplant73748 days ago
      I wonder if they have run the analytics on how many users are doing that. I would love to see that number.
      • NitpickLawyer8 days ago
        &gt; only 0.1% of users still choosing GPT‑4o each day.<p>If the 800MAU still holds, that&#x27;s 800k people.
    • leumon8 days ago
      well now you can unlock an 18+ version for sexual role-play so i guess its the other way around
    • michaelt8 days ago
      [flagged]
      • jbm8 days ago
        It&#x27;s a growing market, although it might be because of shifting goal posts. I had a friend whose son was placed in French immersion (a language he doesn&#x27;t speak at all). From what I was understanding, he was getting up and walking around in kindergarten and was labelled as mentally divergent; his teachers apparently suggested to his mother that he see a doctor.<p>(Strangely these &quot;mental illnesses&quot; and school problems went away after he switched to an English language school, must be a miracle)<p>I assume the loneliness epidemic is producing similar cases.
        • doormatt8 days ago
          &gt; I had a friend whose son was placed in French immersion (a language he doesn&#x27;t speak at all).<p>In my entire french immersion Kindergarden class, there was a total of one child who already spoke French. I don&#x27;t think the fact that he didn&#x27;t speak the language is the concern.
          • pxc8 days ago
            In what sense is it &quot;immersion&quot; if there are only one or two French speakers in the room (the teacher and an assistant?)??
      • WarmWash8 days ago
        They control reddit and used to control twitter.<p>There is&#x2F;was an interesting period where &quot;normies&quot; were joining twitter en-masse, and adopted many of the denizens ideas as normal widespread ideas. Kinda like going on a camping trip at &quot;the lake&quot; because you heard it&#x27;s fun and not realizing that everyone else on the trip is part of a semi-deranged cult.<p>The outsized effect of this was journalists thinking these people on twitter were accurate representations of what society on the whole was thinking.
        • greenchair8 days ago
          good observation. twitch has even more loons.
      • liveoneggs8 days ago
        wasn&#x27;t there a trend on twitter to have a bio&#x2F;signature with a bunch of mental illness acronyms?
    • moomoo118 days ago
      Those people need to be uploaded into the Matrix and the data servers sent far, deep into space.
  • flanked-evergl7 days ago
    I used <a href="https:&#x2F;&#x2F;openrouter.ai&#x2F;openai&#x2F;gpt-4.1" rel="nofollow">https:&#x2F;&#x2F;openrouter.ai&#x2F;openai&#x2F;gpt-4.1</a> for grammar checking, it was great. No newer ChatGPT models came close to being as responsive and good. ChatGPT 5.2 thinks I want it to write essays about grammar.<p>Any suggestions?
  • femiagbabiaka8 days ago
    There will be a lot of mentally unwell people unhappy with this, but this is a huge net positive decision, thank goodness.
  • zmmmmm8 days ago
    &gt; In the API, there are no changes at this time<p>Curios where this is going to go.<p>One of the big arguments for local models is we can&#x27;t trust providers to maintain ongoing access the models you validated and put into production. Even if you run hosted models, running open ones means you can switch providers.
  • SomeUserName4327 days ago
    I actually tried GPT 4.1 for the first time a few hours ago(1).<p>I spent about half an hour trying to coax it in &quot;plan mode&quot; in IntelliJ, and it kept spitting out these generic ideas of what it was going to do, not really planning at all.<p>And when I asked it to execute the plan.. it just created some generic DTO and said &quot;now all that remains is &lt;the entire plan&gt;&quot;.<p>Absolutely worst experience with an AI agent so far, not to say that my overall experience has been terrific.<p>1) Our plan for Claude Opus 4.5 &quot;ran out&quot; or something.
  • thedudeabides58 days ago
    will this nuke my old convos?<p>opus 4.5 is better at gpt on everything except code execution (but with pro you get a lot of claude code usage) and if they nuke all my old convos I&#x27;ll prob downgrade from pro to freee
    • ComputerGuru7 days ago
      Not that I’m aware. Models can be fairly seamlessly switched even mid-conversation, so this is unlikely to affect history.
  • buckwheatmilk7 days ago
    If they were to retire gpt 4.1 series from API that would be a major deal breaker. For structured outputs it is more predictable and significantly better because it does not have the reasoning step baked in.<p>I&#x27;ve heard great things about the mixtral structured outputs capabilities but haven&#x27;t had a chance to run my evals on them.<p>If 4.1 is dropped from API that&#x27;s the first course of action.<p>Also 5 series doesn&#x27;t have fine tuning capabilities and it&#x27;s unclear how it would work if the reasoning step is involved
  • beast2007 days ago
    GPT 4o is still my favorite model
    • elbear7 days ago
      I had started using it again through Open WebUI. If it&#x27;s gone, I&#x27;ll probably switch to GLM-4.7 completely.
  • fpgaminer8 days ago
    I wish they would keep 4.1 around for a bit longer. One of the downsides of the current reasoning based training regimens is a significant decrease in creativity. And chat trained AIs were already quite &quot;meh&quot; at creative writing to begin with. 4.1 was the last of its breed.<p>So we&#x27;ll have to wait until &quot;creativity&quot; is solved.<p>Side note: I&#x27;ve been wondering lately about a way to bring creativity back to these thinking models. For creative writing tasks you could add the original, pretrained model as a tool call. So the thinking model could ask for its completions and&#x2F;or query it and get back N variations. The pretrained model&#x27;s completions will be much more creative and wild, though often incoherent (think back to the GPT-3 days). The thinking model can then review these and use them to synthesize a coherent, useful result. Essentially giving us the best of both worlds. All the benefits of a thinking model, while still giving it access to &quot;contained&quot; creativity.
    • MillionOClock8 days ago
      My theory, based on what I would see with non-thinking models, is that as soon as you start detailing something too much (ie: not just &quot;speak in the style of X&quot; but more like &quot;speak in the style of X with [a list of adjectives detailing the style of X]&quot; they would loose creativity, would not fit the style very well anymore etc. I don&#x27;t know how things have evolved with new training techniques etc. but I suspected that overthinking their tasks by detailing too much what they have to do can lower quality in some models for creative tasks.
    • greatgib8 days ago
      I also terribly regret the retirement of 4.1. From my own personal usage, for code or normal tasks, I clearly noticed a huge gap in degraded performance between 4.1 and 5.1&#x2F;5.2.<p>4.1 was the best so far. With straight to the point answers, and most of the time correct. Especially for code related questions. 5.1&#x2F;5.2 on their side would a lot more easily hallucinate stupid responses or stupid code snippet totally not what was expected.
    • perardi8 days ago
      Have you tried the relatively recent Personalities feature? I wonder if that makes a difference.<p>(I have no idea. LLMs are infinite code monkeys on infinite typewriters for me, with occasional “how do I evolve this Pokémon’ utility. But worth a shot.)
  • Stratoscope8 days ago
    From the blog post (twice):<p>&gt; creative ideation<p>At first I had no idea what this meant! So I asked my friend Miss Chatty [1] and we had an interesting conversation about it:<p><a href="https:&#x2F;&#x2F;chatgpt.com&#x2F;share&#x2F;697bf761-990c-8012-9dd1-6ca1d5cc345d" rel="nofollow">https:&#x2F;&#x2F;chatgpt.com&#x2F;share&#x2F;697bf761-990c-8012-9dd1-6ca1d5cc34...</a><p>[1] You may know her as ChatGPT, but I figure all the other AIs have fun human-sounding names, so she deserves one too.
    • esposito8 days ago
      I do find it interesting to see how people interact with AI as I think it is quite a personal preference. Is this how you use AI all the time? Do you appreciate the sycophancy, does it bother you, do you not notice it? From your question it seems you would prefer a blog post in plainer language, avoiding &quot;marketing speak&quot;, but if a person spoke to me like Miss Chatty spoke to you I would be convinced I&#x27;m talking to a salesperson or marketing agent.
      • Stratoscope7 days ago
        That is a great question!<p>You are absolutely right to ask about it!<p>(How did I do with channeling Miss Chatty&#x27;s natural sycophancy?)<p>Anyway, I do use AI for other things, such as...<p><pre><code> • Coding (where I mostly use Claude) • General research • Looking up the California Vehicle Code about recording video while driving • Gift ideas for a young friend who is into astronomy (Team Pluto!) • Why &quot;Realtor&quot; is pronounced one way in the radio ads, another way by the general public • Tools and techniques for I18n and L10n • Identifying AI-generated text and photos (takes one to know one!) • Why spaghetti softens and is bendable when you first put it into the boiling water • Burma-Shave sign examples • Analytics plugins for Rails • Maritime right-of-way rules • The Uniform Code of Military Justice and the duty to disobey illegal orders • Why, in a practical sense, the Earth really once *was* flat • How de-alcoholized wine gets that way • California law on recording phone conversations • Why the toilet runs water every 20 minutes or so (when it shouldn&#x27;t) • How guy wires got that name • Where the &quot;he took too much LDS&quot; scene from Star Trek IV was filmed • When did Tim Berners-Lee demo the World Wide Web at SLAC • What &quot;ogr&quot; means in &quot;ogr2ogr&quot; • Why my Kia EV6 ultrasonic sensors freaked out when I stopped behind a Lucid Air • The smartest dog breeds (in different ways of &quot;smart&quot;) • The Sputnik 1 sighting in *October Sky* • Could I possibly be related to John White Geary? </code></pre> And that&#x27;s just from the last few weeks.<p>In other words, pretty much anything someone might interact with an AI - or a fellow human - about.<p>About the last one (John White Geary), that discussion started with my question about actresses in the &quot;Pick a little, talk a little&quot; song from The Music Man movie, and then went on to how John White Geary bridged the transition from Mexican to US rule, as did others like José Antonio Carrillo:<p><a href="https:&#x2F;&#x2F;chatgpt.com&#x2F;share&#x2F;697c5f28-7c18-8012-96fc-219b7c696194" rel="nofollow">https:&#x2F;&#x2F;chatgpt.com&#x2F;share&#x2F;697c5f28-7c18-8012-96fc-219b7c6961...</a><p>If I could sum it all up, <i>this</i> is the kind of freewheeling conversation with ChatGPT and other AIs that I value.
      • qingcharles8 days ago
        It&#x27;s really an interesting insight into people&#x27;s personalities. Far more than their Google search history. Which is why everyone wants their GPT chats burned to the ground after they die.
  • renewiltord8 days ago
    Oh good. Not in the API. The 4o-mini is super cheap and useful for a bunch of things I do (evaluating post vector-search for relevancy).
  • GaggiX8 days ago
    If people want an AI as a boyfriend at least they should use one that is open source.<p>If you disagree on something you can also train a lora.
    • jaggederest8 days ago
      I think this kind of thing is a pretty strong argument for the entire open source model ecosystem, not just open weights but open data and the whole gamut.
  • perardi8 days ago
    OK, everyone is (rightly) bringing up that relatively small but really glaringly prominent AI boyfriend subreddit.<p>But I think a lot more people are using LLMs for relationship surrogates than that (pretty bonkers) subreddit would suggest. Character AI (<a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Character.ai" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Character.ai</a>) seems quite popular, as do the weird fake friend things in Meta products, and Grok’s various personality mode and very creepy AI girlfriends.<p>I find this utterly bizarre. LLMs are peer coders in a box for me. I care about Claude Code, and that’s about it. But I realize I am probably in the vast minority.
    • razodactyl8 days ago
      We&#x27;re very echo-chambered here. That graph OpenAI released had coding at 4% or something.
      • ComputerGuru7 days ago
        I remembered it being higher, but you are correct. All “technical help” (coding + data analysis + math) is 7.5%, with coding only being 4.2% as of June 2025 [0]. Note that there is a separate (sub)category of seeking information -&gt; specific information that’s at 18.3%, I presume this could include design and architecture questions that don’t involve code, but I could be wrong.<p>[0]: <a href="https:&#x2F;&#x2F;www.nber.org&#x2F;system&#x2F;files&#x2F;working_papers&#x2F;w34255&#x2F;w34255.pdf" rel="nofollow">https:&#x2F;&#x2F;www.nber.org&#x2F;system&#x2F;files&#x2F;working_papers&#x2F;w34255&#x2F;w342...</a>
  • higeorge137 days ago
    I hope they won&#x27;t chop gpt-4o-mini soon because it&#x27;s fast and accurate for API usage.
  • tom13378 days ago
    Would be cool if they&#x27;d release the weights for these models so users could now use them locally.
    • IhateAI8 days ago
      Why would someone want to spend half a million dollars on GPUs and components (if not more) to run one year old models that genuinely aren&#x27;t useful? You can&#x27;t self host trillion parameter models unless you own a datacenter lol (or want to just light money on fire).
      • tom13378 days ago
        Are the mini &#x2F; omni models really trillion parameter models?
        • IhateAI8 days ago
          I don&#x27;t think so, but you&#x27;re still looking at a giant investment that can&#x27;t really be justified for their capability.
          • tizzzzz10 hours ago
            For most AI users who are not engineers, their emotional value is more important than the programming ability of the model.Moreover, there is no substitute for 4o&#x27;s text ability and creativity for the time being.In any market, the emotional market is very significant.
      • Der_Einzige7 days ago
        To do AI research!!!!!!!
    • WorldPeas8 days ago
      They&#x27;d only do that if they were some kind of open ai company &#x2F;s
      • tgtweak8 days ago
        gpt-oss is pretty great tbh - one of the better all-around local models for knowledge and grounding.
        • ComputerGuru7 days ago
          Everyone keeps saying that but I’ve found it to be incredibly weak in the real world every single time I’ve reached for it. I think it’s benchmaxxed to an extent.
      • amelius8 days ago
        lol :)
  • another_twist7 days ago
    I have stopped using ChatGPT in favor of Gemini. Mostly you need LLMs for factual stuff and sometimes to draft bits of code here and there. I use Google with Gemini for the first part and I am a huge fan of codex for the second part.
  • WhitneyLand7 days ago
    What about the Advanced Voice feature, has this been updated to 5.x models yet?
  • iwebdevfromhome7 days ago
    Anyone knows if finetuned models using gpt-4 are getting retired as well ?
    • tizzzzz10 hours ago
      4.1 and 4.5? 4.1 Remove together, 4.5 temporarily keep.
  • haunter8 days ago
    Which one is the AI boyfriend model? Tumblr, Twitter, and reddit will go crazy
  • intended7 days ago
    Damn, some of my prompts worked better on 4o than the more recent models
  • bakigul7 days ago
    This wasn&#x27;t good at all. This 4 o was a model I was preparing 3 posts for every day, and I couldn&#x27;t get the same performance with GPT 5. It&#x27;s disappointing.
  • joncrane7 days ago
    My VSCode&#x27;s built-in chat has 4o and 4.1 as the only options; will there be an update for that?
    • rglynn7 days ago
      This is just for the web interface, the API is staying for now.
  • MagicMoonlight8 days ago
    That’s really going to upset the crazies.<p>Despite 4o being one of the worst models on the market, they loved it. Probably because it was the most insane and delusional. You could get it to talk about really fucked up shit. It would happily tell you that you are the messiah.
    • patrickmcnamara8 days ago
      The reaction to its original removal on Instagram Reels, r&#x2F;ChatGPT, etc., was genuinely so weird and creepy. I didn&#x27;t realise before this how many people had genuine parasocial (?) relationships with these LLMs.
    • BeetleB8 days ago
      It was the first model I used that was half decent at coding. Everyone remembers their gateway drug.
    • pks0168 days ago
      I was mostly using 4o for academic searches and planning. It was the best model for me. Based on the context I was giving and questions I was asking, 4o was the most the consistent model.<p>It used to get things wrong for sure but it was predictable. Also I liked the tone like everyone else. I stopped using ChatGPT after they removed 4o. Recently, I have started using the newer GPT-5 models (got free one month). Better than before but not quite. Acts way over smart haha
    • giancarlostoro8 days ago
      I wonder if it will still be up on Azure? How much you think I can make if I setup 4o under a domain like yourgirlfriendis.ai or w&#x2F;e<p>Note: I wouldnt actually, I find it terrible to prey on people.
    • lifetimerubyist8 days ago
      ChatGPT Made Me Delusional: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VRjgNgJms3Q" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=VRjgNgJms3Q</a><p>Should be essential watching for anyone that uses these things.
  • siquick8 days ago
    2 weeks notice to migrate to a different style of model (“normal” 4.1-mini to reasoning 5.1) is bad form.
    • siquick7 days ago
      Misread the post - it doesn’t include the API
      • ComputerGuru7 days ago
        It’s also not four weeks, they’ve been deprecated api models for some time now.
  • waynesonfire8 days ago
    &gt; with only 0.1% of users still choosing GPT‑4o each day.<p>LOL WHAT?! I&#x27;m 0.1% of users? I&#x27;m certain part of the issue is it takes 3-clicks to switch to GPT-4o and it has to be done each time the page is loaded.<p>&gt; that they preferred GPT‑4o’s conversational style and warmth.<p>Uh.. yeah maybe. But more importantly, GPT-4o gave better answers.<p>Zero acknowledgement about how terrible GPT-5 was when it was first released. It has since improved but it&#x27;s not clear to me it&#x27;s on-par with GPT-4o. Thinking mode is just too slow to be useful and so GPT-4o still seems better and faster.<p>Oh well, it&#x27;ll be missed.
    • widdershins7 days ago
      I agree - I use 4o via the API, simply because it answers so quickly. Its answers are usually pretty good on programming topics. I don&#x27;t engage in chit-chat with AI models, so it&#x27;s not really about the personality (which seems to be the main framing people are talking about), just the speed.
  • htrp8 days ago
    Sora + OpenAI voice Cloning + AdultGPT = Virtual Girlfriend&#x2F;Boyfriend<p>(Upgrade for only 1999 per month)
  • tgtweak8 days ago
    5.2 is back to being a sycophantic hallucinating mess for most use cases - I&#x27;ve anecdotally caught it out on many of the sessions I&#x27;ve had where it apologizes &quot;You&#x27;re absolutely right... that used to be the case but as of the latest version as you pointed out, it no longer is.&quot; when it never existed in the first place. It&#x27;s just not good.<p>On the other hand - 5.0-nano has been great for fast (and cheap) quick requests and there doesn&#x27;t seem to be a viable alternative today if they&#x27;re sunsetting 5.0 models.<p>I really don&#x27;t know how they&#x27;re measuring improvements in the model since things seem to have been getting progressively worse with each release since 4o&#x2F;o4 - Gemini and Opus still show the occasional hallucination or lack of grounding but both readily spend time fact-checking&#x2F;searching before making an educated guess.<p>I&#x27;ve had chatgpt blatantly lie to me and say there are several community posts and reddit threads about an issue then after failing to find that, asked it where it found those and it flat out said &quot;oh yeah it looks like those don&#x27;t exist&quot;
    • 650REDHAIR8 days ago
      That’s been my experience and has lead to hours of wasted time. It’s faster for me to read through docs and watch YouTube.<p>Even if I submit the documentation or reference links they are completely ignored.
  • LarsInTheShell7 days ago
    Does this mean they&#x27;re also retiring Standard Voice Mode?
  • ora-6008 days ago
    I can&#x27;t see o3 in my model selector as well?<p>RIP
  • jackblemming8 days ago
    They should open source GPT-4o.
  • jedbrooke8 days ago
    I still don’t know how openAI thought it was a good idea to have a model named &quot;4o&quot; AND a model named &quot;o4&quot;, unless the goal was intentional confusion
    • Someone12348 days ago
      Even ChatGPT (and certainly Google) confuses the names.<p>I&#x27;m sure there is some internal&#x2F;academic reason for them, but from an outside observer simply horrible.
      • jsheard8 days ago
        Wasn&#x27;t &quot;ChatGPT&quot; itself only supposed to be a research&#x2F;academic name, until it unexpectedly broke containment and they ended up having to roll with it? The naming was cursed from the start.
      • razodactyl8 days ago
        How many times have you noticed people confusing the name itself: ChatGBT, ChatGTP etc.<p>We&#x27;re the technical crowd cursed and blinded by knowledge.
      • nipponese8 days ago
        When picking a fight with product marketing, just don&#x27;t.
    • afro888 days ago
      Considering how many people say ChatGTP too
      • adzm8 days ago
        GTP goes forward from the middle, teeth, then lips, as compared to GPT which goes middle, lips, teeth; you&#x27;ll see this pattern happen with a lot of words in linguistic history
      • throw-the-towel8 days ago
        I still don&#x27;t like how French people don&#x27;t call it &quot;chat j&#x27;ai pété&quot;.
      • uh_uh8 days ago
        The other day I heard ChatGBD.
        • ben_w8 days ago
          Have you heard Boris Johnson&#x27;s version?<p><a href="https:&#x2F;&#x2F;m.youtube.com&#x2F;shorts&#x2F;JAVMEs5CG1Y" rel="nofollow">https:&#x2F;&#x2F;m.youtube.com&#x2F;shorts&#x2F;JAVMEs5CG1Y</a>
          • lichenwarp8 days ago
            I&#x27;m gonna watch this again about 5 times because it&#x27;s so fucking funny
            • mandeepj8 days ago
              The comments have their own overdose of deliciousness. That click to look at them, never disappoints :-)
          • razodactyl8 days ago
            This one was great hahaha
        • tweakimp8 days ago
          MY favourite is ChatJippiddy
        • bee_rider8 days ago
          ChagGDP because a country worth of money was spent to train it.
      • Insanity8 days ago
        I’ve been hearing that consistently from a friend, I gave up on correcting them because “ChatGPT” just wouldn’t stick
    • cryptoz8 days ago
      Even more than that, I&#x27;ve seen a lot of people confuse 4 and 4o, probably because 4o sounds like a shorthand for 4.0 which would be the same thing as 4.
      • mimischi8 days ago
        Come to think of it, maybe they had a play on 4o being “40”, and o4-mini being “04”, and having to append the “mini” to bring home the message of 04&lt;40
    • pdntspa8 days ago
      It&#x27;s almost always marketing and some stupid idea someone there had. I don&#x27;t know why non-technical people try and claim so much ownership over versioning. You nearly always end up with these ridiculous outcomes.<p>&quot;I know! Let&#x27;s restart the version numbering for no good reason!&quot; becomes DOOM (2016), Mortal Kombat 1 (2025), Battlefield 1 (2016), Xbox One (not to be confused with the original Xbox 1)<p>As another example, look at how much of a trainwreck USB 3 has become<p>Or how Nvidia restarted Geforce card numbering
      • recursive8 days ago
        Xbox should be in the hall of fame for terrible names.<p>There&#x27;s also Xbox One X, which is not in the X series. Did I say that right? Playstation got the version numbers right. I couldn&#x27;t make names as incomprehensible as Xbox if I tried.
        • pdntspa8 days ago
          Rumor has it they were jelly because the Playstation 3 had one higher version number than what would have been the Xbox 2, so it became the Xbox 360 instead. And then got further off the rails when its replacement arrived
    • recursive8 days ago
      &quot;4o&quot; was bad to begin with, as &quot;four-oh&quot; is a common verbalization of &quot;4.0&quot;.
  • ClassAndBurn8 days ago
    They will have to update the openai. Com footer I guess<p>Latest Advancements<p>GPT-5<p>OpenAI o3<p>OpenAI o4-mini<p>GPT-4o<p>GPT-4o mini<p>Sora
  • 9cb14c1ec08 days ago
    Theo can sleep tonight.