10 comments

  • adverbly4 hours ago
    This is very cool but it&#x27;s not quite what I expected out of economic primitives.<p>I expected to see measures of the economic productivity generated as a result of artificial intelligence use.<p>Instead, what I&#x27;m seeing is measures of artificial intelligence use.<p>I don&#x27;t really see how this is measuring the most important economic primitives. Nothing related to productivity at all actually. Everything about how and where and who... This is just demographics and usage statistics...
    • p1necone3 hours ago
      &gt; I expected to see measures of the economic productivity generated as a result of artificial intelligence use.<p>&gt;Instead, what I&#x27;m seeing is measures of artificial intelligence use.<p>Fun fact: this is also how most large companies are measuring their productivity increases from AI usage ;), alongside asking employees to tell them how much faster AI is making them while simultaneously telling them they&#x27;re expected to go faster with AI.
      • hazyc2 hours ago
        productivity is such a nebulous concept in knowledge work - an amalgamation of mostly-qualitative measures that get baked into quantitative measures that are mostly just bad data
        • reactordev1 hour ago
          You can thank agile for that
          • janwirth22 minutes ago
            You don&#x27;t seem to like agile, whatever that word even means.
    • johnrob3 hours ago
      Until AI is used to generate new revenue streams (i.e. acquire new customers), I don’t think the economic impact is going to impress. My two cents.
    • kurttheviking4 hours ago
      agree, was similarly hoping for something akin to a total factor productivity argument
  • siliconc0w4 hours ago
    Skimmed, some notes for a more &#x27;bear&#x27; case:<p>* value seems highly concentrated in a sliver of tasks - the top ten accounting for 32%, suggesting a fat long-tail where it may be less useful&#x2F;relevant.<p>* productivity drops to a more modest 1-1.2% productivity gain once you account for humans correcting AI failure. 1% is still plenty good, especially given the historical malaise here of only like 2% growth but it&#x27;s not like industrial revolution good.<p>* reliability wall - 70% success rate is still problematic and we&#x27;re getting down to 50% with just 2+ hours of task duration or about &quot;15 years&quot; of schooling in terms of complexity for API. For web-based multi-turn it&#x27;s a bit better but I&#x27;d imagine that would at least partly due to task-selection bias.
    • storystarling31 minutes ago
      I&#x27;ve found that architecting around that reliability wall is where the margins fall apart. You end up chaining verification steps and retries to get a usable result, which multiplies inference costs until the business case just doesn&#x27;t work for a bootstrapped product.
    • xiphias24 hours ago
      ,,1% is still plenty good, especially given the historical malaise here of only like 2% growth but it&#x27;s not like industrial revolution good.&#x27;&#x27;<p>You can&#x27;t compare the speed of AI improvements to the speed of technical improvements during the industrial revolution. ChatGPT is 3 years old.
  • mlsu4 hours ago
    &gt; This also highlights the importance of model design and training. While Claude is able to respond in a highly sophisticated manner, it tends to do so only when users input sophisticated prompts.<p>If the output of the model depends on the intelligence of the person picking outputs out of its training corpus, is the model intelligent?<p>This is kind of what I don&#x27;t quite understand when people talk about the models being intelligent. There&#x27;s a huge blindspot, which is that the prompt entirely determines the output.
    • TrainedMonkey4 hours ago
      Humans also respond differently when prompted in different ways. For example, politeness often begets politeness. I would expect that to be reflected in training data.
      • b00ty4breakfast2 minutes ago
        that seems like something you wouldn&#x27;t want from your tools. humans have that and that&#x27;s fine, people are people and have emotions but I don&#x27;t want my power-drill asking me why I only call when I need something.
      • mlsu4 hours ago
        If I, a moron, hire a PhD to crack a tough problem for me, I don&#x27;t need to go back and forth prompting him at a PhD level. I can set him loose on my problem and he&#x27;ll come back to me with a solution.
        • nl2 hours ago
          &gt; hire a PhD to crack a tough problem for me, I don&#x27;t need to go back and forth prompting him at a PhD level. I can set him loose on my problem and he&#x27;ll come back to me with a solution.<p>In my experience with many PhDs they are just as prone to getting off track or using their pet techniques as LLMs! And many find it very hard to translate their work into everyday language too...
        • Herring3 hours ago
          Well if it ever gets to be a full replacement for phds, you’ll know cause it will have already replaced you.
      • HPsquared3 hours ago
        I think that&#x27;s what is happening. It&#x27;s simulating a conversation, after all. A bit like code switching.
    • zozbot2344 hours ago
      What is a &quot;sophisticated prompt&quot;? What if I just tack on &quot;please think about this a lot and respond in a highly sophisticated manner&quot; to my question&#x2F;prompt? Anyone can do this once they&#x27;re made aware of this potential issue. Sometimes the UX layer even adds this for you in the system prompt, you just have to tick the checkbox for &quot;I want a long, highly sophisticated answer&quot;.
      • nl2 hours ago
        In general it will match the language style you use.<p>If you ask a sophisticated question (lots of clauses, college reading level or above) it will respond in kind.<p>You are basically moving where the generation happens in the latent space. By asking in a sophisticated way you are moving the latent space away from say children&#x27;s books and towards say PhD dissertations.
      • mlsu4 hours ago
        They have a chart that shows it. The education level of the input determines the education level of the output.<p>These things are supposed to have intelligence on tap. I&#x27;ll imagine this in a very simple way. Let&#x27;s say &quot;intellignce&quot; is like a fluid. It&#x27;s a finite thing. Intelligence is very valuable, it&#x27;s the substrate for real-world problem solving that makes these things ostensibly worth trillions of dollars. Intelligence comes from interaction with the world; someone&#x27;s education and experience. You spend some effort and energy feeding someone, clothing them, sending them to college. And then you get something out, which is intelligence that can create value for society.<p>When you are having a conversation with the AI, is the intelligence flowing out of the AI? Or is it flowing out of the human operator?<p>The answer to this question is extremely important. If the AI can be intelligent &quot;on its own&quot; without a human operator, then it will be very valuable -- feed electricity into a datacenter and out comes business value. But if a model is only intelligent as someone using it, well, the utility seems to be very harshly capped. At best it saves a bit of time, but it will never do anything novel, it will never create value on its own, independently, it will never scale beyond a 1:1 &quot;human picking outputs&quot;.<p>If you must encode intelligence into the prompt to get intelligence out of the model, well, this doesn&#x27;t quite look like AGI does it?
        • mlsu4 hours ago
          ofc what I&#x27;m getting at is, you can&#x27;t get something from nothing. There is no free lunch.<p>You spend energy distilling the intelligence of the entire internet into a set of weights, but you still had to expend the energy to have humans create the internet first. And on top of this, in order to pick out what you want from the corpus, you have to put some energy in: first, the energy of inference, but second and far more importantly, the energy of prompting. The model is valuable because the dataset is valuable; the model output is valuable because the prompt is valuable.<p>So wait then, where does this exponential increase in value come from again?
          • felixgallo3 hours ago
            the same place an increase in power comes from when you use a lever.
    • wat100004 hours ago
      A smart person will tailor their answers to the perceived level of knowledge of the person asking, and the sophistication of the question is a big indicator of this.
    • thousand_nights4 hours ago
      i don&#x27;t know, are we intelligent?<p>you could argue that our input (senses) entirely define the output (thoughts, muscle movements, etc)
      • HPsquared3 hours ago
        There&#x27;s a bit of baked-in stuff as well. We are a full culture-mind-body[-spirit] system.
    • sjajshha4 hours ago
      [dead]
  • dingdingdang4 hours ago
    The title actually cringes me out a bit, it reads like early report titles in academia where young students (myself no doubt incl back when) try their hardest at making a title sound clever but in actuality only achieve obscuration of their own material.
    • bilsbie3 hours ago
      Reminds me of psychohistory.
  • bix65 hours ago
    &gt; These “primitives”—simple, foundational measures of how Claude is used, which we generate by asking Claude specific questions about anonymized Claude.ai and first-party (1P) API transcripts<p>I just skimmed but is there any manual verification &#x2F; human statistical analysis done on this or we just taking Claude’s word for it?
    • sdwr4 hours ago
      Looks like they are relying on Claude for it, which is interesting. I bet social scientists are going to love this approach
  • ossa-ma4 hours ago
    I&#x27;m not an economist so can someone explain whether this stat is significant:<p>&gt; a sustained increase of 1.0 percentage point per year for the next ten years would return US productivity growth to rates that prevailed in the late 1990s and early 2000s<p>What can it be compared to? Is it on the same level of productivity growth as computers? The internet? Sliced bread?
  • mips_avatar5 hours ago
    Every single AI economic analysis talks about travel planning but none of the AI labs have the primitives (transit routing, geocoding, etc.) in a semantic interface for the models to use.
    • malshe4 hours ago
      Coincidentally, YouTube demos on vibe coding commonly make travel planning apps!
  • brap4 hours ago
    All of this performative bullshit coming out of Anthropic is slowly but surely making them my least favorite AI company.<p>We get it guys the very scary future is here any minute now and you’re the only ones taking it super seriously and responsibly and benevolently. That’s great. Now please just build the damn thing
    • ossa-ma4 hours ago
      These are economic studies on AI&#x27;s impact on productivity, jobs, wages, global inequality. It&#x27;s important to UNDERSTAND who benefits from technology and who gets left behind. Even putting the positive impacts of a study like this aside - this kinda due diligence is critical for them to understand developing markets and how to reach them.
      • futuraperdita2 hours ago
        But the thing is that they really aren&#x27;t rigorous economic studies. They&#x27;re a sort of UX research-like sociological study with some statistics, but don&#x27;t actually approach the topic with any sort of econometric modeling or give more than loose correlations to past economic data. So it does appear performative: it&#x27;s &quot;pop science&quot; using a quantitative veneer to push a marketing message to <i>business leaders</i> in a way that looks well-optimised mathematically.<p>Note the papers cited are nearly all ones <i>about AI use</i>, and align more closely with management case studies vs. economics.
      • brap4 hours ago
        Ok Dario
  • blibble4 hours ago
    &gt; How is AI reshaping the economy?<p>oh I know this one!<p>it&#x27;s created mountains of systemic risk for absolutely no payoff whatsoever!
    • andy_xor_andrew4 hours ago
      no payoff whatsoever? I just asked Claude to do a task that would have previously taken me four days. Then I got up and got lunch, and when I was back, it was done.<p>I would never make the argument that there are no risks. But there&#x27;s also no way you can make the argument there are no payoffs!
      • blibble3 hours ago
        &gt; I just asked Claude to do a task that would have previously taken me four days.<p>I think this probably says more about you than the &quot;AI&quot;
        • steve_adams_863 hours ago
          That&#x27;s not a very constructive thought given you don&#x27;t know what the task is or why it could have taken them days. In a field as large and complex as software, there are myriad reasons why any single person could find substantial time-saving opportunities with LLMs, and it doesn&#x27;t have to point to their own inadequacies.