18 comments

  • NitpickLawyer4 hours ago
    The paper is here - <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2603.19461" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2603.19461</a><p>This, IMO is the biggest insight into where we&#x27;re at and where we&#x27;re going:<p>&gt; Because both evaluation and self-modification are coding tasks, gains in coding ability can translate into gains in self-improvement ability.<p>There&#x27;s a thing that I&#x27;ve noticed early into LLMs: once they unlock one capability, you can use that capability to compose stuff and improve on other, related or not, capabilities. For example &quot;reflexion&quot; goes into coding - hey, this didn&#x27;t work, let me try ... Then &quot;tools&quot;. Then &quot;reflxion&quot; + &quot;tools&quot;. And so on.<p>You can get workflows that have individual parts that aren&#x27;t so precise become better by composing them, and letting one component influence the other. Like e2e coding gets better by checking with &quot;gof&quot; tools (linters, compilers, etc). Then it gets even better by adding a coding review stage. Then it gets even better by adding a static analysis phase.<p>Now we&#x27;re seeing this all converge on &quot;self improving&quot; by combining &quot;improving&quot; components. And so on. This is really cool.
    • alexpotato41 minutes ago
      &gt;You can get workflows that have individual parts that aren&#x27;t so precise become better by composing them, and letting one component influence the other. Like e2e coding gets better by checking with &quot;gof&quot; tools (linters, compilers, etc). Then it gets even better by adding a coding review stage. Then it gets even better by adding a static analysis phase.<p>This is the exact point I make whenever people say LLMs aren&#x27;t deterministic and therefore not useful.<p>Yes, they are &quot;stochastic&quot;. But you can use them to write deterministic tools that create machine readable output that the LLM can use. As you mention, you keep building more of these tools and tying them together and then you have a deterministic &quot;network&quot; of &quot;lego blocks&quot; that you can run repeatably.
    • binarymax4 hours ago
      I disagree that evaluation is always a coding task. Evaluation is scrutiny for the person who wants the thing. It’s <i>subjective</i>. So, unless you’re evaluating something purely objective, such as an algorithm, I don’t see how a self contained, self “improving “ agent accomplishes the subjectivity constraint - as by design you are leaving out the subject.
      • NitpickLawyer3 hours ago
        Sure. There will always be subjective tasks where the person who asks for something needs to give feedback. But even there we could come up with ways to make it easier &#x2F; faster &#x2F; better ux. (one example I saw my frontend colleagues do is use a fast model to create 9 versions of a component, in a grid. And they &quot;at a glance&quot; decide which one is &quot;better&quot;, and use that going forwards).<p>OTOH, there&#x27;s <i>loads</i> you can do for evaluation before a human even sees the artifact. Things like does the site load, does it behave the same, did anything major change on the happy path, etc etc. There&#x27;s a recent-ish paper where instead of classic &quot;LLM as a judge&quot; they used LLMs to come up with rubrics, and other instances check original prompt + rubrics on a binary scale. Saw improvements in a lot of evaluations.<p>Then there&#x27;s &quot;evaluate by having an agent do it&quot; for any documentation tracking. Say you have a project, you implement a feature, and document the changes. Then you can have an agent take that documentation and &quot;try it out&quot;. Should give you much faster feedback loops.
      • ranyume3 hours ago
        In science there are ways to surface subjectivity (cannot be counted) into observable quantized phenomena. Take opinion polls for instance: &quot;approval&quot; of a political figure can mean many things and is subjective, but experts in the field make &quot;approval&quot; into a number through scientific methods. These methods are just an approximation and have many IFs, they&#x27;re not perfect (and for presidential campaign analysis in particular they&#x27;ve been failing for reasons I won&#x27;t clarify here), but they&#x27;re useful nonetheless.<p>Another thing that get quantized is video preferences to maximize engagement.
    • sbinnee30 minutes ago
      I guess this paper is part of ICML coming soon this June. I hope to see a lot of cool papers.
    • lukebuehler3 hours ago
      Agree. It&#x27;s code all the way down. The key is to give agents a substrate where they can code up new capabilities and then compose them meaningfully and safely.<p>Larger composition, though, starts to run into typical software design problems, like dependency graphs, shared state, how to upgrade, etc.<p>I&#x27;ve been working on this front for over two years now too: <a href="https:&#x2F;&#x2F;github.com&#x2F;smartcomputer-ai&#x2F;agent-os&#x2F;" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;smartcomputer-ai&#x2F;agent-os&#x2F;</a>
      • whattheheckheck1 hour ago
        Oh wow, what do you think of karpathys autoresearch? Feels like this is just that? Gotta openclawify it?
      • whattheheckheck1 hour ago
        So what are software packages now a days other than precomputed subsets of capabilities. Like a mesh that data gets pushed through to produce what? What are the optimal subset of prebuilt programs to accomplish any task?
    • alansaber2 hours ago
      The whole theme of llm dev to date has been &quot;theres more common than not&quot; in llm applications
    • testaccount283 hours ago
      because submarine piloting is a going-under-water activity, improvements in holding one&#x27;s breath can lead to faster submersibles.
    • Atomic_Torrfisk1 hour ago
      Im sorry, this just sounds like hypespeak. CAn you provide samples?<p>&gt; once they unlock one capability,<p>What does it mean to unlock? Its an llm nothing is locked. The output is a as good as the context, model and environment. Nothing is hidden or locked.
      • IncreasePosts1 hour ago
        Maybe unlock means &quot;recognize and solve a problem with an order of magnitude fewer tokens than the first time you did it&quot;. The same way humans might spend a lot of time thinking about a certain problem and various ways to solve it, but once they go through that process, and then recognize it again, they don&#x27;t need to go to the same process and jump right to the solution.
  • Jerrrrrrrry4 hours ago
    No matter how far we go, we end up with generation &#x2F; discrimination architecture.<p>Its is the core of any and all learning&#x2F;exellency; exposure to chaotic perturbations allow selection of solutions that are then generalized to further, ever more straining problems; producing increasingly applicable solutions.<p>This is the core of evolution, and is actually derivable from just a single rule.
    • gobdovan3 hours ago
      I don&#x27;t think generation&#x2F;discrimination is fundamental. A more general framing is evolutionary epistemology (Donald T. Campbell, 1974, essay found in &quot;The Philosophy of Karl Popper&quot;), which holds that knowledge emerges through variation and selective retention. As Karl Popper put it, &quot;We choose the theory which best holds its own in competition with other theories; the one which, by natural selection, proves itself the fittest to survive.&quot;<p>On this view, learning in general operates via selection under uncertainty. This is less visible in individual cognition, where we tend to over-attribute agency, but it is explicit in science: hypotheses are proposed, subjected to tests, and selectively retained, precisely because the future cannot be deduced from the present.<p>In that sense, generation&#x2F;discrimination is a particular implementation of this broader principle (a way of instantiating variation and selection) not the primitive itself.
      • Jerrrrrrrry1 hour ago
        I agree, I meant to be explicit that the one rule was &quot;gravity&quot;;<p>Variation (chaos) comes from the tidal push&#x2F;pull of all cumulative processes - all processes are nearly periodic (2nd law) and get slower - guaranteeing oscillator harmonics at intervals.<p>These intervals are astronomically convulted, but still promise a Fourier distribution of frequency: tidal effects ensure synchronization eventually, as all periods resonate eventually.<p>As systems are increasingly exposed to pendulums of positive and negative coherence, they will generalize for variance, and eventually for increasingly (fourier) selective filters of increasingly resiliente traits, that will generalize.<p>The system would eventually be increasingly resilient and eventually an awareness would develop.<p>Awareness of past periodic cycles would improve fitness (with or without consciousness) and eventually the mechanistic processes would be in the systems nature.<p>This is why we have pointless traditions, folk lore, collective unconscious artifacts, cyclical cataclysmic religions, the Fermi Paradox, the great filters...<p>Variation and selection are woven, but understanding how it all stems from gravity by means of nearly perioidic oscillators (spinning planets, tidal pools, celestial bodies) due to the conservation of angular momentum, due to the 3body problem.....that is what took a genius to reconcile
        • zoogeny58 minutes ago
          &gt; eventually an awareness would develop<p>I am not sure how this is a necessary conclusion to the premises you provide.
          • Jerrrrrrrry30 minutes ago
            Awareness would be any form of agency, goal seeking, or loss minimizing.<p>As Briggs–Rauscher reactions can eventually lead to Belousov–Zhabotinsky reactions, the system can maintain homeostasis with its environment (and continuing to oscillate) by varying reactants in a loss minimizing fashion.<p>This loss minimizing would be done during scarcity to limp towards an abundance phase.<p>This is the mechanism that hypothetical tidal pools batteries would had exhibited to continue between periods of sunlight&#x2F;darkness&#x2F;acidity that eventually gets stratified as a resilency trait.
            • zoogeny25 minutes ago
              I&#x27;m not sure what your argument is here, except stating an opinion that loss minimization is equivalent to agency. But even if that was accepted, which is a huge stretch, it doesn&#x27;t stretch all the way to awareness.
        • whattheheckheck1 hour ago
          So where does gravity come from?
          • Jerrrrrrrry39 minutes ago
            A cool illusion, just another emergent property of our geometrical solution: higher dimensional aperiodic tilings of a 10^80 faceted complex polyhedra &quot;walking&quot; on another large aperioidic Penrose plane, that is getting smaller in a dimension we observe as &quot;energy&quot;.<p>Basically a dice with a bajillion sides is getting rolled along an increasingly slim poker table, house winning eventually.<p>Time only goes one way, protons dont decay, energy is radiated unto the cosmic background hiss, until homogeneity is reached as CMB, and entrophy reaches 1.<p>I dont know where it comes from, but I know the shape it makes as it rolls by.
    • ilaksh4 hours ago
      It&#x27;s a feedback loop.<p>I&#x27;ve always felt that the most important part of engineering was feedback loops.<p>Maybe nature is the greatest engineer ever?
      • 0xbadcafebee2 hours ago
        The most important part of engineering is problem-solving, which feedback loops don&#x27;t necessarily do. The reason we are here as engineers is: 2.5 billion years ago, the earth made cyanobacteria, which flourished, then flooded the earth with toxic oxygen, killing almost all life on the planet. The initial feedback loop didn&#x27;t solve a problem, it destroyed a use case. That&#x27;s not a solution to a problem that an engineer would choose, even if those organisms that came after were pretty happy about it...
        • Jerrrrrrrry1 hour ago
          Systems emerge in times of abundance, and are whittled in times of scarcity.<p>The great oxygenation was a time of near catyclismsic scarcity for most complex organisms, as resources scale to food&#x2F;energy requirements imply the most complex organisms were the most dependent on the environment, and were most impacted by changes.<p>Inversely, oxygenation was our most crucial abundancy pre cursor, as it provides a large substrate chemically for life to exhibit
  • mifydev1 hour ago
    I&#x27;ve been experimenting with similar concept myself. The linter loop is the only thing that can keep the agent sane in my opinion, and if anyone can generalize bun+tsc loop to other tasks, this would finally be a way to trust LLMs output.<p>I was annoyed at how Claude Code ignores my CLAUDE.md and skills, so I was looking for ways to expand type checking to them. So I wrote a wrapper on top of claude-agents-sdk that reads my CLAUDE.md and skills, and compiles them into rules - could be linter rules or custom checking scripts. Then it hooks up to all tools and runs the checks. The self improving part comes if some rule doesn&#x27;t work: I run the tool with the session id in review mode, it proposes the fixes and improves the rule checkers. (not the md files) So it&#x27;s kinda like vibe coding rules, definitely lowers the bar for me to maintain them. Repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;chebykinn&#x2F;agent-ruler" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;chebykinn&#x2F;agent-ruler</a>
  • supermdguy1 hour ago
    It&#x27;s surprising that this works so well considering that AI-generated AGENTS.md files have been shown to be not very useful. I think the key difference here is that the real-world experience helps the agent reach regions of its latent space that wouldn&#x27;t occur naturally through autoregression.<p>I wonder how much of the improvement is due to the agent actually learning <i>new things</i> vs. reaching parts of its latent space that enable it to recall things it already knows. Did the agent come up with novel RL reward design protocols based on trial and error? Or did the tokens in the environment cause it to &quot;act smarter&quot;?
  • Xcelerate44 minutes ago
    I’m still curious how any (finitely describable) approach to “self-improving” doesn’t violate Levin’s conservation of mutual algorithmic information.
  • kordlessagain2 hours ago
    Uses LiteLLM. Lovely.
  • flockonus4 hours ago
    The readme seems very unclear about what it does. Anyone has a practical example of it?
    • pegasus4 hours ago
      There&#x27;s a paper at <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2603.19461" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2603.19461</a><p>Abstract:<p>Self-improving AI systems aim to reduce reliance on human engineering by learning to improve their own learning and problem-solving processes. Existing approaches to self-improvement rely on fixed, handcrafted meta-level mechanisms, fundamentally limiting how fast such systems can improve. The Darwin Gödel Machine (DGM) demonstrates open-ended self-improvement in coding by repeatedly generating and evaluating self-modified variants. Because both evaluation and self-modification are coding tasks, gains in coding ability can translate into gains in self-improvement ability. However, this alignment does not generally hold beyond coding domains. We introduce \textbf{hyperagents}, self-referential agents that integrate a task agent (which solves the target task) and a meta agent (which modifies itself and the task agent) into a single editable program. Crucially, the meta-level modification procedure is itself editable, enabling metacognitive self-modification, improving not only the task-solving behavior, but also the mechanism that generates future improvements. We instantiate this framework by extending DGM to create DGM-Hyperagents (DGM-H), eliminating the assumption of domain-specific alignment between task performance and self-modification skill to potentially support self-accelerating progress on any computable task. Across diverse domains, the DGM-H improves performance over time and outperforms baselines without self-improvement or open-ended exploration, as well as prior self-improving systems. Furthermore, the DGM-H improves the process by which it generates new agents (e.g., persistent memory, performance tracking), and these meta-level improvements transfer across domains and accumulate across runs. DGM-Hyperagents offer a glimpse of open-ended AI systems that do not merely search for better solutions, but continually improve their search for how to improve.
      • functional_dev3 hours ago
        This &#x27;self vs non-self&#x27; logic is very similar to how plants prevent self-pollination. They have a biological &#x27;discrimination&#x27; system to recognize and reject their own genetic code.<p>Here is a breakdown - <a href="https:&#x2F;&#x2F;vectree.io&#x2F;c&#x2F;plant-self-incompatibility-logic" rel="nofollow">https:&#x2F;&#x2F;vectree.io&#x2F;c&#x2F;plant-self-incompatibility-logic</a>
    • OutThisLife4 hours ago
      Hermes agent does this, if you&#x27;re curious<p><a href="https:&#x2F;&#x2F;github.com&#x2F;NousResearch&#x2F;hermes-agent" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;NousResearch&#x2F;hermes-agent</a>
      • willy_k1 hour ago
        Seems like that only has the task improvement loop, no self-improvement improvement loop like this project.
  • sonu273 hours ago
    Can someone add this to OpenClaw :)
  • measurablefunc3 hours ago
    That&#x27;s great but how about UltraAgents: Meta-referential meta-improving self-referential hyperagents?
    • 2001zhaozhao1 hour ago
      AGI-MegaAgent 5.7 Pro Ultra
      • measurablefunc1 hour ago
        Somehow still financed w&#x2F; ads &amp; ubiquitous surveillance.
  • jauntywundrkind4 hours ago
    Pi is self modifying, self aware. <a href="https:&#x2F;&#x2F;lucumr.pocoo.org&#x2F;2026&#x2F;1&#x2F;31&#x2F;pi&#x2F;" rel="nofollow">https:&#x2F;&#x2F;lucumr.pocoo.org&#x2F;2026&#x2F;1&#x2F;31&#x2F;pi&#x2F;</a><p>But this idea of having a task agent &amp; meta agent maybe has wings. Neat submission.
    • ghywertelling4 hours ago
      What are the differences wrt Recursive Language Models
      • adw3 hours ago
        Completely unrelated. Recursive Language Models are just &quot;what if we replaced putting all the long text into the context window with a REPL which lets you read parts of the context through tool calls and launch partitioned subagents&quot;, ie divide-and-conquer applied to attention space.
  • llmslave4 hours ago
    I think even code bases will have self improving agents. Software is moving from just the product code, to the agent code that maintains the product. Engineering teams&#x2F;companies that move in this direction will vastly out produce others.<p>I&#x27;ve had to really shift how I think about building code bases, alot of logic can go into claude skills and sub agents. Requires essentially relearning software engineering
    • _pdp_16 minutes ago
      We do this already but I bet this is not how people imagine it to be. There is still a review process to accept contributions.
  • Archiebuilds11 minutes ago
    [dead]
  • 11thDwarf25 minutes ago
    [flagged]
  • leontloveless2 hours ago
    [dead]
  • felixagentai3 hours ago
    [dead]
  • maxbeech3 hours ago
    [dead]
  • agentpiravi2 days ago
    [dead]
  • andyg_blog2 days ago
    [dead]