11 comments

  • al-yak18 minutes ago
    Eternal question about pre-training, i.e. on what medium this pre-training or pre-pre-training has been performed ... if we assume that the platform is CMOS-based piece of hardware, then the most primitive trainable element would be a two-state learning automaton, that can be assembled out of a couple of dozen transistors. This is effectively a kind of a single cell bacteria. You then need to organize these automata into collectives which can start to exchange tokens and reinforcements between themselves and the training environment, and these collectives do so at a higher level and so on, and that&#x27;s what can be seen as NCAs. (*) A good source of idea on this approach can be found in the book of M.L. Tsetlin: Automata Theory and Modelling of Biological Systems, 1973 - <a href="https:&#x2F;&#x2F;shop.elsevier.com&#x2F;books&#x2F;automation-theory-and-modeling-of-biological-systems&#x2F;tsetlin&#x2F;978-0-12-701650-4" rel="nofollow">https:&#x2F;&#x2F;shop.elsevier.com&#x2F;books&#x2F;automation-theory-and-modeli...</a>
  • benob3 hours ago
    Reminds me of &quot;Universal pre-training by iterated random computation&quot; <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2506.20057" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2506.20057</a>, with bit less formal approach.<p>I wonder if there is a closed-form solution for those kinds of initialization methods (call them pre-training if you wish). A solution that would allow attention heads to detect a variety of diverse patterns, yet more structured than random init.
  • stanfordkid3 hours ago
    I did a similar project but using 3D fractals I found on shadertoy feeding into ViTs. They are extremely simple iterative functions that produce a ton of scene like complexity.<p>I have a pet theory that the visual cortex when developing is linked to some kind of mechanism such as this. You just need proteins that create some sort of resonating signal that feed into the neurons as they grow (obviously this is hand-wavy) but similar feedback loops guide nervous system growth in Zebra fish for example.
    • heyitsguay2 hours ago
      What were the results of 3d fractal shader pretraining?
    • andai3 hours ago
      I like your funny words, magic man!
  • gavinray1 hour ago
    Can someone ELI5 how this hypothesis could ever be true?<p><pre><code> &gt; &quot;The core hypothesis: what makes language useful for pre-training is its structure, not its semantics.&quot; </code></pre> As a layman, I&#x27;ve always held the intuition that semantics are the only meaningful thing.<p>&quot;Structure without semantics&quot; = form without function, symmetric&#x2F;regular noise, right?<p>My naive bet is on compressing semantics into mediums more expressive&#x2F;information dense than text. Like how some languages have single words&#x2F;symbols to represent entire sentence-long concepts.
    • andy12_40 minutes ago
      I think what they mean by this is that, for example, in &quot;If it&#x27;s raining the outside is wet. It&#x27;s raining, so the outside is wet&quot;, it&#x27;s more important for the model to learn &quot;If A then B. A, therefore B&quot; than to learn what &quot;raining&quot; , &quot;outside&quot; and &quot;wet&quot; mean.
  • andai3 hours ago
    &gt; The key: since every sequence has a unique latent rule, the model must infer that rule in-context to predict what comes next. This in-context learning ability underpins many of the key reasoning capabilities observed in language models.<p>This is a remarkable paper. This is the first time I&#x27;ve heard someone training the actual thing we&#x27;re trying to get this stuff to do!<p>---<p>&gt; This raises a radical question: Is natural language the only path to intelligence?<p>Of course not! We have octopi, ravens etc., which in many domain display higher intelligence than frontier AIs.<p>&quot;Embodied reasoning&quot; (genetic algorithm brute force solving physical tasks for a billion years, to name one solution) is definitely one very practical form of intelligence, although we&#x27;re taking some shortcuts in replicating it.<p>I&#x27;m wondering if simplified analog tasks like Box2D puzzled would help too (or perhaps even simpler? Hanoi? Block worlds?). I know many companies are using simulations of 3D worlds for that.<p>What I don&#x27;t understand is how that can integrate with the LLM (physical intelligence would seem to require specialized circuitry, if only for the latency). But maybe once we have good specialized models, LLMs can be trained on their synthetic data?
  • voxleone5 hours ago
    Neural cellular automata are interesting because they shift learning from “predict tokens” to “model state evolution.” That feels much closer to a transition-based view of systems, where structure emerges from repeated local updates (transitions) rather than being encoded explicitly.<p>I&#x27;m working on a theoretical&#x2F;computational framework, the Functional Universe, intended for modeling physical reality as functional state evolution. i would say it could be used to replicate your CA process. Won&#x27;t link it here to signal my good faith discussing this issue - it&#x27;s on my GH.
    • troelsSteegin3 hours ago
      from <a href="https:&#x2F;&#x2F;voxleone.github.io&#x2F;FunctionalUniverse&#x2F;pages&#x2F;executive-summary.html" rel="nofollow">https:&#x2F;&#x2F;voxleone.github.io&#x2F;FunctionalUniverse&#x2F;pages&#x2F;executiv...</a>, &quot;The Functional Universe models reality as a history built from irreversible transitions, with time emerging from the accumulation of causal commitments rather than flowing as a primitive parameter.&quot; Is it fair to say that time is simply a way of organizing a log file on a dynamic reality? I interpreted &quot;composition of transitions&quot; as a system of processes. I think the hard modeling problem is interpreting interactions between processes - that transitions don&#x27;t simply compose, that observed transitions may be confounded views of more complex transitions. I gather NCA would be granular enough to overcome that.
      • voxleone1 hour ago
        That’s a very good objection, and it’s pointing at a real pressure point in our framework.<p>Short answer: it’s close, but incomplete. It’s not that time organizes a log of reality; rather, reality <i>is</i> the accumulation of committed transitions. What you’re calling a ‘log’ it’s the ontological structure itself.<p>I gather you&#x27;re basically saying: what we see as a transition ≠ what’s actually happening at the fundamental level. This is a legitimate and deep problem.<p>You’re right that observed transitions may not compose cleanly. In the Functional Universe, composition is a property of fundamental transitions. What we observe are often coarse-grained projections of many underlying transitions, which can obscure compositional structure.
  • dzink4 hours ago
    “The long-term vision is: foundation models that acquire reasoning from fully synthetic data, then learn semantics from a small, curated corpus of natural language. This would help us build models that reason without inheriting human biases from inception.”
    • qsera4 hours ago
      I think this is a bit risky, because it assumes that all knowledge that a human posses about nature is acquired after birth.<p>But is that correct? I think organisms also come with a partial built in understanding of nature at birth.
      • throw-qqqqq3 hours ago
        &gt; I think organisms also come with a partial built in understanding of nature at birth<p>I agree. Most organisms are quite pre-trained: they have “instincts” and natural behaviors.<p>E.g. newly hatched turtles know to crawl towards the ocean immediately when they hatch. They don’t learn that on their way.<p>It seems to me that most lifeforms come into this world pre-trained.
      • jamilton3 hours ago
        I don’t think that assumption is being made, why do you think that? In terms of metaphor, training a model could be considered both knowledge acquired after birth and its evolution. But I don’t think it’s particularly useful to stay thinking in metaphors.
  • dmos623 hours ago
    Honestly, I never thought about reasoning this way, but it&#x27;s kind of obvious now that someone did it. Very interesting.
  • builderhq_io3 hours ago
    [dead]
  • Heer_J3 hours ago
    [dead]