30 comments

  • devnullbrain13 hours ago
    I don&#x27;t see why this wouldn&#x27;t just lead to model collapse:<p><a href="https:&#x2F;&#x2F;www.nature.com&#x2F;articles&#x2F;s41586-024-07566-y" rel="nofollow">https:&#x2F;&#x2F;www.nature.com&#x2F;articles&#x2F;s41586-024-07566-y</a><p>If you&#x27;ve spent any time using LLMs to write documentation you&#x27;ll see this for yourself: the compounding will just be rewriting valid information with less terse information.<p>I find it concerning Karpathy doesn&#x27;t see this. But I&#x27;m not surprised, because AI maximalists seem to find it really difficult to be... &quot;normal&quot;?<p>Rule of thumb: if you find yourself needing to broadcast the special LLM sauce you came up with instead of what it helped you produce, ask yourself why.
    • gojomo10 hours ago
      Here in 2026, many forms of training LLMs on (well-chosen) outputs of themselves, or other LLMs, have delivered gigantic wins. So 2024 &amp; earlier fears of &#x27;model collapse&#x27; will lead your intuition astray about what&#x27;s productive.<p>It is unlikely you are accurately perceiving some limitation that Karpathy does not.
    • ChrisGreenHeur11 hours ago
      The article is not on training LLMs. it is about using LLMs to write a wiki for personal use. The article assumes a fully trained LLM such as ChatGPT or Claude already exists to be used.
      • khalic6 hours ago
        Don&#x27;t even try, after vibe coding, people seem to be adopting vibe thinking. &quot;Model Collapse sounds cool, I&#x27;m gonna use it without looking up&quot;
        • mikkupikku3 hours ago
          Vibe thinking... that&#x27;s an interesting premise. I&#x27;ll have to build up my new llm-wiki before I&#x27;ll know what to think about &quot;vibe thinking.&quot;
          • mikkupikku1 hour ago
            I was joking but also not joking, this llm-wiki idea is fun. I fed into it it&#x27;s own llm-wiki.md, Foucault&#x27;s Pendulum, randomly collected published papers about the philosophy of GiTS, several CCRU essays, and Manufacturing Consent. It drew fun red yarn between all of them, about the topic of red yarn (e.g. schizos drawing connections out of nothing, particularly through the use of computers, and how this relates to itself doing literally this as it does it.)<p>I&#x27;ll spare you most of the slop but.. <i>&quot;The Case That I Am Abulafia: The parallel is uncomfortable and precise. [...]&quot;</i><p>Yeah... It&#x27;s fun though.
    • jahala7 hours ago
      I did a proof of concept for self-updating html files (polyglot bash&#x2F;html) some weeks ago. It actually works quite well, with simple prompting it seems to not just go in circles (<a href="https:&#x2F;&#x2F;github.com&#x2F;jahala&#x2F;o-o" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;jahala&#x2F;o-o</a>)
    • sebmellen11 hours ago
      <i>Edit for context: the sibling comment from karpathy is gone after being flagged to oblivion. Not sure if he deleted it or if it was just removed based on the number of flags? He had copy-pasted a few snarky responses from Claude and essentially said “Claude has this to say to you:” followed by a super long run on paragraph of slop.</i><p>————<p>Wow, I respect karpathy so much and have learned a ton from him. But WTF is the sibling comment he wrote as a response to you? Just pasting a Claude-written slop retort… it’s sad.<p>Maybe we need to update that old maxim about “if you don’t have something nice to say, don’t say it” to “if you don’t have something human to say, don’t say it.”<p>So many really smart people I know have seen the ‘ghost in the machine’ and as a result have slowly lost their human faculties. Ezra Klein, of all people, had a great article about this recently titled “I Saw Something New in San Francisco” (gift link if you want to read it): <a href="https:&#x2F;&#x2F;www.nytimes.com&#x2F;2026&#x2F;03&#x2F;29&#x2F;opinion&#x2F;ai-claude-chatgpt-gemini-mcluhan.html?unlocked_article_code=1.YlA.WVXx.E5Q4sB2BnVSU" rel="nofollow">https:&#x2F;&#x2F;www.nytimes.com&#x2F;2026&#x2F;03&#x2F;29&#x2F;opinion&#x2F;ai-claude-chatgpt...</a>
      • prodigycorp9 hours ago
        It&#x27;s not sad. He&#x27;s a person like you and me. devnullbrain&#x27;s comment is snarky. He invoked model collapse which has nothing to do with the topic of a wiki&#x2F;kb, he wrote that karpathy is not normal, and then seemed to imply that the idea was useless. I&#x27;d be pretty in my feels and the fact that he wrote it and deleted it seems like a +1 normal guy thing.
      • girvo7 hours ago
        Eh, he’s just a person. I’m not surprised he posted a rude comment haha, and it got rightfully flagged off the site for being AI slop.<p>Appreciate the gift link, I’ll give it a read!
      • moralestapia10 hours ago
        Lol at that.<p>It&#x27;s weird how some people cover the whole range of putting out some really good stuff and other times the complete opposite.<p>Feels as if they were two different people ... or three, or four.
        • iamflimflam16 hours ago
          Emotional state, tiredness, drunkenness, a goods nights sleep… the number of factors that drive our responses is ridiculous.
    • kwar1311 hours ago
      also my experience. it can&#x27;t even keep up with a simple claude.md let alone a whole wiki...
  • Imanari9 hours ago
    Isn’t this just kicking the can down the road?<p>&gt; but the LLM is rediscovering knowledge from scratch on every question<p>Unless the wiki stays fully in context now the LLM hast to re-read the wiki instead of re-reading the source files. Also this will introduce and accumulate subtle errors as we start to regurgitate 2nd-order information.<p>I totally get the idea but I think next gen models with 10M context and&#x2F;or 1000tps will make this obsolete.
    • lelanthran5 hours ago
      &gt; I totally get the idea but I think next gen models with 10M context and&#x2F;or 1000tps will make this obsolete.<p>We&#x27;ve already got 1m context, 800k context, and they still start &quot;forgetting&quot; things around the 200k - 300k mark.<p>What use is 10M context if degradation starts at 200k - 300k?
    • SOLAR_FIELDS8 hours ago
      I use a home baked system based on obsidian that is essentially just “obsidian but with structured format on top with schemas” and I deploy this in multiple places with ranges of end users. It is more valuable than you think. The intermediary layer is great for capturing intent of design and determining when implementation diverges from that. There will always be a divergence from the intent of a system and how it actually behaves, and the code itself doesn’t capture that. The intermediate layer is lossy, it’s messy, it goes out of date, but it’s highly effective.<p>It’s not what this person is describing though. A self referential layer like this that’s entirely autonomous does feel completely valueless - because what is it actually solving? Making itself more efficient? The frontier model providers will be here in 3 weeks doing it better than you on that front. The real value is having a system that supports a human coming in and saying “this is how the system should actually behave”, and having the system be reasonably responsive to that.<p>I feel like a lot of the exercises like op are interesting but ultimately futile. You will not have the money these frontier providers do, and you do not have remotely the amount of information that they do on how to squeeze the most efficiency in how they work. Best bet is to just stick with the vanilla shit until the firehose of innovation slows down to something manageable, because otherwise the abstraction you build is gonna be completely irrelevant in two months
    • dennisy2 hours ago
      The “next gen of models” argument is a valid one and one I think of often, but if you truly buy it, it would stop do creating anything - since the next gen of models could make it obsolete.
    • khalic6 hours ago
      The goal isn’t to keep the context every time, it’s to make the memory queryable. Like a data lake but for your ideas and decisions
    • 0123456789ABCDE6 hours ago
      this solves for now, and this solves for the future.<p>now you get to condense the findings that interest from a handful of papers<p>in the future it solves for condensing your interests in a whole field to a handful of papers or less
    • nidnogg6 hours ago
      It is how I feel when I do it. And it certainly shows over time.
  • Vetch15 hours ago
    This sounds very like Licklider&#x27;s essay on Intelligence Amplification: Man Computer Symbiosis, from 1960:<p>&gt; Men will set the goals and supply the motivations, of course, at least in the early years. They will formulate hypotheses. They will ask questions. They will think of mechanisms, procedures, and models. They will remember that such-and-such a person did some possibly relevant work on a topic of interest back in 1947, or at any rate shortly after World War II, and they will have an idea in what journals it might have been published. In general, they will make approximate and fallible, but leading, contributions, and they will define criteria and serve as evaluators, judging the contributions of the equipment and guiding the general line of thought.<p>&gt; In addition, men will handle the very-low-probability situations when such situations do actually arise. (In current man-machine systems, that is one of the human operator&#x27;s most important functions. The sum of the probabilities of very-low-probability alternatives is often much too large to neglect. ) Men will fill in the gaps, either in the problem solution or in the computer program, when the computer has no mode or routine that is applicable in a particular circumstance.<p>&gt; The information-processing equipment, for its part, will convert hypotheses into testable models and then test the models against data (which the human operator may designate roughly and identify as relevant when the computer presents them for his approval). The equipment will answer questions. It will simulate the mechanisms and models, carry out the procedures, and display the results to the operator. It will transform data, plot graphs (&quot;cutting the cake&quot; in whatever way the human operator specifies, or in several alternative ways if the human operator is not sure what he wants). The equipment will interpolate, extrapolate, and transform. It will convert static equations or logical statements into dynamic models so the human operator can examine their behavior. In general, it will carry out the routinizable, clerical operations that fill the intervals between decisions.<p><a href="https:&#x2F;&#x2F;www.organism.earth&#x2F;library&#x2F;document&#x2F;man-computer-symbiosis" rel="nofollow">https:&#x2F;&#x2F;www.organism.earth&#x2F;library&#x2F;document&#x2F;man-computer-sym...</a>
    • ramoz14 hours ago
      Wow. fascinating insights he had.<p>e.g. (amongst many others) <i>Desk-Surface Display and Control: Certainly, for effective man-computer interaction, it will be necessary for the man and the computer to draw graphs and pictures and to write notes and equations to each other on the same display surface. The man should be able to present a function to the computer, in a rough but rapid fashion, by drawing a graph. The computer should read the man&#x27;s writing, perhaps on the condition that it be in clear block capitals, and it should immediately post, at the location of each hand-drawn symbol, the corresponding character as interpreted and put into precise type-face.</i>
  • kenforthewin15 hours ago
    This is just RAG. Yes, it&#x27;s not using a vector database - but it&#x27;s building an index file of semantic connections, it&#x27;s constructing hierarchical semantic structures in the filesystem to aid retrieval .. this is RAG.<p>On a sidenote, I&#x27;ve been building an AI powered knowledge base (yes, it uses RAG) that has wiki synthesis and similar ideas, take a look at <a href="https:&#x2F;&#x2F;github.com&#x2F;kenforthewin&#x2F;atomic" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;kenforthewin&#x2F;atomic</a>
    • panarky11 hours ago
      There&#x27;s nothing about RAG that requires embeddings.<p>The retrieval part can be grep if you don&#x27;t care about semantic search.
    • Jet_Xu14 hours ago
      I believe Multimodal KB+Agentic RAG is a suitable solution for personal KB. Imagine you have tons of office docs and want to dig some complex topics within it. You could try <a href="https:&#x2F;&#x2F;github.com&#x2F;JetXu-LLM&#x2F;DocMason" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;JetXu-LLM&#x2F;DocMason</a><p>Fully retrieve all diagram or charts info from ppt and excels, and then leverage Native AI agents(e.g. Codex) to conduct Agentic Rad
    • darkhanakh15 hours ago
      eh i&#x27;d push back on &quot;just RAG&quot;. like yes the retrieval-generation loop is RAG shaped, no ones arguing that. but the interesting bit here is the write loop - the LLM is authoring and maintaining the wiki itself, building backlinks, filing its own outputs back in. thats not retrieval thats knowledge synthesis. in vanilla RAG your corpus is static, here it isnt<p>also the linting pass is doing something genuinely different - auditing inconsistencies, imputing missing data, suggesting connections. thats closer to assistant maintaining a zettelkasten than a search engine returning top-k chunks<p>cool project btw will check it out
      • kenforthewin15 hours ago
        I agree with you, the linting pass seems valuable and it&#x27;s something I&#x27;m thinking about adding - it&#x27;s a great idea.<p>What I&#x27;m pushing back on specifically is the insistence that the core loop - retrieving the most relevant pieces of knowledge for wiki synthesis - is not RAG. In order for the LLM to do a good job at this, it needs some way to retrieve the most relevant info. Whether that&#x27;s via vector DB queries or a structured index&#x2F;filesystem approach, that fundamental problem - retrieving the best data for the LLM&#x27;s context - is RAG. It&#x27;s a problem that has been studied and evaluated for years now.<p>thanks for checking it out
      • Covenant002812 hours ago
        I&#x27;m curious how this linting step scales with larger wikis. Looking for an inconstency across N files requires N*N comparisons, and that&#x27;s assuming each file contains a single idea.
        • ChrisGreenHeur11 hours ago
          Presumably, randomness and only looking at a limited subset will semi-ensure over time that most contradictions will surface. Alternatively, how large do you really expect this kind of thing to be, there is a limit to the amount of facts from Warhammer 40k worth saving in a wiki.
      • devmor14 hours ago
        This is just persistent memory RAG. I have had a setup like this since about a day after I started using copilot, except it&#x27;s an MCP server that uses sqlite-vec and has recall endpoints to contextually load the proper data instead of a bunch of extra files polluting context.<p>OP&#x27;s example isn&#x27;t something new or incredibly thoughtful at all - in fact this pattern gets &quot;discovered&quot; every other day here, reddit or social media in general by people that don&#x27;t have the foresight to just look around and see what other people are doing.
    • alfiedotwtf14 hours ago
      You should have started your comment with “ I have a few qualms with this app”.<p>I’ve been thinking something along the lines of a LLM-WIKI for a while now which could truely act as a wingman-executive-assistant-second-brain, but OP has gone deeper than my ADHD thoughts could have possibly gone.<p>Looking forward to seeing this turn into fruition
    • locknitpicker10 hours ago
      &gt; This is just RAG.<p>More to the point, this is how LLM assistants like GitHub Copilot use their custom instructions file, aka copilot-instructions.md<p><a href="https:&#x2F;&#x2F;docs.github.com&#x2F;en&#x2F;copilot&#x2F;how-tos&#x2F;configure-custom-instructions&#x2F;add-repository-instructions" rel="nofollow">https:&#x2F;&#x2F;docs.github.com&#x2F;en&#x2F;copilot&#x2F;how-tos&#x2F;configure-custom-...</a>
  • nidnogg7 hours ago
    I&#x27;ve recently lazied out big time on a company project going down a similar rabbit hole. After having a burnout episode and dealing with sole caregiver woes in the family for the past year, I&#x27;ve had less and less energy to piece together intense, correct thought sequences at work.<p>As such I&#x27;ve taken to delegating substantial parts architecture and discovery to multiagent workflows that always refer back to a wiki-like castle of markdown files that I&#x27;ve built over time with them, fronted by Obsidian so I can peep efficiently often enough.<p>Now I&#x27;m certainly doing something wrong, but the gaps are just too many to count. If anything, this creates a weird new type of tech debt. Almost like a persistent brain gap. I miss thinking harder and I think it would get me out of this one for sure. But the wiki workflow is just too addictive to stop.
    • stingraycharles4 hours ago
      &gt; I miss thinking harder<p>Me too, and I wonder where this will take us; I worry about losing the ability to think hard.
      • jareklupinski2 hours ago
        im hoping to re-use the newly garbage-collected memory available to me now to rediscover &quot;play hard&quot;
    • kubb5 hours ago
      You’re not doing anything wrong. This isn’t a bulletproof idea. It can work, and this is what a lot of people end up with to manage complexity, but there’s a critical point beyond which things collapse: the agent can’t keep the wiki up to date anymore, the developer can’t grok it anymore.
      • kaashif4 hours ago
        Managing complexity, modularity, separation of concerns, were already critical for ensuring humans could still hold enough of the system in their brains to do something useful.<p>People who do not understand that will continue to not understand that it also applies to AI right now. Maybe at some point in the future it won&#x27;t, not sure. But my impression is that systems grow in complexity far past the point where the system is gummed up and no-one can do anything, unless it&#x27;s actively managed.<p>If a human can understand 10 units of complexity and their LLM can do 20, then they might just build a system that&#x27;s 30 complex and not understand the failure modes until it&#x27;s too late.
        • stingraycharles3 hours ago
          &gt; People who do not understand that will continue to not understand that it also applies to AI right now.<p>I think this is mostly a matter of expectation management. AIs are being positioned as being able to develop software independently, and that’s certainly the end goal.<p>So then people come in with the expectation that the AI is able to manage that, and it fails. Spectacularly.<p>The LLM can certainly not manage any non-local complexity right now, and succeed in increasing the technical debt and complexity faster than ever before.
    • loveparade6 hours ago
      That has been my experience as well. Most of the value of writing docs or a wiki is not in the final artifacts, it&#x27;s that the process of writing docs updates your own mental models and knowledge so that you can make better decisions down the road.<p>Even if you can get an LLM to output good artifacts that don&#x27;t eventually evolve into slop, which is questionable, it&#x27;s really not that useful, especially not for a personal wiki.
      • kilroy1235 hours ago
        Makes me think of all these tools that use AI to make fancy flashcards for you to study.<p>It seems rather silly to me, as _creating_ those flashcards is what helps you learn, with the studying after, cementing that knowledge in your brain.
        • kaashif4 hours ago
          The ultimate connecting up of the dots would be brain implants that just give you knowledge with zero effort.<p>I don&#x27;t know if I&#x27;d ever be comfortable with that, hopefully I&#x27;ll just be retired or dead when that takes off.
      • nidnogg6 hours ago
        And what happens when the bucket of knowledge gets too big and starts to overflow? I feel as if, by delegating that process of building knowledge too much, I end up accruing knowledge gaps of my own. Funnily enough it mirrors the LLM&#x2F;agent&#x27;s performance.<p>Maybe my recent prompts reflects how badly up to speed I am at a given time? I don&#x27;t know. A slightly related note - I recently heard the term AI de-skilling, this treads close to it imo.
      • nidnogg5 hours ago
        The worst part to me, by far, is having nothing more than a bunch of &quot;smart&quot; markdown files to show as my deliverables for the day. Sometimes this stacks for many days on end. Usually the bigger the knowledge gaps are, the more I procrastinate on real work.<p>Talk about back to school feelings (!)
    • mikkupikku2 hours ago
      Just w.r.t having time to think harder, have you considered getting a hobby that forces you to go offline and do something repetitive so your mind can wander? I do this with walks (phone left at home) and sometimes swimming laps. Physical exercise may not seem appealing if you&#x27;re in burnout territory, but I think it&#x27;s worth trying because for me at least it&#x27;s a different, mostly orthogonal, kind of fatigue.
  • mpazik10 hours ago
    Happy to see it gets attention. The friction shows up once you mix docs with structured things like work items or ADRs. Flat markdown doesn&#x27;t query well and gets inconsistent. You can read TASKS.md fine. The agent can&#x27;t ask &quot;show me open tasks blocking this epic&quot; without scanning prose or maintaining a parallel index.<p>The AGENTS.md approach papers over this by teaching the LLM the folder conventions. Works until the data gets complex but gets worse after many iterations.<p>Both are needed: files that open in any editor, and a structured interface the agent can actually query. Been building toward that with Binder (github.com&#x2F;mpazik&#x2F;binder), a local knowledge platform. Data lives in a structured DB but renders to plain markdown with bi-directional sync. LSP gives editors autocomplete and validation. Agents and scripts get the same data through CLI or MCP.
  • gbro3n5 hours ago
    I built AS Notes for VS Code (<a href="https:&#x2F;&#x2F;www.asnotes.io" rel="nofollow">https:&#x2F;&#x2F;www.asnotes.io</a>) with the option for this usage pattern in mind. By augmenting VS Code so it has the tooling we use in personal knowledge management systems, it makes it easy to write, link and update markdown &#x2F; wikilinked notes manually (with mermaid &#x2F; LaTeX rendering capability also) - but by using VS Code we have easy access to an Agent harness that we can direct to work on, or use our notes as context. Others have pointed out that context bloat is an issue, but no more so than when you use the copilot harness (or any other) inside a large codebase. I find I get more value from my AI conversations when I persist the outputs in markdown like this.
  • LeonTing10106 hours ago
    Related to this — I&#x27;ve been experimenting with the idea that AI&#x27;s output should be deterministic programs, not more AI calls. An &quot;idea file&quot; for me is: what&#x27;s the minimal interface between the AI (creative, expensive) and the runtime (deterministic, free)? Turns out 8 core operations are enough to cover any browser interface. The AI observes once, crystallizes a program, then the program runs forever at zero cost. The hard part isn&#x27;t the runtime — it&#x27;s building a good enough feedback loop so the AI can tell whether its program actually works.
  • ractive8 hours ago
    I built a tool to exactly help an LLM navigate and search a knowledgebase of md files. It helps a lot to build such a wiki by providing basic content search à la grep but also structured search for frontmatter properties. It also helps to move files around without breaking links and to fix links automatically. It is a CLI tool, mainly meant to be driven by AI tools.<p>Check it out: <a href="https:&#x2F;&#x2F;github.com&#x2F;ractive&#x2F;hyalo" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;ractive&#x2F;hyalo</a>
  • gchamonlive13 hours ago
    I don&#x27;t think this is taking it as far as it can go.<p>Everything should live in the repo. Code and docs yes. But also the planning files, epics, work items, architectural documentation and decisions. Here is a small example of my Linux system doc: <a href="https:&#x2F;&#x2F;github.com&#x2F;gchamon&#x2F;archie&#x2F;tree&#x2F;main&#x2F;docs" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;gchamon&#x2F;archie&#x2F;tree&#x2F;main&#x2F;docs</a><p>And you don&#x27;t need to reinvent the wheel. Code docs can like either right next to it in the readme or in docs&#x2F; if it&#x27;s too big for a single file or the context spams multiple modules. ADRs live in docs&#x2F;architecture&#x2F;decisions. Epics and Workitems can also live in the docs.<p>Everything is for agents and everything is for humans, unless put in AGENTS.md and docs&#x2F;agents or something similar, and even those are for human too.<p>In a nutshell, put everything in the repo, reuse standards as much as possible, the idea being it&#x27;s likely the structure is already embedded in the model, and always review documentation changes.
    • locknitpicker10 hours ago
      &gt; Everything should live in the repo. Code and docs yes. But also the planning files, epics, work items, architectural documentation and decisions.<p>You just described spec-driven development.
      • gchamonlive3 hours ago
        Yep, and don&#x27;t reinvent the wheel
  • foo428 hours ago
    I built a very similar system into my own assistant type project. In all honesty though I&#x27;ve not used it enough to know how well it works out in practice.
  • estetlinus8 hours ago
    Sounds like a solution in search for a problem.
  • atbpaca15 hours ago
    An LLM that maintains a Confluence space. That looks like an interesting idea!
  • jdthedisciple9 hours ago
    The challenge to me seems quality assurance:<p>I&#x27;d rather have it source the <i>original</i> document everytime, then an LLM-generated wiki which I most likely wouldn&#x27;t have the time to fact-check and review myself.
  • ansc2 hours ago
    The comments in the gist is depressing.
  • mbreese15 hours ago
    I’ve been doing something similar with a RAG system where in addition to storing the documents, we use an LLM to pull out “facts”. We’re using the LLM to look for relationships between different entities. This is then also returned when we query the database.<p>But I like the idea of an LLM generated&#x2F;maintained wiki. That might be a useful addition to allow for more interactive exploration of a document database.
  • serendipty019 hours ago
    Similar: <a href="https:&#x2F;&#x2F;v-it.org&#x2F;doctrine&#x2F;" rel="nofollow">https:&#x2F;&#x2F;v-it.org&#x2F;doctrine&#x2F;</a>
  • Lockal9 hours ago
    This thing already exists for multiple years - see <a href="https:&#x2F;&#x2F;deepwiki.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;deepwiki.com&#x2F;</a> (99% it is autonomous, but it can be manually structured - see <a href="https:&#x2F;&#x2F;docs.devin.ai&#x2F;work-with-devin&#x2F;deepwiki#steering-deepwiki" rel="nofollow">https:&#x2F;&#x2F;docs.devin.ai&#x2F;work-with-devin&#x2F;deepwiki#steering-deep...</a>). There were also multiple attempts to replicate it with local LLMs.<p>The problem is that it is still a slop: not only it adds a lot of noise (&quot;architecture&quot; diagrams based on some cherry-picked filenames, incomplete datatables, hyperfocusing on strange things), it also hallucinates, adding factually incorrect information (while direct questions to LLM shows correct information).
  • argee14 hours ago
    This is what Semiont is trying to do, to some extent [0].<p>Doesn&#x27;t really feel that useful in practice.<p>[0] <a href="https:&#x2F;&#x2F;github.com&#x2F;The-AI-Alliance&#x2F;semiont" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;The-AI-Alliance&#x2F;semiont</a>
  • mememememememo16 hours ago
    This sounds like compaction for RAG.
  • 0123456789ABCDE11 hours ago
    this is so validating•<p><a href="https:&#x2F;&#x2F;grimoire-pt5.sprites.app&#x2F;" rel="nofollow">https:&#x2F;&#x2F;grimoire-pt5.sprites.app&#x2F;</a>
  • voidhorse11 hours ago
    This makes me feel like karpathy is behind on the times a tad. Many agent users I know already do precisely this as part of &quot;agentic&quot; development. If you use a harness, the harness is already empowered to do much of this under the hood, no fancy instruction file required. Just ask the agent to update some knowledge directory at the end of each convo, done. If you really need to automate it, write some scheduling tool that tells the agent to read past convos and summarize. It really is that easy.
  • qaadika16 hours ago
    &gt; You never (or rarely) write the wiki yourself — the LLM writes and maintains all of it. You&#x27;re in charge of sourcing, exploration, and asking the right questions. The LLM does all the grunt work — the summarizing, cross-referencing, filing, and bookkeeping that makes a knowledge base actually useful over time.<p>I&#x27;m not sure how you can get any closer to &quot;turning your thinking over to machines.&quot; These tasks may be &quot;grunt work,&quot; but it&#x27;s while doing these things that new ideas pop in, or you decide on a particular or novel way to organize or frame information. Many of my insights in my (analog? vanilla? my human-written) Obsidian vault (that I consider my &quot;personal wiki&quot;) have been made or expanded on because I happened to see one note after another in doing the &quot;grunt work&quot;, or just by opening one note and seeing its title right beside a previously forgotten one.<p>There&#x27;s nothing &quot;personal&quot; about a knowledge base you filled by asking AI questions. It&#x27;s the AI&#x27;s database, you just ask it to write stuff. Learn how to learn and answer your own damn questions.<p>Soon pedagogy will be a piece of paper that says &quot;Ask AI.&quot;<p>I hate this idea that a result is all that matters, and the quicker you can get the result the better, at any cost (mental or financial, short-term or long-term).<p>If we optimized showers to be 20 seconds, we&#x27;d stop having shower thoughts. I like my shower thoughts. And so too my grunt-work thoughts.<p>---<p>As an aside, I&#x27;m not <i>totally</i> against AI writing in a personal knowledgebase. I include it at times in my own. But since I started my current obsidian vault in 2023 (now 4100 self-written notes, including maybe up to 5% Web Clipper notes), I&#x27;ve had a Templater (Obsidian plugin) template I wrap around anything AI-written to &#x27;quarantine&#x27; it from my own words:<p>==BEGIN AI-GENERATED CONTENT==<p>&lt;% tp.file.cursor(1) %&gt;<p>==END AI-GENERATED CONTENT==<p>I&#x27;ve used this consistently and it&#x27;s helped me keep (and develop) my own writing voice apart from any of my AI usage. It actually motivates me to write <i>more</i>, because I know I could always take the easy route and chunk whatever I&#x27;m thinking into the AI, but I&#x27;m choosing not to by writing it myself, with my own vocabulary, in my own voice, with my own framing. I trick myself into writing because my pride tells me I can express my knowledge better than the AI can.<p>I also manually copy and paste from wherever I&#x27;m using AI into my notes. Nothing automated. The friction keeps me from sliding into the happy path of turning my brain off.
    • mold_aid15 hours ago
      Since you&#x27;re a fellow Obsidian user, you likely remember the early days of back-linking note-taking software like Roam and such. I remember just seeing pictures of the graph being the primary visual symbol representing the depth of learning. I thought &quot;ok well people just want to accumulate stuff.&quot; AI tools certainly help with creating a mass of notes.<p>There&#x27;s a comment above how this is reminiscent of Licklider&#x27;s work, but it reminds of the early print culture era, where books were a consumer item, and people just purchased a lot of them to put on shelves built to display them.
      • qaadika15 hours ago
        I actually never got into note-taking before I found Obsidian. I used Google Drive all throughout college and up to 2023, so any knowledge I had written down was sequestered by an ad-hoc folder structure that was mostly chronological by year, or in my physical notebooks by subject. It was also limited by what I felt was worth writing down enough that it merited a Doc, or what I could write in one session before getting distracted and never touching it again. And mentally I limited myself by always wanting to write something down &quot;right&quot;, all spell-checked and grammatically correct and sensically organized, which led to often not writing anything down at all. Now I dump the words down and come back to it later when I want to &quot;garden.&quot;<p>My brother showed me stuff like Trilium circa 2017, but I hadn&#x27;t the thinking process I do now to even know where to start racking my brain for stuff to write down.<p>When I read Chernow&#x27;s biography of Alexander Hamilton, I was in awe of his ability to write so damn much. I never thought I&#x27;d be able to do that. Turns out the secret is just three things: have stuff you&#x27;re passionate about, be able to recall it, and write so damn much about it. When your thinking process is based around &quot;how would I phrase&#x2F;word&#x2F;frame this to write it down,&quot; if you have the right process and organization to be able to access it later, it&#x27;s even easier than talking. For some they can keep everything they know in their head. I&#x27;m not one of them. but I can write everything down, and in writing it down I end up remembering it better. My professors were right all along.<p>And if one looks at the actual written letters people like Hamilton [1] and co. wrote, or even back to Isaaci Newtoni [2][3], They&#x27;re riddled with spelling errors and strange latinesque grammar and informal formalities. Yet they&#x27;re revered for their ability to write. Because it really is the thought that countts [4], not the words.<p>(Very little of these comments is new thoughts I&#x27;m having now. Most of it is thoughts I had and documented when I was super into PKM in 2023 and since, and now comes back up as those neurons fire again and I consider the new idea of &quot;should AI <i>be</i> my PKM?&quot; after reading the post.)<p>---<p>Yeah, the graphs are cool for a little bit. But only post-facto, once one has an amount of data points where it might become useful. If the AI is doing the organization then any personal significance is lost. Or rather it was never there to begin with.<p>Wikilinks is the feature I use most often, outside of my folder organization (PARA). Now when I have a thought, it goes down a chain of &quot;do I have a note already this can go into,&quot; to &quot;no, it should be a new note. are there any notes I should wikilink in this one, or link this one in?&quot;<p>I think I made a good decision early on when I was inspired by the Emacs documentation to add a basic &quot;Related: &quot; line before the first header (and after the YAML). There I dump any wikilink I think <i>might</i> possibly be something I want to reference, or find this note via a backlink, without having to think about where to put it in the body.<p>E.g.<p>{YAML header}<p>Related: [[Artemis]], [[Artemis II]], [[NASA Engineering]], [[Space MOC]], [[NASA CAPCOM]],<p># Artemis II Mission Timeline and Notes<p>{body and rest of note, my own record of things that happened as I watch the stream}<p>---<p>&gt; AI tools certainly help with creating a mass of notes.<p>Agreed. Presuming the implication is it creates a mass of notes, but of generic information stated generically. I&#x27;m really proud of my 4100 notes, because I know (aside from a few catagories like web clipping) even if they&#x27;re a mess, they&#x27;re <i>my</i> mess. I definitely could have gone the last three years without having found Obsidian, but I wouldn&#x27;t have as clear a record of them as I do now. Or the rest of my life, as I slowly add stuff about my past, or migrate old writing into it. I also definitely repeat myself by saying the same information in different places, but in different ways. It&#x27;s not &#x27;efficient&#x27; information density-wise, but it is designed for a human to read and see the human behind the writing.<p>I also believe I think clearer, as often when I&#x27;m recalling information I&#x27;m actually recalling my note in my head on that subject. I write so much that in conversation &quot;I was thinking the other day&quot; is analogous to &quot;I wrote down in my notes&quot;.<p>I might be crazy but I would put my vault in my will as something to be passed on, because there&#x27;s so much <i>me</i> in them. My yearly journals in &#x2F;02 Areas&#x2F;Journals&#x2F; are the most obvious ones, but I have a &#x2F;02 Areas&#x2F;Writing&#x2F; folder that&#x27;s just notes I consider &quot;writing&quot;, whic is distincy from the contents of &#x2F;03 Resources&#x2F; folder that&#x27;s the &quot;general knowledge&quot; knowledgebase.<p>---<p>Anyway, I guess my tl;dr is that AI can never write about thinking as well as a human can, and in my opinion it&#x27;s the thinking that important, not the writing. the writing or the words is merely a tool in thinking. Karpathy mistakes the words to be the goal, rather than the thinking that caused the words.<p>---<p>One last thing: I just re-read the HN guidelines out of curiosity, and I noticed they recently added &quot;Don&#x27;t post generated comments or AI-edited comments.&quot;<p>I could copy and paste almost anything from my vault into an HN comment without violating this rule. Anybody creating a PKM with this sytem could not. They would have to rewrite it in their own words. So one might as well just right it themself in the first place if they ever think they might want to reuse it in a place like HN.<p>---<p>[1] <a href="https:&#x2F;&#x2F;outhistory.org&#x2F;exhibits&#x2F;show&#x2F;rev&#x2F;hamilton-laurens-letters" rel="nofollow">https:&#x2F;&#x2F;outhistory.org&#x2F;exhibits&#x2F;show&#x2F;rev&#x2F;hamilton-laurens-le...</a><p>[2] <a href="https:&#x2F;&#x2F;www.newtonproject.ox.ac.uk&#x2F;view&#x2F;texts&#x2F;diplomatic&#x2F;MINT01748" rel="nofollow">https:&#x2F;&#x2F;www.newtonproject.ox.ac.uk&#x2F;view&#x2F;texts&#x2F;diplomatic&#x2F;MIN...</a><p>[3] A while back, while in a Newton phase, I decided arbitrarily to refer to him as &quot;Isaaci Newtoni,&quot; as that&#x27;s how he called himself. I reinforced that by using that name for him in my notes. Now I call him that instinctually, not consciously.<p>[4] Intentional.
  • cyanydeez17 hours ago
    Too much context pollution.<p>Start with short text context, and flow through DAGs via choose your own adventure. We alreadybreached context limits. Nows the time to let LLMs build their contexts through decision trees and prune dead ends.
    • j-pb16 hours ago
      In my experience a wiki can actually drastically reduce the amount of dead context.<p>I&#x27;ve handed my local agents a bunch of integrated command line tools (kinda like an office suite for LLMs), including a wiki (<a href="https:&#x2F;&#x2F;github.com&#x2F;triblespace&#x2F;playground&#x2F;blob&#x2F;main&#x2F;faculties&#x2F;wiki.rs" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;triblespace&#x2F;playground&#x2F;blob&#x2F;main&#x2F;facultie...</a> ) and linkage really helps drastically reduce context bloat because they can pull in fragment by fragment incrementally.
      • cyanydeez15 hours ago
        Was also thinking to disambiguate context where you wish to express a tokens function (eg, top) as different from one could use unique ASCII prefix (eg, ∆top) to avoid pollution between the english and the linux binary.<p>Youd then alias these disambiguated terms and theyd still trigger the correct token autocomplete but would reduce overlap which cause misdirection.
  • kmaitreys7 hours ago
    &gt; You never (or rarely) write the wiki yourself — the LLM writes and maintains all of it.<p>Then what is the point? Why be averse so to use your own brain so much? Why are tech bros like this?
  • cookiengineer5 hours ago
    This is essentially the hack that Claude Code did with their dedicated sub-agent that summarizes a list of previous messages and maintains different types of memory (temporary, persistent, cross-agent). They also built tools so that agents can talk to each other and use that summary to propagate knowledge to other agents if needed.<p>Setting aside that their codebase is absolute slopcrap, I think something like this might work nicely if it&#x27;s built from the ground up.<p>For my own test environment I&#x27;m relying on Golang and its conventions (go build, go test, go fmt, gopls etc) which saves a lot of prompts and tokens down the line. Additionally I think that spec driven development might be more successful but I haven&#x27;t found out yet what the right amount of details for specifications is, so that semantic anchors can help summarize it better.<p>Anyways if you&#x27;re curious, it&#x27;s made for short agent lifecycles and it kinda works every time most of the time: <a href="https:&#x2F;&#x2F;github.com&#x2F;cookiengineer&#x2F;exocomp" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;cookiengineer&#x2F;exocomp</a><p>Still need to implement the summarizing agent and memory parts, it&#x27;s a little fiddlework to get that right so I&#x27;m currently experimenting a lot locally with both ollama and vllm as an inference engine.
  • nurettin10 hours ago
    He really wants to shine, but how is this different than claude memory or skills? When I encounter something it had difficulty doing, or consistently start off with incorrect assumptions, I solve for it and tell it to remember this. If it goes on a long trial and error loop to accomplish something, once it works I tell it to create a skill.
    • locknitpicker10 hours ago
      &gt; He really wants to shine, but how is this different than claude memory or skills?<p>It isn&#x27;t different. This just tries to reinvent the wheel that all mainstream coding assistants have been providing for over a year.<p>Even ChatGPT rolled out chat memory in their free tier.
      • TeMPOraL6 hours ago
        The difference obviously being, his way <i>you</i> own the memories; in what&#x27;s currently deployed, it&#x27;s the platform that owns them.
        • nurettin2 hours ago
          Skills and memory are .md files on your system.
        • locknitpicker3 hours ago
          &gt; The difference obviously being, his way you own the memories; in what&#x27;s currently deployed, it&#x27;s the platform that owns them.<p>I don&#x27;t think you are very familiar with the state of LLM assistants. Prompt files were a thing since ever, and all mainstream coding assistants support storing your personal notes in your own files in multiple ways. Prompt files, instruction files, AGENTS.md, etc.
  • MarcelinoGMX3C2 hours ago
    [dead]
  • bambushu2 hours ago
    [dead]
  • meidad_g10 hours ago
    [dead]