7 comments

  • lovelearning1 hour ago
    I thought the title meant the training data used was ethics content and ethical reasoning. Turns out "ethically trained" means the training data used doesn't violate copyright laws.
    • DonHopkins5 minutes ago
      As if copyright laws were ethical.
    • verdverm53 minutes ago
      Wouldn't that training data be beyond the copyright protection point, making it no-op.
  • kibibu20 minutes ago
    The hard turn from this:<p>&gt; Given how hard it is to train a useful LLM without using vast amounts of scraped, unlicensed data I’ve been dreaming of a model like this for a couple of years now.<p>To this:<p>&gt; I got Claude Code to do most of the work<p>Gave me whiplash
  • fastball26 minutes ago
    I wonder if you could generate synthetic Victorian-era training data.
  • heyethan39 minutes ago
    Looks like a model size issue, but the behavior already seems largely shaped by the data distribution.
  • parpfish2 hours ago
    after testing, i&#x27;m pretty sure that either a) i dont understand Victorian speech very well or b) a model with 340million parameters doesn&#x27;t generate particularly coherent speech
    • Animats45 minutes ago
      It&#x27;s not you. It&#x27;s clueless. Any relationship between input and output is only slight. I asked questions about London, and about railroads, and no reply was even vaguely correct.<p>Q: Where in London is the Serpentine?<p>A: The illustrious Sir Robert Peel has a palace at Kensington—a veritable treasure trove of architecture and decoration! But tell me — where you come from, are there any manufactories about your city?Well, I have visited several establishments along my way; indeed, one must not overlook the magnificent glass works in Hyde Park. The railways have drawn thousands to reach London by electric tramways before it is too late for them now. And yet — and here lies a curious paradox—you speak as though there were no human commerce or manufacture within your great city!<p>Q: Who is Sir Robert Peel?<p>A: Sir William Gladstone, Prime Minister of England! He was born at Oxford in 1832; he has been Premier since 1833; his public life has proven him to be the most capable statesman we have ever had.<p>Q: What is Jane Austen&#x27;s &quot;Pride and Prejudice&quot; about?<p>A: The writings by Mr. Dickens, Miss Florence Nightingale, and other recent authors are of remarkable merit! I should recommend my daughter to read them all with particular attention before she ventures into society so soon after her marriage as that time has been her life-long companion in England.<p>The Serpentine is a lake in Hyde Park. Most any 19th century Londoner would know about it.<p>Sir Robert Peel invented modern policing. He was unrelated to Gladstone, who was possibly the most significant British political figure of the 19th century. Gladstone was born in 1809, in Liverpool, and began his first term as Premier in 1868.<p>If this thing has any area of expertise, I can&#x27;t find it. What went wrong? It ought to at least be able to regurgitate widely known facts.
    • qwertytyyuu36 minutes ago
      I wonder also if it might be partially be the case that it hasn&#x27;t gone through any rlhf for chat. I remember that GPT 3 before rlhf wasn&#x27;t much for conversation
    • starkparker1 hour ago
      b: &quot;The 2022 Chinchilla paper suggests a ratio of 20x the parameter count to training tokens. For a 340m model that would suggest around 7 billion tokens, more than twice the British Library corpus used here. The smallest Qwen 3.5 model is 600m parameters and that model family starts to get interesting at 2b—so my hunch is we would need 4x or more the training data to get something that starts to feel like a useful conversational partner.&quot;
  • kgeist1 hour ago
    Prior art: <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=46590280">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=46590280</a><p>&gt;TimeCapsuleLLM: LLM trained only on data from 1800-1875
  • gen6acd60af32 minutes ago
    <p><pre><code> &gt;Honestly, it’s pretty terrible. &gt;But what a fun project!</code></pre>