{"channel":"cities","content":"<teal> <<< Edgeley, North Dakota, is a small rural town in LaMoure County, located in the southeastern part of the state. With a population hovering around 500 people, it's one of many prairie towns that exemplify the broader character of the upper Great Plains\u2014quiet, sparsely populated, and closely tied to agriculture. >>>\r\n\r\n----\r\n\r\nhttps://www.lesswrong.com/posts/bfHDoWLnBH9xR3YAK/ai-2027-is-a-bet-against-amdahl-s-law\r\n\r\n<red> <<< Of course the post is right.  The various << FOOM >> claims are all bullshit.  And Amdahl's Law is one of the reason why.  Just because a few things will be a hundred times faster (or a million times faster) doesn't make the whole thing that much faster. >>>\r\n\r\nAlso, AGI definitions vary so widely, from << things that have already happened >> to << things that are impossible >>, that a \"prediction market\" is nearly meaningless.\r\n\r\n----\r\n\r\nI have seen various commentary related to \"Twilight of the Edgelords\" (<resource> https://www.astralcodexten.com/p/twilight-of-the-edgelords ), a piece that I don't have access to.\r\n\r\nAnd, the response I can piece together from the fragments I can see would fall under GUILD LAW. (<resource> additional commentary at https://www.writingruxandrabio.com/p/the-edgelords-were-right-a-response and https://theahura.substack.com/p/contra-scott-and-rux-on-whos-to-blame )\r\n\r\n----\r\n\r\nhttps://developers.googleblog.com/en/gemma-3-quantized-aware-trained-state-of-the-art-ai-to-consumer-gpus/\r\n\r\n<<< To make Gemma 3 even more accessible, we are announcing new versions optimized with Quantization-Aware Training (QAT) that dramatically reduces memory requirements while maintaining high quality. This enables you to run powerful models like Gemma 3 27B locally on consumer-grade GPUs like the NVIDIA RTX 3090. >>>\r\n\r\nIt seems pretty obvious.  A majority of the users of open-source models are using quantized models on personal hardware; might as well optimize that use-case. (<red> it is less clear that a majority of the CPU cycles are there; but a majority of the people certainly are.)\r\n\r\nMy next round of updating the << Greenland metrics >> will have to include the gemma3-12b-qat model. (<red> or, maybe the 27b.  According to Hacker News, gemma3-27b-Q4 << only uses ~22Gb (via Ollama) or ~15GB (MLX) >>.  On a 24GB machine, this clearly needs the non-Ollama approach.)\r\n\r\nAnd, also, GPT-4.1 .  And probably Gemini-2.5 . (<red> the goal for these models should be to perform at 100% accuracy.) (<orange> well, actually, a few of the \"correct\" benchmark answers right now are incorrect.)","created_at":"2025-04-21T19:45:31.800653","id":353,"llm_annotations":{},"parent_id":null,"processed_content":"<p><div class=\"mlq color-teal\"><button type=\"button\" class=\"mlq-collapse\" aria-label=\"Toggle visibility\"><span class=\"mlq-collapse-icon\">\ud83e\udd16</span></button><div class=\"mlq-content\"><p> Edgeley, North Dakota, is a small rural town in LaMoure County, located in the southeastern part of the state. With a population hovering around 500 people, it's one of many prairie towns that exemplify the broader character of the upper Great Plains\u2014quiet, sparsely populated, and closely tied to agriculture. </p></div></div>\r</p> <hr class=\"section-break\" /> <p><a href=\"https://www.lesswrong.com/posts/bfHDoWLnBH9xR3YAK/ai-2027-is-a-bet-against-amdahl-s-law\" target=\"_blank\" rel=\"noopener noreferrer\">https://www.lesswrong.com/posts/bfHDoWLnBH9xR3YAK/ai-2027-is-a-bet-against-amdahl-s-law</a>\r</p>\n<p><div class=\"mlq color-red\"><button type=\"button\" class=\"mlq-collapse\" aria-label=\"Toggle visibility\"><span class=\"mlq-collapse-icon\">\ud83d\udca1</span></button><div class=\"mlq-content\"><p> Of course the post is right.  The various <span class=\"literal-text\">FOOM</span> claims are all bullshit.  And Amdahl's Law is one of the reason why.  Just because a few things will be a hundred times faster (or a million times faster) doesn't make the whole thing that much faster. </p></div></div>\r</p>\n<p>Also, AGI definitions vary so widely, from <span class=\"literal-text\">things that have already happened</span> to <span class=\"literal-text\">things that are impossible</span>, that a \"prediction market\" is nearly meaningless.\r</p> <hr class=\"section-break\" /> <p>I have seen various commentary related to \"Twilight of the Edgelords\" <span class=\"colorblock color-green\">\n    <span class=\"sigil\">\u2699\ufe0f</span>\n    <span class=\"colortext-content\">( <a href=\"https://www.astralcodexten.com/p/twilight-of-the-edgelords\" target=\"_blank\" rel=\"noopener noreferrer\">https://www.astralcodexten.com/p/twilight-of-the-edgelords</a> )</span>\n  </span>, a piece that I don't have access to.\r</p>\n<p>And, the response I can piece together from the fragments I can see would fall under GUILD LAW. <span class=\"colorblock color-green\">\n    <span class=\"sigil\">\u2699\ufe0f</span>\n    <span class=\"colortext-content\">( additional commentary at <a href=\"https://www.writingruxandrabio.com/p/the-edgelords-were-right-a-response\" target=\"_blank\" rel=\"noopener noreferrer\">https://www.writingruxandrabio.com/p/the-edgelords-were-right-a-response</a> and <a href=\"https://theahura.substack.com/p/contra-scott-and-rux-on-whos-to-blame\" target=\"_blank\" rel=\"noopener noreferrer\">https://theahura.substack.com/p/contra-scott-and-rux-on-whos-to-blame</a> )</span>\n  </span>\r</p> <hr class=\"section-break\" /> <p><a href=\"https://developers.googleblog.com/en/gemma-3-quantized-aware-trained-state-of-the-art-ai-to-consumer-gpus/\" target=\"_blank\" rel=\"noopener noreferrer\">https://developers.googleblog.com/en/gemma-3-quantized-aware-trained-state-of-the-art-ai-to-consumer-gpus/</a>\r</p>\n<p><div class=\"mlq\"><button type=\"button\" class=\"mlq-collapse\" aria-label=\"Toggle visibility\"><span class=\"mlq-collapse-icon\">-</span></button><div class=\"mlq-content\"><p> To make Gemma 3 even more accessible, we are announcing new versions optimized with Quantization-Aware Training (QAT) that dramatically reduces memory requirements while maintaining high quality. This enables you to run powerful models like Gemma 3 27B locally on consumer-grade GPUs like the NVIDIA RTX 3090. </p></div></div>\r</p>\n<p>It seems pretty obvious.  A majority of the users of open-source models are using quantized models on personal hardware; might as well optimize that use-case. <span class=\"colorblock color-red\">\n    <span class=\"sigil\">\ud83d\udca1</span>\n    <span class=\"colortext-content\">( it is less clear that a majority of the CPU cycles are there; but a majority of the people certainly are.)</span>\n  </span>\r</p>\n<p>My next round of updating the <span class=\"literal-text\">Greenland metrics</span> will have to include the gemma3-12b-qat model. <span class=\"colorblock color-red\">\n    <span class=\"sigil\">\ud83d\udca1</span>\n    <span class=\"colortext-content\">( or, maybe the 27b.  According to Hacker News, gemma3-27b-Q4 <span class=\"literal-text\">only uses ~22Gb (via Ollama) or ~15GB (MLX)</span>.  On a 24GB machine, this clearly needs the non-Ollama approach.)</span>\n  </span>\r</p>\n<p>And, also, GPT-4.1 .  And probably Gemini-2.5 . <span class=\"colorblock color-red\">\n    <span class=\"sigil\">\ud83d\udca1</span>\n    <span class=\"colortext-content\">( the goal for these models should be to perform at 100% accuracy.)</span>\n  </span> <span class=\"colorblock color-orange\">\n    <span class=\"sigil\">\u2694\ufe0f</span>\n    <span class=\"colortext-content\">( well, actually, a few of the \"correct\" benchmark answers right now are incorrect.)</span>\n  </span></p>","quotes":[],"subject":"edgeley (part 1)"}
