2 comments

  • westurner22 minutes ago
    How could this lend insight into why Fast Fourier Transform approximates self-attention?<p>&gt; <i>Because self-attention can be replaced with FFT for a loss in accuracy and a reduction in kWh [1], I suspect that the Quantum Fourier Transform can also be substituted for attention in LLMs.</i><p>[1] &quot;Fnet: Mixing tokens with fourier transforms&quot; (2021) <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2105.03824" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2105.03824</a> .. &quot;Google Replaces BERT Self-Attention with Fourier Transform: 92% Accuracy, 7 Times Faster on GPUs&quot; <a href="https:&#x2F;&#x2F;syncedreview.com&#x2F;2021&#x2F;05&#x2F;14&#x2F;deepmind-podracer-tpu-based-rl-frameworks-deliver-exceptional-performance-at-low-cost-19&#x2F;" rel="nofollow">https:&#x2F;&#x2F;syncedreview.com&#x2F;2021&#x2F;05&#x2F;14&#x2F;deepmind-podracer-tpu-ba...</a><p>&quot;Why formalize mathematics – more than catching errors&quot; (2025) <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=45695541">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=45695541</a><p>Can the QFT <i>Quantum</i> Fourier Transform (and IQFT Inverse Quantum Fourier Transform) also be substituted for self-attention in LLMs, and do Lean formalisms provide any insight into how or why?
  • measurablefunc2 hours ago
    I guess the next step would be adding support for quantized arithmetic.
    • pstoll1 hour ago
      And the lower precision float variants.