Superpositional Gradient Descent: Harnessing Quantum Principles for Model Training
arXiv:2511.01918v1 Announce Type: new Abstract: Large language models (LLMs) are increasingly trained with classical optimization techniques like AdamW to improve convergence and generalization. However, the mechanisms by which quantum-inspired methods enhance classical training remain underexplored. We introduce Superpositional Gradient Descent…
