PTQTP: Post-Training Quantization to Trit-Planes for Large Language Models
arXiv:2509.16989v3 Announce Type: replace Abstract: Post-training quantization (PTQ) of large language models (LLMs) to extremely low bit-widths remains challenging due to the fundamental trade-off between computational efficiency and representational capacity. While existing ultra-low-bit methods rely on binary approximations or quantization-aware…
