MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
arXiv:2511.00056v1 Announce Type: new Abstract: The substantial memory demands of pre-training and fine-tuning large language models (LLMs) require memory-efficient optimization algorithms. One promising approach is layer-wise optimization, which treats each transformer block as a single layer and optimizes it sequentially,…
