Don’t be lazy: CompleteP enables compute-efficient deep transformers
arXiv:2505.01618v3 Announce Type: replace Abstract: We study compute efficiency of LLM training when using different parameterizations, i.e., rules for adjusting model and optimizer hyperparameters (HPs) as model size changes. Some parameterizations fail to transfer optimal base HPs (such as learning…
