New Encoders for German Trained from Scratch: Comparing ModernGBERT with Converted LLM2Vec Models
arXiv:2505.13136v2 Announce Type: replace-cross Abstract: Encoders remain essential for efficient German NLP and NLU scenarios despite the rise of decoder-only LLMs. This work studies two routes to high-quality German encoders under identical data and training constraints: 1) training from scratch…
