Archives AI News

CEMTM: Contextual Embedding-based Multimodal Topic Modeling

arXiv:2509.11465v2 Announce Type: replace-cross Abstract: We introduce CEMTM, a context-enhanced multimodal topic model designed to infer coherent and interpretable topic structures from both short and long documents containing text and images. CEMTM builds on fine-tuned large vision language models (LVLMs)…

PT$^2$-LLM: Post-Training Ternarization for Large Language Models

arXiv:2510.03267v1 Announce Type: new Abstract: Large Language Models (LLMs) have shown impressive capabilities across diverse tasks, but their large memory and compute demands hinder deployment. Ternarization has gained attention as a promising compression technique, delivering substantial size reduction and high…

COSMO-RL: Towards Trustworthy LMRMs via Joint Safety and Stability

arXiv:2510.04196v1 Announce Type: cross Abstract: Large Multimodal Reasoning Models (LMRMs) are moving into real applications, where they must be both useful and safe. Safety is especially challenging in multimodal settings: images and text can be combined to bypass guardrails, and…

General Exploratory Bonus for Optimistic Exploration in RLHF

arXiv:2510.03269v1 Announce Type: new Abstract: Optimistic exploration is central to improving sample efficiency in reinforcement learning with human feedback, yet existing exploratory bonus methods to incentivize exploration often fail to realize optimism. We provide a theoretical analysis showing that current…

CoDA: Coding LM via Diffusion Adaptation

arXiv:2510.03270v1 Announce Type: new Abstract: Diffusion language models promise bidirectional context and infilling capabilities that autoregressive coders lack, yet practical systems remain heavyweight. We introduce CoDA, a 1.7B-parameter diffusion coder trained on TPU with a fully open-source training pipeline. CoDA…