Archives AI News

Data-Driven Temperature Modelling of Machine Tools by Neural Networks: A Benchmark

arXiv:2510.03261v1 Announce Type: new Abstract: Thermal errors in machine tools significantly impact machining precision and productivity. Traditional thermal error correction/compensation methods rely on measured temperature-deformation fields or on transfer functions. Most existing data-driven compensation strategies employ neural networks (NNs) to…

Safety Subspaces are Not Linearly Distinct: A Fine-Tuning Case Study

arXiv:2505.14185v2 Announce Type: replace Abstract: Large Language Models (LLMs) rely on safety alignment to produce socially acceptable responses. However, this behavior is known to be brittle: further fine-tuning, even on benign or lightly contaminated data, can degrade safety and reintroduce…

GUIDE: Towards Scalable Advising for Research Ideas

arXiv:2507.08870v2 Announce Type: replace Abstract: The field of AI research is advancing at an unprecedented pace, enabling automated hypothesis generation and experimental design across diverse domains such as biology, mathematics, and artificial intelligence. Despite these advancements, there remains a significant…

Memory Self-Regeneration: Uncovering Hidden Knowledge in Unlearned Models

arXiv:2510.03263v1 Announce Type: new Abstract: The impressive capability of modern text-to-image models to generate realistic visuals has come with a serious drawback: they can be misused to create harmful, deceptive or unlawful content. This has accelerated the push for machine…

Front-Loading Reasoning: The Synergy between Pretraining and Post-Training Data

arXiv:2510.03264v1 Announce Type: new Abstract: The prevailing paradigm for enhancing the reasoning abilities of LLMs revolves around post-training on high-quality, reasoning-intensive data. While emerging literature suggests that reasoning data is increasingly incorporated also during the mid-training stage-a practice that is…

Don’t Pay Attention, PLANT It: Pretraining Attention via Learning-to-Rank

arXiv:2410.23066v2 Announce Type: replace-cross Abstract: State-of-the-art Extreme Multi-Label Text Classification models rely on multi-label attention to focus on key tokens in input text, but learning good attention weights is challenging. We introduce PLANT – Pretrained and Leveraged Attention – a…

MindCraft: How Concept Trees Take Shape In Deep Models

arXiv:2510.03265v1 Announce Type: new Abstract: Large-scale foundation models demonstrate strong performance across language, vision, and reasoning tasks. However, how they internally structure and stabilize concepts remains elusive. Inspired by causal inference, we introduce the MindCraft framework built upon Concept Trees.…