Archives AI News

FedHFT: Efficient Federated Finetuning with Heterogeneous Edge Clients

arXiv:2510.14054v1 Announce Type: new Abstract: Fine-tuning pre-trained large language models (LLMs) has become a common practice for personalized natural language understanding (NLU) applications on downstream tasks and domain-specific datasets. However, there are two main challenges: (i) limited and/or heterogeneous data…

LLM-guided Chemical Process Optimization with a Multi-Agent Approach

arXiv:2506.20921v2 Announce Type: replace Abstract: Chemical process optimization maximizes production efficiency and economic performance, but optimization algorithms, including gradient-based solvers, numerical methods, and parameter grid searches, become impractical when operating constraints are ill-defined or unavailable. We present a multi-agent LLM…

Beyond Linear Probes: Dynamic Safety Monitoring for Language Models

arXiv:2509.26238v2 Announce Type: replace Abstract: Monitoring large language models’ (LLMs) activations is an effective way to detect harmful requests before they lead to unsafe outputs. However, traditional safety monitors often require the same amount of compute for every query. This…

Weight Weaving: Parameter Pooling for Data-Free Model Merging

arXiv:2510.13921v1 Announce Type: new Abstract: Model merging provides a cost-effective and data-efficient combination of specialized deep neural networks through parameter integration. This technique leverages expert models across downstream tasks without requiring retraining. Most model merging approaches critically depend on scaling…

Symmetry-Aware GFlowNets

arXiv:2506.02685v3 Announce Type: replace-cross Abstract: Generative Flow Networks (GFlowNets) offer a powerful framework for sampling graphs in proportion to their rewards. However, existing approaches suffer from systematic biases due to inaccuracies in state transition probability computations. These biases, rooted in…