Archives AI News

Sculpting Latent Spaces With MMD: Disentanglement With Programmable Priors

arXiv:2510.11953v1 Announce Type: new Abstract: Learning disentangled representations, where distinct factors of variation are captured by independent latent variables, is a central goal in machine learning. The dominant approach has been the Variational Autoencoder (VAE) framework, which uses a Kullback-Leibler…

Y-shaped Generative Flows

arXiv:2510.11955v1 Announce Type: new Abstract: Modern continuous-time generative models often induce V-shaped transport: each sample travels independently along nearly straight trajectories from prior to data, overlooking shared structure. We introduce Y-shaped generative flows, which move probability mass together along shared…

LayerSync: Self-aligning Intermediate Layers

arXiv:2510.12581v1 Announce Type: cross Abstract: We propose LayerSync, a domain-agnostic approach for improving the generation quality and the training efficiency of diffusion models. Prior studies have highlighted the connection between the quality of generation and the representations learned by diffusion…

QLENS: Towards A Quantum Perspective of Language Transformers

arXiv:2510.11963v1 Announce Type: new Abstract: In natural language processing, current methods for understanding Transformers are successful at identifying intermediate predictions during a model’s inference. However, these approaches function as limited diagnostic checkpoints, lacking a mathematical framework for mechanistically modeling how…

WW-FL: Secure and Private Large-Scale Federated Learning

arXiv:2302.09904v4 Announce Type: replace Abstract: Federated learning (FL) is an efficient approach for large-scale distributed machine learning that promises data privacy by keeping training data on client devices. However, recent research has uncovered vulnerabilities in FL, impacting both security and…

Learning Dynamics of VLM Finetuning

arXiv:2510.11978v1 Announce Type: new Abstract: Preference-based finetuning of vision–language models (VLMs) is brittle: trivially wrong negatives inject uninformative gradients that destabilize training. We recast alignment as textbf{learning-dynamics–aware optimization} and introduce textbf{Cooling-Weighted DPO (CW-DPO)}, a two-stage recipe that explicitly models and…

Toward Fair Graph Neural Networks Via Dual-Teacher Knowledge Distillation

arXiv:2412.00382v2 Announce Type: replace Abstract: Graph Neural Networks (GNNs) have demonstrated strong performance in graph representation learning across various real-world applications. However, they often produce biased predictions caused by sensitive attributes, such as religion or gender, an issue that has…