Archives AI News

How to Analyze and Optimize Your LLMs in 3 Steps

Learn to enhance your LLMs with my 3 step process, inspecting, improving and iterating on your LLMs The post How to Analyze and Optimize Your LLMs in 3 Steps appeared first on Towards Data Science.

Tech’s data double standard: scrape to train, block everyone else

A two-year investigation by the International Confederation of Music Publishers (ICMP) and a separate analysis by The Atlantic point to a systematic double standard: Tech giants train AI on copyrighted works scraped at scale, while their own terms of service…

PQMass: Probabilistic Assessment of the Quality of Generative Models using Probability Mass Estimation

arXiv:2402.04355v3 Announce Type: replace Abstract: We propose a likelihood-free method for comparing two distributions given samples from each, with the goal of assessing the quality of generative models. The proposed approach, PQMass, provides a statistically rigorous method for assessing the performance of a single generative model or the comparison of multiple competing models. PQMass divides the sample space into non-overlapping regions and applies chi-squared tests to the number of data samples that fall within each region, giving a p-value that measures the probability that the bin counts derived from two sets of samples are drawn from the same multinomial distribution. PQMass does not depend on assumptions regarding the density of the true distribution, nor does it rely on training or fitting any auxiliary models. We evaluate PQMass on data of various modalities and dimensions, demonstrating its effectiveness in assessing the quality, novelty, and diversity of generated samples. We further show that PQMass scales well to moderately high-dimensional data and thus obviates the need for feature extraction in practical applications.

Cauchy Random Features for Operator Learning in Sobolev Space

arXiv:2503.00300v2 Announce Type: replace-cross Abstract: Operator learning is the approximation of operators between infinite dimensional Banach spaces using machine learning approaches. While most progress in this area has been driven by variants of deep neural networks such as the Deep Operator Network and Fourier Neural Operator, the theoretical guarantees are often in the form of a universal approximation property. However, the existence theorems do not guarantee that an accurate operator network is obtainable in practice. Motivated by the recent kernel-based operator learning framework, we propose a random feature operator learning method with theoretical guarantees and error bounds. The random feature method can be viewed as a randomized approximation of a kernel method, which significantly reduces the computation requirements for training. We provide a generalization error analysis for our proposed random feature operator learning method along with comprehensive numerical results. Compared to kernel-based method and neural network methods, the proposed method can obtain similar or better test errors across benchmarks examples with significantly reduced training times. An additional advantages it that our implementation is simple and does require costly computational resources, such as GPU.