Search:
Match:
3 results

Analysis

This paper introduces a novel sampling method, Schrödinger-Föllmer samplers (SFS), for generating samples from complex distributions, particularly multimodal ones. It improves upon existing SFS methods by incorporating a temperature parameter, which is crucial for sampling from multimodal distributions. The paper also provides a more refined error analysis, leading to an improved convergence rate compared to previous work. The gradient-free nature and applicability to the unit interval are key advantages over Langevin samplers.
Reference

The paper claims an enhanced convergence rate of order $\mathcal{O}(h)$ in the $L^2$-Wasserstein distance, significantly improving the existing order-half convergence.

GLUE: Gradient-free Expert Unification

Published:Dec 27, 2025 04:59
1 min read
ArXiv

Analysis

This paper addresses the challenge of combining multiple pre-trained specialist models for new target domains. It proposes a novel method, GLUE, that avoids the computational cost of full backpropagation by using a gradient-free optimization technique (SPSA) to learn the mixture coefficients of expert models. This is significant because it allows for efficient adaptation to new domains without requiring extensive training. The results demonstrate improved accuracy compared to baseline methods, highlighting the practical value of the approach.
Reference

GLUE improves test accuracy by up to 8.5% over data-size weighting and by up to 9.1% over proxy-metric selection.

Research#llm🔬 ResearchAnalyzed: Dec 25, 2025 11:49

Random Gradient-Free Optimization in Infinite Dimensional Spaces

Published:Dec 25, 2025 05:00
1 min read
ArXiv Stats ML

Analysis

This paper introduces a novel random gradient-free optimization method tailored for infinite-dimensional Hilbert spaces, addressing functional optimization challenges. The approach circumvents the computational difficulties associated with infinite-dimensional gradients by relying on directional derivatives and a pre-basis for the Hilbert space. This is a significant improvement over traditional methods that rely on finite-dimensional gradient descent over function parameterizations. The method's applicability is demonstrated through solving partial differential equations using a physics-informed neural network (PINN) approach, showcasing its potential for provable convergence. The reliance on easily obtainable pre-bases and directional derivatives makes this method more tractable than approaches requiring orthonormal bases or reproducing kernels. This research offers a promising avenue for optimization in complex functional spaces.
Reference

To overcome this limitation, our framework requires only the computation of directional derivatives and a pre-basis for the Hilbert space domain.