Skip to main content

Optimization

The Many Faces of Optimal Weak-to-Strong Learning
·1344 words·7 mins· loading · loading
Machine Learning Optimization šŸ¢ Aarhus University
A new, surprisingly simple boosting algorithm achieves provably optimal sample complexity and outperforms existing algorithms on large datasets.
The Implicit Bias of Heterogeneity towards Invariance: A Study of Multi-Environment Matrix Sensing
·1551 words·8 mins· loading · loading
AI Theory Optimization šŸ¢ Peking University
Leveraging data heterogeneity, this study reveals that standard SGD implicitly learns invariant features across multiple environments, achieving robust generalization without explicit regularization.
The Implicit Bias of Adam on Separable Data
·1356 words·7 mins· loading · loading
AI Theory Optimization šŸ¢ Hong Kong University of Science and Technology
Adam’s implicit bias revealed: On separable data, Adam converges towards the maximum lāˆž-margin solution, a finding contrasting with gradient descent’s l2-margin preference. This polynomial-time conver…
The High Line: Exact Risk and Learning Rate Curves of Stochastic Adaptive Learning Rate Algorithms
·1819 words·9 mins· loading · loading
Machine Learning Optimization šŸ¢ McGill University
Researchers developed a framework for analyzing stochastic adaptive learning rate algorithms, providing exact risk and learning rate curves, revealing the importance of data covariance and uncovering …
The Empirical Impact of Neural Parameter Symmetries, or Lack Thereof
·2500 words·12 mins· loading · loading
AI Theory Optimization šŸ¢ MIT
Breaking neural network parameter symmetries leads to faster training, better generalization, and improved loss landscape behavior, as demonstrated by novel asymmetric network architectures.
The Challenges of the Nonlinear Regime for Physics-Informed Neural Networks
·2142 words·11 mins· loading · loading
AI Generated AI Theory Optimization šŸ¢ BMW AG
Physics-Informed Neural Networks (PINNs) training dynamics for nonlinear PDEs are fundamentally different than linear ones; this paper reveals why using second-order methods is crucial for solving non…
The Bayesian sampling in a canonical recurrent circuit with a diversity of inhibitory interneurons
·1541 words·8 mins· loading · loading
AI Theory Optimization šŸ¢ UT Southwestern Medical Center
Diverse inhibitory neurons in brain circuits enable faster Bayesian computation via Hamiltonian sampling.
Symmetries in Overparametrized Neural Networks: A Mean Field View
·2636 words·13 mins· loading · loading
AI Theory Optimization šŸ¢ University of Chile
Overparametrized neural networks’ learning dynamics are analyzed under data symmetries using mean-field theory, revealing that data augmentation, feature averaging, and equivariant architectures asymp…
SymILO: A Symmetry-Aware Learning Framework for Integer Linear Optimization
·1858 words·9 mins· loading · loading
AI Theory Optimization šŸ¢ Shenzhen Research Institute of Big Data
SymILO: A novel symmetry-aware learning framework dramatically improves integer linear program (ILP) solutions by addressing data variability caused by ILP symmetry.
Strategic Littlestone Dimension: Improved Bounds on Online Strategic Classification
·389 words·2 mins· loading · loading
AI Theory Optimization šŸ¢ Toyota Technological Institute at Chicago
This paper introduces the Strategic Littlestone Dimension, a novel complexity measure for online strategic classification, proving instance-optimal mistake bounds in the realizable setting and improve…
Strategic Linear Contextual Bandits
·1349 words·7 mins· loading · loading
AI Theory Optimization šŸ¢ Alan Turing Institute
Strategic agents gaming recommender systems is solved by a novel mechanism that incentivizes truthful behavior while minimizing regret, offering a solution to a key challenge in online learning.
Stopping Bayesian Optimization with Probabilistic Regret Bounds
·3802 words·18 mins· loading · loading
Machine Learning Optimization šŸ¢ Morgan Stanley
This paper presents a novel probabilistic regret bound (PRB) framework for Bayesian optimization, replacing the traditional fixed-budget stopping rule with a criterion based on the probability of find…
Stochastic Zeroth-Order Optimization under Strongly Convexity and Lipschitz Hessian: Minimax Sample Complexity
·361 words·2 mins· loading · loading
AI Theory Optimization šŸ¢ UC Santa Barbara
Stochastic zeroth-order optimization of strongly convex functions with Lipschitz Hessian achieves optimal sample complexity, as proven by matching upper and lower bounds with a novel two-stage algorit…
Stochastic Optimization Schemes for Performative Prediction with Nonconvex Loss
·2232 words·11 mins· loading · loading
AI Generated Machine Learning Optimization šŸ¢ Chinese University of Hong Kong
Bias-free performative prediction is achieved using a novel lazy deployment scheme with SGD, handling non-convex loss functions.
Stochastic Optimal Control Matching
·1801 words·9 mins· loading · loading
AI Theory Optimization šŸ¢ Meta AI
Stochastic Optimal Control Matching (SOCM) significantly reduces errors in stochastic optimal control by learning a matching vector field using a novel iterative diffusion optimization technique.
Stochastic Optimal Control and Estimation with Multiplicative and Internal Noise
·2159 words·11 mins· loading · loading
AI Theory Optimization šŸ¢ Pompeu Fabra University
A novel algorithm significantly improves stochastic optimal control by accurately modeling sensorimotor noise, achieving substantially lower costs than current state-of-the-art solutions, particularly…
Stochastic Newton Proximal Extragradient Method
·1769 words·9 mins· loading · loading
AI Generated AI Theory Optimization šŸ¢ University of Texas at Austin
Stochastic Newton Proximal Extragradient (SNPE) achieves faster global and local convergence rates for strongly convex functions, improving upon existing stochastic Newton methods by requiring signifi…
Stochastic Extragradient with Flip-Flop Shuffling & Anchoring: Provable Improvements
·1705 words·9 mins· loading · loading
AI Generated AI Theory Optimization šŸ¢ KAIST
Stochastic extragradient with flip-flop shuffling & anchoring achieves provably faster convergence in minimax optimization.
Statistical-Computational Trade-offs for Density Estimation
·433 words·3 mins· loading · loading
AI Theory Optimization šŸ¢ MIT
Density estimation algorithms face inherent trade-offs: reducing sample needs often increases query time. This paper proves these trade-offs are fundamental, showing limits to how much improvement is…
Statistical Estimation in the Spiked Tensor Model via the Quantum Approximate Optimization Algorithm
·1516 words·8 mins· loading · loading
AI Theory Optimization šŸ¢ University of California, Los Angeles
Quantum Approximate Optimization Algorithm (QAOA) achieves weak recovery in spiked tensor models matching classical methods, but with potential constant factor advantages for certain parameters.