Optimization
Are Graph Neural Networks Optimal Approximation Algorithms?
·2409 words·12 mins·
loading
·
loading
AI Theory
Optimization
π’ Massachusetts Institute of Technology
Graph Neural Networks (GNNs) learn optimal approximation algorithms for combinatorial optimization problems, achieving high-quality solutions for Max-Cut, Min-Vertex-Cover, and Max-3-SAT, while also p…
Approximation-Aware Bayesian Optimization
·1889 words·9 mins·
loading
·
loading
Optimization
π’ University of Pennsylvania
Approximation-Aware Bayesian Optimization (AABO) boosts high-dimensional Bayesian optimization by jointly optimizing model approximation and data acquisition, achieving superior efficiency and perform…
Approximating the Top Eigenvector in Random Order Streams
·341 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Google Research
Random-order stream data necessitates efficient top eigenvector approximation; this paper presents novel algorithms with improved space complexity, achieving near-optimal bounds.
An Improved Empirical Fisher Approximation for Natural Gradient Descent
·2632 words·13 mins·
loading
·
loading
Machine Learning
Optimization
π’ University of Cambridge
Improved Empirical Fisher (iEF) approximation significantly boosts the performance of Natural Gradient Descent (NGD) optimizers, offering superior convergence and generalization.
An Equivalence Between Static and Dynamic Regret Minimization
·321 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ UniversitΓ Degli Studi Di Milano
Dynamic regret minimization equals static regret in an extended space; this equivalence reveals a trade-off between loss variance and comparator variability, leading to a new algorithm achieving impro…
An Efficient High-dimensional Gradient Estimator for Stochastic Differential Equations
·1548 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Stanford University
New unbiased gradient estimator for high-dimensional SDEs drastically reduces computation time without sacrificing estimation accuracy.
An Analysis of Elo Rating Systems via Markov Chains
·2046 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ University of Warwick
Elo rating system’s convergence rigorously analyzed via Markov chains under the Bradley-Terry-Luce model, demonstrating competitive learning rates and informing efficient tournament design.
An Accelerated Gradient Method for Convex Smooth Simple Bilevel Optimization
·448 words·3 mins·
loading
·
loading
Machine Learning
Optimization
π’ University of Texas at Austin
Accelerated Gradient Method for Bilevel Optimization (AGM-BiO) achieves state-of-the-art convergence rates for simple bilevel optimization problems, requiring fewer iterations than existing methods to…
Amortized Eigendecomposition for Neural Networks
·2211 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Sea AI Lab
Accelerate neural network training using ‘amortized eigendecomposition’ β a novel method replacing expensive eigendecomposition with faster QR decomposition while preserving accuracy.
Almost Free: Self-concordance in Natural Exponential Families and an Application to Bandits
·359 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Alberta
Generalized linear bandits with subexponential reward distributions are self-concordant, enabling second-order regret bounds free of exponential dependence on problem parameters.
Aggregating Quantitative Relative Judgments: From Social Choice to Ranking Prediction
·2425 words·12 mins·
loading
·
loading
AI Theory
Optimization
π’ Carnegie Mellon University
This paper introduces Quantitative Relative Judgment Aggregation (QRJA), a novel social choice model, and applies it to ranking prediction, yielding effective and interpretable results on various real…
Adjust Pearson's $r$ to Measure Arbitrary Monotone Dependence
·1286 words·7 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Beijing University of Posts and Telecommunications
Researchers refine Pearson’s correlation coefficient to precisely measure arbitrary monotone dependence, expanding its applicability beyond linear relationships.
Addressing Bias in Online Selection with Limited Budget of Comparisons
·2019 words·10 mins·
loading
·
loading
AI Theory
Optimization
π’ ENSAE
This paper introduces efficient algorithms for online selection with a budget constraint when comparing candidates from different groups has a cost, improving fairness and efficiency.
Adaptive Variance Reduction for Stochastic Optimization under Weaker Assumptions
·1456 words·7 mins·
loading
·
loading
Machine Learning
Optimization
π’ Nanjing University
Adaptive STORM achieves optimal convergence rates for stochastic optimization of non-convex functions under weaker assumptions, eliminating the need for bounded gradients or function values and removi…
Adaptive Sampling for Efficient Softmax Approximation
·1972 words·10 mins·
loading
·
loading
Machine Learning
Optimization
π’ Stanford University
AdaptiveSoftmax: Achieve 10x+ speedup in softmax computation via adaptive sampling!
Adaptive Proximal Gradient Method for Convex Optimization
·1541 words·8 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Vienna
Adaptive gradient descent methods are improved by leveraging local curvature information for entirely adaptive algorithms without added computational cost, proving convergence with only local Lipschit…
Adaptive and Optimal Second-order Optimistic Methods for Minimax Optimization
·1754 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ University of Texas at Austin
New adaptive second-order optimistic methods for minimax optimization achieve optimal convergence without line search, simplifying updates and improving efficiency.
Adam with model exponential moving average is effective for nonconvex optimization
·281 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Microsoft Research
Clipped Adam with EMA achieves optimal convergence rates for smooth and non-smooth nonconvex optimization, particularly when scales vary across different coordinates.
Achieving Linear Convergence with Parameter-Free Algorithms in Decentralized Optimization
·1448 words·7 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
π’ Innopolis University
A novel parameter-free decentralized optimization algorithm achieves linear convergence for strongly convex, smooth objectives, eliminating the need for hyperparameter tuning and improving scalability…
Acceleration Exists! Optimization Problems When Oracle Can Only Compare Objective Function Values
·1792 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ Moscow Institute of Physics and Technology
Accelerated gradient-free optimization is achieved using only function value comparisons, significantly improving black-box optimization.