Optimization
UDC: A Unified Neural Divide-and-Conquer Framework for Large-Scale Combinatorial Optimization Problems
·5789 words·28 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
๐ข School of System Design and Intelligent Manufacturing, Southern University of Science and Technology
A unified neural divide-and-conquer framework (UDC) achieves superior performance on large-scale combinatorial optimization problems by employing a novel Divide-Conquer-Reunion training method and a h…
Truthfulness of Calibration Measures
·337 words·2 mins·
loading
·
loading
AI Theory
Optimization
๐ข UC Berkeley
Researchers developed Subsampled Smooth Calibration Error (SSCE), a new truthful calibration measure for sequential prediction, solving the problem of existing measures being easily gamed.
Transductive Learning is Compact
·325 words·2 mins·
loading
·
loading
AI Theory
Optimization
๐ข USC
Supervised learning’s sample complexity is compact: a hypothesis class is learnable if and only if all its finite projections are learnable, simplifying complexity analysis.
Trading off Consistency and Dimensionality of Convex Surrogates for Multiclass Classification
·1544 words·8 mins·
loading
·
loading
AI Theory
Optimization
๐ข Harvard University
Researchers achieve a balance between accuracy and efficiency in multiclass classification by introducing partially consistent surrogate losses and novel methods.
Trade-Offs of Diagonal Fisher Information Matrix Estimators
·2685 words·13 mins·
loading
·
loading
AI Theory
Optimization
๐ข Australian National University
This paper examines the trade-offs between two popular diagonal Fisher Information Matrix (FIM) estimators in neural networks, deriving variance bounds and highlighting the importance of considering e…
Trace is the Next AutoDiff: Generative Optimization with Rich Feedback, Execution Traces, and LLMs
·2686 words·13 mins·
loading
·
loading
AI Theory
Optimization
๐ข Microsoft Research
Trace: Automating AI workflow design with LLMs.
Toward Global Convergence of Gradient EM for Over-Paramterized Gaussian Mixture Models
·345 words·2 mins·
loading
·
loading
Machine Learning
Optimization
๐ข University of Washington
Gradient EM for over-parameterized Gaussian Mixture Models globally converges with a sublinear rate, solving a longstanding open problem in machine learning.
Topological obstruction to the training of shallow ReLU neural networks
·1553 words·8 mins·
loading
·
loading
AI Theory
Optimization
๐ข Politecnico Di Torino
Shallow ReLU neural networks face topological training obstructions due to gradient flow confinement on disconnected quadric hypersurfaces.
Tighter Convergence Bounds for Shuffled SGD via Primal-Dual Perspective
·1717 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
๐ข University of Wisconsin-Madison
Shuffled SGD’s convergence is now better understood through a primal-dual analysis, yielding tighter bounds that align with its superior empirical performance.
Tight Rates for Bandit Control Beyond Quadratics
·406 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
๐ข Princeton University
This paper presents an algorithm achieving ร(โT) optimal regret for bandit non-stochastic control with strongly-convex and smooth cost functions, overcoming prior limitations of suboptimal bounds.
Tight Bounds for Learning RUMs from Small Slates
·255 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
๐ข Google Research
Learning user preferences accurately from limited data is key; this paper shows that surprisingly small datasets suffice for precise prediction, and provides efficient algorithms to achieve this.
This Too Shall Pass: Removing Stale Observations in Dynamic Bayesian Optimization
·4337 words·21 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
๐ข IC, EPFL
W-DBO efficiently tackles stale data in dynamic Bayesian Optimization by leveraging a novel Wasserstein distance-based criterion to remove irrelevant observations, maintaining high sampling frequency …
Theoretical Characterisation of the Gauss Newton Conditioning in Neural Networks
·2952 words·14 mins·
loading
·
loading
AI Theory
Optimization
๐ข University of Basel
New theoretical bounds reveal how neural network architecture impacts the Gauss-Newton matrix’s conditioning, paving the way for improved optimization.
The Surprising Effectiveness of SP Voting with Partial Preferences
·3640 words·18 mins·
loading
·
loading
AI Theory
Optimization
๐ข Penn State University
Partial preferences and noisy votes hinder accurate ranking recovery; this paper introduces scalable SP voting variants, empirically demonstrating superior performance in recovering ground truth ranki…
The Space Complexity of Approximating Logistic Loss
·359 words·2 mins·
loading
·
loading
AI Theory
Optimization
๐ข LinkedIn Corporation
This paper proves fundamental space complexity lower bounds for approximating logistic loss, revealing that existing coreset constructions are surprisingly optimal.
The Secretary Problem with Predicted Additive Gap
·1651 words·8 mins·
loading
·
loading
AI Theory
Optimization
๐ข Institute of Computer Science, University of Bonn
Beat the 1/e barrier in the secretary problem using only an additive gap prediction!
The Sample Complexity of Gradient Descent in Stochastic Convex Optimization
·336 words·2 mins·
loading
·
loading
AI Theory
Optimization
๐ข Tel Aviv University
Gradient descent’s sample complexity in non-smooth stochastic convex optimization is ร(d/m+1/โm), matching worst-case ERMs and showing no advantage over naive methods.
The Road Less Scheduled
·2275 words·11 mins·
loading
·
loading
Optimization
๐ข Princeton University
Revolutionizing machine learning, Schedule-Free optimization achieves state-of-the-art results without needing learning rate schedules, simplifying training and improving efficiency.
The Reliability of OKRidge Method in Solving Sparse Ridge Regression Problems
·2340 words·11 mins·
loading
·
loading
AI Theory
Optimization
๐ข Wuhan University
OKRidge’s reliability for solving sparse ridge regression problems is rigorously proven through theoretical error analysis, enhancing its applicability in machine learning.
The Minimax Rate of HSIC Estimation for Translation-Invariant Kernels
·215 words·2 mins·
loading
·
loading
AI Theory
Optimization
๐ข Karlsruhe Institute of Technology
Researchers found the minimax optimal rate of HSIC estimation for translation-invariant kernels is O(nโปยน/ยฒ), settling a two-decade-old open question and validating many existing HSIC estimators.