Skip to main content

AI Theory

Adam with model exponential moving average is effective for nonconvex optimization
·281 words·2 mins· loading · loading
AI Theory Optimization 🏒 Microsoft Research
Clipped Adam with EMA achieves optimal convergence rates for smooth and non-smooth nonconvex optimization, particularly when scales vary across different coordinates.
Achieving Domain-Independent Certified Robustness via Knowledge Continuity
·2020 words·10 mins· loading · loading
AI Theory Robustness 🏒 Carnegie Mellon University
Certifying neural network robustness across diverse domains, this paper introduces knowledge continuityβ€”a novel framework ensuring model stability independent of input type, norms, and distribution.
Achievable Fairness on Your Data With Utility Guarantees
·6805 words·32 mins· loading · loading
AI Generated AI Theory Fairness 🏒 ByteDance Research
This paper introduces a computationally efficient method to approximate the optimal accuracy-fairness trade-off curve for various datasets, providing rigorous statistical guarantees and quantifying un…
Achievable distributional robustness when the robust risk is only partially identified
·1876 words·9 mins· loading · loading
AI Generated AI Theory Robustness 🏒 ETH Zurich
This paper introduces a novel framework for evaluating the robustness of machine learning models when the true data distribution is only partially known. It defines a new risk measure (‘identifiable r…
Acceleration Exists! Optimization Problems When Oracle Can Only Compare Objective Function Values
·1792 words·9 mins· loading · loading
AI Theory Optimization 🏒 Moscow Institute of Physics and Technology
Accelerated gradient-free optimization is achieved using only function value comparisons, significantly improving black-box optimization.
Accelerating Nash Equilibrium Convergence in Monte Carlo Settings Through Counterfactual Value Based Fictitious Play
·1841 words·9 mins· loading · loading
AI Theory Optimization 🏒 Huazhong University of Science and Technology
MCCFVFP, a novel Monte Carlo-based algorithm, accelerates Nash equilibrium convergence in large-scale games by combining CFR’s counterfactual value calculations with fictitious play’s best response st…
Accelerating Matroid Optimization through Fast Imprecise Oracles
·485 words·3 mins· loading · loading
AI Theory Optimization 🏒 Technical University of Berlin
Fast imprecise oracles drastically reduce query times in matroid optimization, achieving near-optimal performance with few accurate queries.
Accelerating ERM for data-driven algorithm design using output-sensitive techniques
·366 words·2 mins· loading · loading
AI Theory Optimization 🏒 Carnegie Mellon University
Accelerating ERM for data-driven algorithm design using output-sensitive techniques achieves computationally efficient learning by scaling with the actual number of pieces in the dual loss function, n…
Accelerated Regularized Learning in Finite N-Person Games
·1352 words·7 mins· loading · loading
AI Theory Optimization 🏒 Stanford University
Accelerated learning in games achieved! FTXL algorithm exponentially speeds up convergence to Nash equilibria in finite N-person games, even under limited feedback.
Abductive Reasoning in Logical Credal Networks
·1676 words·8 mins· loading · loading
AI Generated AI Theory Optimization 🏒 IBM Research
This paper presents efficient algorithms for abductive reasoning in Logical Credal Networks (LCNs), addressing the MAP and Marginal MAP inference tasks to enable scalable solutions for complex real-wo…
A Walsh Hadamard Derived Linear Vector Symbolic Architecture
·1922 words·10 mins· loading · loading
AI Theory Representation Learning 🏒 University of Maryland, Baltimore County
Hadamard-derived Linear Binding (HLB): A novel, efficient vector symbolic architecture surpassing existing methods in classical AI tasks and deep learning applications.
A Universal Growth Rate for Learning with Smooth Surrogate Losses
·1364 words·7 mins· loading · loading
AI Generated AI Theory Optimization 🏒 Courant Institute
This paper reveals a universal square-root growth rate for H-consistency bounds of smooth surrogate losses in classification, significantly advancing our understanding of loss function selection.
A Unifying Post-Processing Framework for Multi-Objective Learn-to-Defer Problems
·1692 words·8 mins· loading · loading
AI Generated AI Theory Fairness 🏒 Max Planck Institute for Intelligent Systems
A novel post-processing framework, based on a d-dimensional generalization of the Neyman-Pearson lemma, optimally solves multi-objective learn-to-defer problems under various constraints, improving co…
A Theory of Optimistically Universal Online Learnability for General Concept Classes
·411 words·2 mins· loading · loading
AI Generated AI Theory Optimization 🏒 Purdue University
This paper fully characterizes concept classes optimistically universally learnable online, introducing novel algorithms and revealing equivalences between agnostic and realizable settings.
A theoretical design of concept sets: improving the predictability of concept bottleneck models
·1648 words·8 mins· loading · loading
AI Theory Interpretability 🏒 University of Cambridge
Boosting concept bottleneck model predictability, this paper introduces a theoretical framework linking concept set properties to model performance, proposing a method for effective concept identifica…
A Simple yet Scalable Granger Causal Structural Learning Approach for Topological Event Sequences
·1497 words·8 mins· loading · loading
AI Theory Causality 🏒 East China Normal University
SΒ²GCSL: A novel scalable Granger causal structural learning approach efficiently identifies root causes of telecommunication network alarms by leveraging a linear kernel and incorporating expert knowl…
A Simple Remedy for Dataset Bias via Self-Influence: A Mislabeled Sample Perspective
·3350 words·16 mins· loading · loading
AI Theory Fairness 🏒 Korea Advanced Institute of Science and Technology
This paper introduces Bias-Conditioned Self-Influence (BCSI) for precise bias-conflicting sample detection and model rectification, enhancing fairness in machine learning.
A Simple and Optimal Approach for Universal Online Learning with Gradient Variations
·244 words·2 mins· loading · loading
AI Theory Optimization 🏒 Nanjing University
A novel universal online learning algorithm achieves optimal gradient-variation regret across diverse function curvatures, boasting efficiency with only one gradient query per round.
A Simple and Adaptive Learning Rate for FTRL in Online Learning with Minimax Regret of Θ(T^{2/3}) and its Application to Best-of-Both-Worlds
·334 words·2 mins· loading · loading
AI Theory Optimization 🏒 University of Tokyo
A new adaptive learning rate for FTRL achieves minimax regret of O(TΒ²/Β³) in online learning, improving existing best-of-both-worlds algorithms for various hard problems.
A Separation in Heavy-Tailed Sampling: Gaussian vs. Stable Oracles for Proximal Samplers
·1758 words·9 mins· loading · loading
AI Theory Optimization 🏒 Georgia Institute of Technology
Stable oracles outperform Gaussian oracles in high-accuracy heavy-tailed sampling, overcoming limitations of Gaussian-based proximal samplers.