AI Theory
Adam with model exponential moving average is effective for nonconvex optimization
·281 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Microsoft Research
Clipped Adam with EMA achieves optimal convergence rates for smooth and non-smooth nonconvex optimization, particularly when scales vary across different coordinates.
Achieving Domain-Independent Certified Robustness via Knowledge Continuity
·2020 words·10 mins·
loading
·
loading
AI Theory
Robustness
π’ Carnegie Mellon University
Certifying neural network robustness across diverse domains, this paper introduces knowledge continuityβa novel framework ensuring model stability independent of input type, norms, and distribution.
Achievable Fairness on Your Data With Utility Guarantees
·6805 words·32 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
π’ ByteDance Research
This paper introduces a computationally efficient method to approximate the optimal accuracy-fairness trade-off curve for various datasets, providing rigorous statistical guarantees and quantifying un…
Achievable distributional robustness when the robust risk is only partially identified
·1876 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
π’ ETH Zurich
This paper introduces a novel framework for evaluating the robustness of machine learning models when the true data distribution is only partially known. It defines a new risk measure (‘identifiable r…
Acceleration Exists! Optimization Problems When Oracle Can Only Compare Objective Function Values
·1792 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ Moscow Institute of Physics and Technology
Accelerated gradient-free optimization is achieved using only function value comparisons, significantly improving black-box optimization.
Accelerating Nash Equilibrium Convergence in Monte Carlo Settings Through Counterfactual Value Based Fictitious Play
·1841 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ Huazhong University of Science and Technology
MCCFVFP, a novel Monte Carlo-based algorithm, accelerates Nash equilibrium convergence in large-scale games by combining CFR’s counterfactual value calculations with fictitious play’s best response st…
Accelerating Matroid Optimization through Fast Imprecise Oracles
·485 words·3 mins·
loading
·
loading
AI Theory
Optimization
π’ Technical University of Berlin
Fast imprecise oracles drastically reduce query times in matroid optimization, achieving near-optimal performance with few accurate queries.
Accelerating ERM for data-driven algorithm design using output-sensitive techniques
·366 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Carnegie Mellon University
Accelerating ERM for data-driven algorithm design using output-sensitive techniques achieves computationally efficient learning by scaling with the actual number of pieces in the dual loss function, n…
Accelerated Regularized Learning in Finite N-Person Games
·1352 words·7 mins·
loading
·
loading
AI Theory
Optimization
π’ Stanford University
Accelerated learning in games achieved! FTXL algorithm exponentially speeds up convergence to Nash equilibria in finite N-person games, even under limited feedback.
Abductive Reasoning in Logical Credal Networks
·1676 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ IBM Research
This paper presents efficient algorithms for abductive reasoning in Logical Credal Networks (LCNs), addressing the MAP and Marginal MAP inference tasks to enable scalable solutions for complex real-wo…
A Walsh Hadamard Derived Linear Vector Symbolic Architecture
·1922 words·10 mins·
loading
·
loading
AI Theory
Representation Learning
π’ University of Maryland, Baltimore County
Hadamard-derived Linear Binding (HLB): A novel, efficient vector symbolic architecture surpassing existing methods in classical AI tasks and deep learning applications.
A Universal Growth Rate for Learning with Smooth Surrogate Losses
·1364 words·7 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Courant Institute
This paper reveals a universal square-root growth rate for H-consistency bounds of smooth surrogate losses in classification, significantly advancing our understanding of loss function selection.
A Unifying Post-Processing Framework for Multi-Objective Learn-to-Defer Problems
·1692 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
π’ Max Planck Institute for Intelligent Systems
A novel post-processing framework, based on a d-dimensional generalization of the Neyman-Pearson lemma, optimally solves multi-objective learn-to-defer problems under various constraints, improving co…
A Theory of Optimistically Universal Online Learnability for General Concept Classes
·411 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Purdue University
This paper fully characterizes concept classes optimistically universally learnable online, introducing novel algorithms and revealing equivalences between agnostic and realizable settings.
A theoretical design of concept sets: improving the predictability of concept bottleneck models
·1648 words·8 mins·
loading
·
loading
AI Theory
Interpretability
π’ University of Cambridge
Boosting concept bottleneck model predictability, this paper introduces a theoretical framework linking concept set properties to model performance, proposing a method for effective concept identifica…
A Simple yet Scalable Granger Causal Structural Learning Approach for Topological Event Sequences
·1497 words·8 mins·
loading
·
loading
AI Theory
Causality
π’ East China Normal University
SΒ²GCSL: A novel scalable Granger causal structural learning approach efficiently identifies root causes of telecommunication network alarms by leveraging a linear kernel and incorporating expert knowl…
A Simple Remedy for Dataset Bias via Self-Influence: A Mislabeled Sample Perspective
·3350 words·16 mins·
loading
·
loading
AI Theory
Fairness
π’ Korea Advanced Institute of Science and Technology
This paper introduces Bias-Conditioned Self-Influence (BCSI) for precise bias-conflicting sample detection and model rectification, enhancing fairness in machine learning.
A Simple and Optimal Approach for Universal Online Learning with Gradient Variations
·244 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Nanjing University
A novel universal online learning algorithm achieves optimal gradient-variation regret across diverse function curvatures, boasting efficiency with only one gradient query per round.
A Simple and Adaptive Learning Rate for FTRL in Online Learning with Minimax Regret of Ξ(T^{2/3}) and its Application to Best-of-Both-Worlds
·334 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Tokyo
A new adaptive learning rate for FTRL achieves minimax regret of O(TΒ²/Β³) in online learning, improving existing best-of-both-worlds algorithms for various hard problems.
A Separation in Heavy-Tailed Sampling: Gaussian vs. Stable Oracles for Proximal Samplers
·1758 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ Georgia Institute of Technology
Stable oracles outperform Gaussian oracles in high-accuracy heavy-tailed sampling, overcoming limitations of Gaussian-based proximal samplers.