Spotlight AI Theories
2024
Honor Among Bandits: No-Regret Learning for Online Fair Division
·357 words·2 mins·
loading
·
loading
AI Theory
Fairness
π’ Harvard University
Online fair division algorithm achieves Γ(TΒ²/Β³) regret while guaranteeing envy-freeness or proportionality in expectation, a result proven tight.
Get rich quick: exact solutions reveal how unbalanced initializations promote rapid feature learning
·2253 words·11 mins·
loading
·
loading
AI Theory
Optimization
π’ Stanford University
Unbalanced initializations dramatically accelerate neural network feature learning by modifying the geometry of learning trajectories, enabling faster feature extraction and improved generalization.
Generalization Analysis for Label-Specific Representation Learning
·269 words·2 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Southeast University
Researchers derived tighter generalization bounds for label-specific representation learning (LSRL) methods, improving understanding of LSRL’s success and offering guidance for future algorithm develo…
Exploring Jacobian Inexactness in Second-Order Methods for Variational Inequalities: Lower Bounds, Optimal Algorithms and Quasi-Newton Approximations
·349 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Mohamed Bin Zayed University of Artificial Intelligence
VIJI, a novel second-order algorithm, achieves optimal convergence rates for variational inequalities even with inexact Jacobian information, bridging the gap between theory and practice in machine le…
Enhancing Robustness of Graph Neural Networks on Social Media with Explainable Inverse Reinforcement Learning
·1977 words·10 mins·
loading
·
loading
AI Theory
Robustness
π’ Hong Kong University of Science and Technology
MoE-BiEntIRL: A novel explainable inverse reinforcement learning method enhances GNN robustness against diverse social media attacks by reconstructing attacker policies and generating more robust trai…
Energy-based Epistemic Uncertainty for Graph Neural Networks
·4139 words·20 mins·
loading
·
loading
AI Theory
Robustness
π’ Technical University of Munich
GEBM: a novel graph-based energy model for robust GNN uncertainty estimation.
Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space
·3166 words·15 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Harvard University
Generative models learn hidden capabilities suddenly during training, which can be explained and predicted using a novel ‘concept space’ framework that analyzes learning dynamics and concept signal.
ECLipsE: Efficient Compositional Lipschitz Constant Estimation for Deep Neural Networks
·2852 words·14 mins·
loading
·
loading
AI Theory
Robustness
π’ Purdue University
ECLipsE: A novel compositional approach drastically accelerates Lipschitz constant estimation for deep neural networks, achieving speedups of thousands of times compared to the state-of-the-art while …
Dimension-free deterministic equivalents and scaling laws for random feature regression
·1898 words·9 mins·
loading
·
loading
AI Theory
Generalization
π’ Γcole Normale SupΓ©rieure
This work delivers dimension-free deterministic equivalents for random feature regression, revealing sharp excess error rates and scaling laws.
Continual learning with the neural tangent ensemble
·1983 words·10 mins·
loading
·
loading
AI Theory
Generalization
π’ Cold Spring Harbor Laboratory
Neural networks, viewed as Bayesian ensembles of fixed classifiers, enable continual learning without forgetting; posterior updates mirror stochastic gradient descent, offering insights into optimizat…
Constrained Adaptive Attack: Effective Adversarial Attack Against Deep Neural Networks for Tabular Data
·2605 words·13 mins·
loading
·
loading
AI Theory
Robustness
π’ University of Luxembourg
Constrained Adaptive Attack (CAA) significantly improves adversarial attacks on deep learning models for tabular data by combining gradient and search-based methods, achieving up to 96.1% accuracy dro…
Can Transformers Smell Like Humans?
·2615 words·13 mins·
loading
·
loading
AI Theory
Representation Learning
π’ KTH Royal Institute of Technology
Pre-trained transformer models can predict human smell perception by encoding odorant chemical structures, aligning with expert labels, continuous ratings, and similarity assessments.
Benign overfitting in leaky ReLU networks with moderate input dimension
·366 words·2 mins·
loading
·
loading
AI Theory
Generalization
π’ University of California, Los Angeles
Leaky ReLU networks exhibit benign overfitting under surprisingly relaxed conditions: input dimension only needs to linearly scale with sample size, challenging prior assumptions in the field.
Barely Random Algorithms and Collective Metrical Task Systems
·315 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Inria
Randomness-efficient algorithms are developed for online decision making, requiring only 2log n random bits and achieving near-optimal competitiveness for metrical task systems.
Axioms for AI Alignment from Human Feedback
·1869 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ Harvard University
This paper revolutionizes AI alignment by applying social choice theory axioms to RLHF, exposing flaws in existing methods and proposing novel, axiomatically guaranteed reward learning rules.
Automated Efficient Estimation using Monte Carlo Efficient Influence Functions
·1825 words·9 mins·
loading
·
loading
AI Theory
Causality
π’ Basis Research Institute
MC-EIF automates efficient statistical estimation for high-dimensional models, integrating seamlessly with existing differentiable probabilistic programming systems and achieving optimal convergence r…
Auditing Privacy Mechanisms via Label Inference Attacks
·1460 words·7 mins·
loading
·
loading
AI Theory
Privacy
π’ Google Research
New metrics audit label privatization, revealing differentially private schemes often outperform heuristic methods in the privacy-utility tradeoff.
Are Graph Neural Networks Optimal Approximation Algorithms?
·2409 words·12 mins·
loading
·
loading
AI Theory
Optimization
π’ Massachusetts Institute of Technology
Graph Neural Networks (GNNs) learn optimal approximation algorithms for combinatorial optimization problems, achieving high-quality solutions for Max-Cut, Min-Vertex-Cover, and Max-3-SAT, while also p…
Approximating the Top Eigenvector in Random Order Streams
·341 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Google Research
Random-order stream data necessitates efficient top eigenvector approximation; this paper presents novel algorithms with improved space complexity, achieving near-optimal bounds.
Approximating mutual information of high-dimensional variables using learned representations
·2528 words·12 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Harvard University
Latent Mutual Information (LMI) approximation accurately estimates mutual information in high-dimensional data using low-dimensional learned representations, solving a critical problem in various scie…