Skip to main content

AI Theory

Bayes-optimal learning of an extensive-width neural network from quadratically many samples
·1425 words·7 mins· loading · loading
AI Theory Optimization 🏒 ETH Zurich
This study solves a key challenge in neural network learning, deriving a closed-form expression for the Bayes-optimal test error of extensive-width networks with quadratic activation functions from qu…
Barely Random Algorithms and Collective Metrical Task Systems
·315 words·2 mins· loading · loading
AI Theory Optimization 🏒 Inria
Randomness-efficient algorithms are developed for online decision making, requiring only 2log n random bits and achieving near-optimal competitiveness for metrical task systems.
Bandit-Feedback Online Multiclass Classification: Variants and Tradeoffs
·455 words·3 mins· loading · loading
AI Generated AI Theory Optimization 🏒 Faculty of Computer Science,Technion,Israel
This paper reveals the optimal mistake bounds for online multiclass classification under bandit feedback, showing the cost of limited feedback is at most O(k) times higher than full information, where…
Banded Square Root Matrix Factorization for Differentially Private Model Training
·4880 words·23 mins· loading · loading
AI Theory Privacy 🏒 Institute of Science and Technology (ISTA)
This paper introduces BSR, a novel banded square root matrix factorization for differentially private model training. Unlike existing methods, BSR avoids computationally expensive optimization, enabli…
Back to the Continuous Attractor
·5636 words·27 mins· loading · loading
AI Generated AI Theory Generalization 🏒 Champalimaud Centre for the Unknown
Despite their brittleness, continuous attractors remain functionally robust analog memory models due to persistent slow manifolds surviving bifurcations, enabling accurate approximation and generaliza…
B-cosification: Transforming Deep Neural Networks to be Inherently Interpretable
·2514 words·12 mins· loading · loading
AI Theory Interpretability 🏒 Max Planck Institute for Informatics
B-cosification: cheaply transform any pre-trained deep neural network into an inherently interpretable model.
Axioms for AI Alignment from Human Feedback
·1869 words·9 mins· loading · loading
AI Theory Optimization 🏒 Harvard University
This paper revolutionizes AI alignment by applying social choice theory axioms to RLHF, exposing flaws in existing methods and proposing novel, axiomatically guaranteed reward learning rules.
Average gradient outer product as a mechanism for deep neural collapse
·2027 words·10 mins· loading · loading
AI Theory Optimization 🏒 UC San Diego
Deep Neural Collapse (DNC) explained via Average Gradient Outer Product (AGOP).
Automating Data Annotation under Strategic Human Agents: Risks and Potential Solutions
·6858 words·33 mins· loading · loading
AI Theory Fairness 🏒 Ohio State University
AI models retraining with model-annotated data incorporating human strategic responses can lead to unexpected outcomes, potentially reducing the proportion of agents with positive labels over time, wh…
Automatic Outlier Rectification via Optimal Transport
·2826 words·14 mins· loading · loading
AI Theory Optimization 🏒 Stanford University
This study presents a novel single-step outlier rectification method using optimal transport with a concave cost function, surpassing the limitations of conventional two-stage approaches by jointly op…
Automated Efficient Estimation using Monte Carlo Efficient Influence Functions
·1825 words·9 mins· loading · loading
AI Theory Causality 🏒 Basis Research Institute
MC-EIF automates efficient statistical estimation for high-dimensional models, integrating seamlessly with existing differentiable probabilistic programming systems and achieving optimal convergence r…
Autobidder's Dilemma: Why More Sophisticated Autobidders Lead to Worse Auction Efficiency
·332 words·2 mins· loading · loading
AI Theory Optimization 🏒 Google Research
More sophisticated autobidders surprisingly worsen online auction efficiency; a fine-grained analysis reveals that less powerful, uniform bidders lead to better market outcomes.
Auditing Privacy Mechanisms via Label Inference Attacks
·1460 words·7 mins· loading · loading
AI Theory Privacy 🏒 Google Research
New metrics audit label privatization, revealing differentially private schemes often outperform heuristic methods in the privacy-utility tradeoff.
Auditing Local Explanations is Hard
·1271 words·6 mins· loading · loading
AI Theory Interpretability 🏒 University of Tübingen and Tübingen AI Center
Auditing local explanations is surprisingly hard: proving explanation trustworthiness requires far more data than previously thought, especially in high dimensions, challenging current AI explainabil…
Attack-Aware Noise Calibration for Differential Privacy
·2558 words·13 mins· loading · loading
AI Generated AI Theory Privacy 🏒 Lausanne University Hospital
Boosting machine learning model accuracy in privacy-preserving applications, this research introduces novel noise calibration methods directly targeting desired attack risk levels, bypassing conventio…
Are Your Models Still Fair? Fairness Attacks on Graph Neural Networks via Node Injections
·2258 words·11 mins· loading · loading
AI Theory Fairness 🏒 Huazhong University of Science and Technology
Node Injection-based Fairness Attack (NIFA) reveals GNNs’ vulnerability to realistic fairness attacks by injecting a small percentage of nodes, significantly undermining fairness even in fairness-awar…
Are High-Degree Representations Really Unnecessary in Equivariant Graph Neural Networks?
·2234 words·11 mins· loading · loading
AI Theory Representation Learning 🏒 Gaoling School of Artificial Intelligence, Renmin University of China
High-degree representations significantly boost the expressiveness of E(3)-equivariant GNNs, overcoming limitations of lower-degree models on symmetric structures, as demonstrated theoretically and em…
Are Graph Neural Networks Optimal Approximation Algorithms?
·2409 words·12 mins· loading · loading
AI Theory Optimization 🏒 Massachusetts Institute of Technology
Graph Neural Networks (GNNs) learn optimal approximation algorithms for combinatorial optimization problems, achieving high-quality solutions for Max-Cut, Min-Vertex-Cover, and Max-3-SAT, while also p…
Approximating the Top Eigenvector in Random Order Streams
·341 words·2 mins· loading · loading
AI Theory Optimization 🏒 Google Research
Random-order stream data necessitates efficient top eigenvector approximation; this paper presents novel algorithms with improved space complexity, achieving near-optimal bounds.
Approximating mutual information of high-dimensional variables using learned representations
·2528 words·12 mins· loading · loading
AI Theory Representation Learning 🏒 Harvard University
Latent Mutual Information (LMI) approximation accurately estimates mutual information in high-dimensional data using low-dimensional learned representations, solving a critical problem in various scie…