Skip to main content

AI Theory

On Weak Regret Analysis for Dueling Bandits
·1775 words·9 mins· loading · loading
AI Generated AI Theory Optimization 🏢 KAUST
New algorithms achieve optimal weak regret in K-armed dueling bandits by leveraging the full problem structure, improving upon state-of-the-art methods.
On Tractable $
hi$-Equilibria in Non-Concave Games
·1428 words·7 mins· loading · loading
AI Theory Optimization 🏢 Yale University
This paper presents efficient algorithms for approximating equilibria in non-concave games, focusing on tractable ɸ-equilibria and addressing computational challenges posed by infinite strategy sets.
On the Sparsity of the Strong Lottery Ticket Hypothesis
·1303 words·7 mins· loading · loading
AI Theory Optimization 🏢 Université Côte D'Azur
Researchers rigorously prove the Strong Lottery Ticket Hypothesis, offering the first theoretical guarantees on the sparsity of winning neural network subnetworks.
On the Saturation Effects of Spectral Algorithms in Large Dimensions
·1464 words·7 mins· loading · loading
AI Theory Generalization 🏢 Tsinghua University
High-dimensional spectral algorithms show saturation effects: Kernel Ridge Regression underperforms optimal algorithms like gradient flow when regression functions are very smooth.
On the Role of Attention Masks and LayerNorm in Transformers
·2522 words·12 mins· loading · loading
AI Generated AI Theory Representation Learning 🏢 MIT
Transformers’ self-attention mechanism, while powerful, suffers from rank collapse with increasing depth. This paper reveals that while masked attention still leads to exponential collapse, sparse att…
On the Robustness of Spectral Algorithms for Semirandom Stochastic Block Models
·1629 words·8 mins· loading · loading
AI Theory Robustness 🏢 University of Utah
Spectral algorithms for graph bisection show surprising robustness to helpful adversaries in semirandom models, with unnormalized Laplacian consistently outperforming the normalized one.
On the Power of Small-size Graph Neural Networks for Linear Programming
·2361 words·12 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Peking University
Small-size Graph Neural Networks effectively solve Linear Programs!
On the Parameter Identifiability of Partially Observed Linear Causal Models
·3769 words·18 mins· loading · loading
AI Generated AI Theory Causality 🏢 Carnegie Mellon University
Researchers achieve full parameter identifiability in partially observed linear causal models using novel graphical conditions and a likelihood-based estimation method, addressing previous limitations…
On the Optimality of Dilated Entropy and Lower Bounds for Online Learning in Extensive-Form Games
·1661 words·8 mins· loading · loading
AI Generated AI Theory Optimization 🏢 MIT
Researchers discover Dilated Entropy is the optimal distance-generating function for solving extensive-form games using first-order methods, achieving near-optimal regret bounds.
On the Impacts of the Random Initialization in the Neural Tangent Kernel Theory
·1555 words·8 mins· loading · loading
AI Theory Generalization 🏢 Tsinghua University
Standard initialization in neural networks negatively impacts generalization ability under Neural Tangent Kernel theory, contradicting real-world performance, urging the development of improved theore…
On the Impact of Feature Heterophily on Link Prediction with Graph Neural Networks
·2063 words·10 mins· loading · loading
AI Theory Representation Learning 🏢 University of Michigan
Graph Neural Networks (GNNs) struggle with heterophilic link prediction; this paper introduces formal definitions, theoretical analysis, improved designs, and real-world benchmarks to address this cha…
On the Identifiability of Poisson Branching Structural Causal Model Using Probability Generating Function
·2064 words·10 mins· loading · loading
AI Theory Causality 🏢 Guangdong University of Technology
Researchers developed a novel, efficient causal discovery method using Probability Generating Functions to identify causal structures within Poisson Branching Structural Causal Models, overcoming limi…
On the Expressivity and Sample Complexity of Node-Individualized Graph Neural Networks
·2191 words·11 mins· loading · loading
AI Generated AI Theory Generalization 🏢 Max Planck Institute of Biochemistry
Boosting GNN expressivity and generalization: Novel node individualization schemes lower sample complexity, improving substructure identification.
On the Expressive Power of Tree-Structured Probabilistic Circuits
·1425 words·7 mins· loading · loading
AI Theory Optimization 🏢 University of Illinois Urbana-Champaign
Tree-structured probabilistic circuits are surprisingly efficient: this paper proves a quasi-polynomial upper bound on their size, showing they’re almost as expressive as more complex DAG structures.
On the Computational Landscape of Replicable Learning
·348 words·2 mins· loading · loading
AI Theory Optimization 🏢 Yale University
This paper reveals surprising computational connections between algorithmic replicability and other learning paradigms, offering novel algorithms and demonstrating separations between replicability an…
On the Computational Complexity of Private High-dimensional Model Selection
·1501 words·8 mins· loading · loading
AI Theory Privacy 🏢 University of Michigan
This paper proposes a computationally efficient, differentially private best subset selection method for high-dimensional sparse linear regression, achieving both strong statistical utility and provab…
On the Complexity of Identification in Linear Structural Causal Models
·1403 words·7 mins· loading · loading
AI Theory Causality 🏢 Saarland University
New polynomial-space algorithm for causal parameter identification in linear models vastly improves upon existing methods, showing that this crucial task is computationally hard.
On the cohesion and separability of average-link for hierarchical agglomerative clustering
·1663 words·8 mins· loading · loading
AI Theory Optimization 🏢 Departmento De Informática, PUC-RIO
Average-link hierarchical clustering gets a comprehensive evaluation using new criteria, showing it outperforms other methods when both cohesion and separability matter.
On the Benefits of Public Representations for Private Transfer Learning under Distribution Shift
·1927 words·10 mins· loading · loading
AI Generated AI Theory Privacy 🏢 Carnegie Mellon University
Public data boosts private AI accuracy even with extreme distribution shifts, improving private model training by up to 67% in three tasks. This is due to shared low-dimensional representations betwe…
On the Adversarial Robustness of Benjamini Hochberg
·1747 words·9 mins· loading · loading
AI Generated AI Theory Robustness 🏢 Operations Research Department Naval Postgraduate School
Even a few data changes can break the Benjamini-Hochberg (BH) procedure, a widely used multiple testing method, highlighting a critical vulnerability.