AI Theory
The Intelligible and Effective Graph Neural Additive Network
·2248 words·11 mins·
loading
·
loading
AI Theory
Interpretability
🏢 Tel Aviv University
GNAN: a novel interpretable graph neural network achieving accuracy comparable to black-box models.
The Implicit Bias of Heterogeneity towards Invariance: A Study of Multi-Environment Matrix Sensing
·1551 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 Peking University
Leveraging data heterogeneity, this study reveals that standard SGD implicitly learns invariant features across multiple environments, achieving robust generalization without explicit regularization.
The Implicit Bias of Gradient Descent toward Collaboration between Layers: A Dynamic Analysis of Multilayer Perceptions
·1405 words·7 mins·
loading
·
loading
AI Theory
Robustness
🏢 Department of Computer Science University of Exeter
Deep learning models’ success hinges on understanding gradient descent’s implicit bias. This study reveals how this bias influences layer collaboration, revealing a decreasing trend in adversarial rob…
The Implicit Bias of Adam on Separable Data
·1356 words·7 mins·
loading
·
loading
AI Theory
Optimization
🏢 Hong Kong University of Science and Technology
Adam’s implicit bias revealed: On separable data, Adam converges towards the maximum l∞-margin solution, a finding contrasting with gradient descent’s l2-margin preference. This polynomial-time conver…
The Group Robustness is in the Details: Revisiting Finetuning under Spurious Correlations
·2643 words·13 mins·
loading
·
loading
AI Theory
Fairness
🏢 Google DeepMind
Finetuning’s impact on worst-group accuracy is surprisingly nuanced, with common class-balancing methods sometimes hurting performance; a novel mixture method consistently outperforms others.
The Fairness-Quality Tradeoff in Clustering
·2122 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
🏢 Columbia University
Novel algorithms trace the optimal balance between clustering quality and fairness, revealing all non-dominated solutions for various objectives.
The Empirical Impact of Neural Parameter Symmetries, or Lack Thereof
·2500 words·12 mins·
loading
·
loading
AI Theory
Optimization
🏢 MIT
Breaking neural network parameter symmetries leads to faster training, better generalization, and improved loss landscape behavior, as demonstrated by novel asymmetric network architectures.
The Challenges of the Nonlinear Regime for Physics-Informed Neural Networks
·2142 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 BMW AG
Physics-Informed Neural Networks (PINNs) training dynamics for nonlinear PDEs are fundamentally different than linear ones; this paper reveals why using second-order methods is crucial for solving non…
The Bayesian sampling in a canonical recurrent circuit with a diversity of inhibitory interneurons
·1541 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 UT Southwestern Medical Center
Diverse inhibitory neurons in brain circuits enable faster Bayesian computation via Hamiltonian sampling.
Testing Calibration in Nearly-Linear Time
·1823 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Interpretability
🏢 Harvard University
This paper presents nearly-linear time algorithms for testing model calibration, improving upon existing methods and providing theoretical lower bounds for various calibration measures.
Testably Learning Polynomial Threshold Functions
·248 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Generalization
🏢 ETH Zurich
Testably learning polynomial threshold functions efficiently, matching agnostic learning’s best guarantees, is achieved, solving a key problem in robust machine learning.
Temporal Graph Neural Tangent Kernel with Graphon-Guaranteed
·1777 words·9 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Meta AI
Temp-G³NTK: a novel temporal graph neural tangent kernel guarantees convergence to graphon NTK, offering superior performance in temporal graph classification and node-level tasks.
Targeted Sequential Indirect Experiment Design
·2254 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 Technical University of Munich
Adaptive experiment design optimizes indirect experiments in complex systems by sequentially narrowing the gap between upper and lower bounds on a targeted query, providing more efficient and informat…
Tangent Space Causal Inference: Leveraging Vector Fields for Causal Discovery in Dynamical Systems
·1953 words·10 mins·
loading
·
loading
AI Theory
Causality
🏢 Stony Brook University
Tangent Space Causal Inference (TSCI) enhances causal discovery in dynamical systems by leveraging vector fields, outperforming existing methods in accuracy and interpretability.
Symmetries in Overparametrized Neural Networks: A Mean Field View
·2636 words·13 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Chile
Overparametrized neural networks’ learning dynamics are analyzed under data symmetries using mean-field theory, revealing that data augmentation, feature averaging, and equivariant architectures asymp…
SymILO: A Symmetry-Aware Learning Framework for Integer Linear Optimization
·1858 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Shenzhen Research Institute of Big Data
SymILO: A novel symmetry-aware learning framework dramatically improves integer linear program (ILP) solutions by addressing data variability caused by ILP symmetry.
SureMap: Simultaneous mean estimation for single-task and multi-task disaggregated evaluation
·2443 words·12 mins·
loading
·
loading
AI Theory
Fairness
🏢 Princeton University
SureMap, a new method, significantly boosts accuracy in single and multi-task disaggregated evaluations of AI models using limited data by transforming the problem into Gaussian mean estimation and cl…
SuperDeepFool: a new fast and accurate minimal adversarial attack
·4315 words·21 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 EPFL
SuperDeepFool: a fast, accurate algorithm generating minimal adversarial perturbations, significantly improving deep learning model robustness evaluation and adversarial training.
Structured Learning of Compositional Sequential Interventions
·2166 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 University College London
Predicting outcomes of combined sequential interventions is challenging, especially in sparse data. This paper introduces CSI-VAE, a novel compositional model that provides reliable predictions for u…
Structured flexibility in recurrent neural networks via neuromodulation
·1567 words·8 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Stanford University
Neuromodulated RNNs (NM-RNNs) enhance RNN flexibility by dynamically scaling recurrent weights using a neuromodulatory subnetwork, achieving higher accuracy and generalizability on various tasks compa…