AI Theory
A provable control of sensitivity of neural networks through a direct parameterization of the overall bi-Lipschitzness
·4589 words·22 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Tokyo
New framework directly controls neural network sensitivity by precisely parameterizing overall bi-Lipschitzness, offering improved robustness and generalization.
A Primal-Dual-Assisted Penalty Approach to Bilevel Optimization with Coupled Constraints
·2217 words·11 mins·
loading
·
loading
AI Theory
Optimization
π’ Rensselaer Polytechnic Institute
BLOCC, a novel first-order algorithm, efficiently solves bilevel optimization problems with coupled constraints, offering improved scalability and convergence for machine learning applications.
A Non-parametric Direct Learning Approach to Heterogeneous Treatment Effect Estimation under Unmeasured Confounding
·1578 words·8 mins·
loading
·
loading
AI Theory
Causality
π’ State University of New York at Binghamton
Estimating heterogeneous treatment effects (CATE) under unmeasured confounding is revolutionized by a novel non-parametric direct learning approach using instrumental variables, offering efficient and…
A Neural Network Approach for Efficiently Answering Most Probable Explanation Queries in Probabilistic Models
·11719 words·56 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Texas at Dallas
A novel neural network efficiently answers arbitrary Most Probable Explanation (MPE) queries in large probabilistic models, eliminating the need for slow inference algorithms.
A Huber Loss Minimization Approach to Mean Estimation under User-level Differential Privacy
·334 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Privacy
π’ Zhejiang Lab
Huber loss minimization ensures accurate and robust mean estimation under user-level differential privacy, especially for imbalanced datasets and heavy-tailed distributions.
A generalized neural tangent kernel for surrogate gradient learning
·1667 words·8 mins·
loading
·
loading
AI Theory
Generalization
π’ University of Bern
Researchers introduce a generalized neural tangent kernel for analyzing surrogate gradient learning in neural networks with non-differentiable activation functions, providing a strong theoretical foun…
A Fast Convoluted Story: Scaling Probabilistic Inference for Integer Arithmetics
·2476 words·12 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ KU Leuven
Revolutionizing probabilistic inference, PLIAβ uses tensor operations and FFT to scale integer arithmetic, achieving orders-of-magnitude speedup in inference and learning times.
A Comprehensive Analysis on the Learning Curve in Kernel Ridge Regression
·2395 words·12 mins·
loading
·
loading
AI Theory
Generalization
π’ University of Basel
This study provides a unified theory for kernel ridge regression’s learning curve, improving existing bounds and validating the Gaussian Equivalence Property under minimal assumptions.
A Compositional Atlas for Algebraic Circuits
·1573 words·8 mins·
loading
·
loading
AI Theory
Causality
π’ UC Los Angeles
This paper introduces a compositional framework for algebraic circuits, deriving novel tractability conditions for compositional inference queries and unifying existing results.
A Combinatorial Algorithm for the Semi-Discrete Optimal Transport Problem
·1938 words·10 mins·
loading
·
loading
AI Theory
Optimization
π’ Duke University
A new combinatorial algorithm dramatically speeds up semi-discrete optimal transport calculations, offering an efficient solution for large datasets and higher dimensions.
A Closer Look at AUROC and AUPRC under Class Imbalance
·2353 words·12 mins·
loading
·
loading
AI Theory
Fairness
π’ Harvard University
Debunking a common myth, this paper proves that AUPRC is not superior to AUROC for imbalanced datasets, and in fact, can worsen algorithmic bias.
A Boosting-Type Convergence Result for AdaBoost.MH with Factorized Multi-Class Classifiers
·358 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Wuhan University
Solved a long-standing open problem: Factorized ADABOOST.MH now has a proven convergence rate!
4+3 Phases of Compute-Optimal Neural Scaling Laws
·3282 words·16 mins·
loading
·
loading
AI Theory
Optimization
π’ McGill University
Researchers discovered four distinct compute-optimal phases for training neural networks, offering new predictions for resource-efficient large model training.
2D-OOB: Attributing Data Contribution Through Joint Valuation Framework
·2147 words·11 mins·
loading
·
loading
AI Theory
Interpretability
π’ University of Illinois Urbana-Champaign
2D-OOB: a novel framework for jointly attributing data values to individual features, enabling fine-grained outlier detection and improved model performance.