Skip to main content

AI Theory

Distribution Learning with Valid Outputs Beyond the Worst-Case
·320 words·2 mins· loading · loading
AI Theory Optimization 🏢 UC San Diego
Generative models often produce invalid outputs; this work shows that ensuring validity is easier than expected when using log-loss and carefully selecting model classes and data distributions.
Dissecting the Interplay of Attention Paths in a Statistical Mechanics Theory of Transformers
·2710 words·13 mins· loading · loading
AI Generated AI Theory Interpretability 🏢 Harvard University
Researchers dissected attention paths in Transformers using statistical mechanics, revealing a task-relevant kernel combination mechanism boosting generalization performance.
Dissecting the Failure of Invariant Learning on Graphs
·4452 words·21 mins· loading · loading
AI Generated AI Theory Generalization 🏢 Peking University
Cross-environment Intra-class Alignment (CIA) and its label-free variant, CIA-LRA, significantly improve node-level OOD generalization on graphs by aligning representations and eliminating spurious fe…
Disentangled Representation Learning in Non-Markovian Causal Systems
·2882 words·14 mins· loading · loading
AI Theory Causality 🏢 Columbia University
This paper introduces graphical criteria and an algorithm for disentangling causal factors from heterogeneous data in non-Markovian settings, advancing causal representation learning.
Discretely beyond $1/e$: Guided Combinatorial Algortihms for Submodular Maximization
·3091 words·15 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Texas A&M University
Researchers surpass the 1/e barrier in submodular maximization with novel combinatorial algorithms!
Directional Smoothness and Gradient Methods: Convergence and Adaptivity
·1502 words·8 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Stanford University
New sub-optimality bounds for gradient descent leverage directional smoothness, a localized gradient variation measure, achieving tighter convergence guarantees and adapting to optimization paths.
Direct Preference-Based Evolutionary Multi-Objective Optimization with Dueling Bandits
·5743 words·27 mins· loading · loading
AI Generated AI Theory Optimization 🏢 School of Computer Science and Engineering, University of Electronic Science and Technology of China
D-PBEMO: A novel framework for preference-based multi-objective optimization using clustering-based stochastic dueling bandits to directly leverage human feedback, improving efficiency and managing co…
Dimension-free Private Mean Estimation for Anisotropic Distributions
·233 words·2 mins· loading · loading
AI Generated AI Theory Privacy 🏢 UC Berkeley
Dimension-free private mean estimation is achieved for anisotropic data, breaking the curse of dimensionality in privacy-preserving high-dimensional analysis.
Dimension-free deterministic equivalents and scaling laws for random feature regression
·1898 words·9 mins· loading · loading
AI Theory Generalization 🏢 École Normale Supérieure
This work delivers dimension-free deterministic equivalents for random feature regression, revealing sharp excess error rates and scaling laws.
Diffusion Models are Certifiably Robust Classifiers
·1735 words·9 mins· loading · loading
AI Theory Robustness 🏢 Tsinghua University
Diffusion models are certifiably robust classifiers due to their inherent O(1) Lipschitzness, a property further enhanced by generalizing to noisy data, achieving over 80% certified robustness on CIFA…
DiffHammer: Rethinking the Robustness of Diffusion-Based Adversarial Purification
·3686 words·18 mins· loading · loading
AI Theory Robustness 🏢 Hong Kong University of Science and Technology
DiffHammer unveils weaknesses in diffusion-based adversarial defenses by introducing a novel attack bypassing existing evaluation limitations, leading to more robust security solutions.
Differentially Private Stochastic Gradient Descent with Fixed-Size Minibatches: Tighter RDP Guarantees with or without Replacement
·2045 words·10 mins· loading · loading
AI Theory Privacy 🏢 Texas State University
Tighter differential privacy (RDP) guarantees for DP-SGD with fixed-size minibatches are achieved, improving private deep learning model training.
Differentially Private Set Representations
·1424 words·7 mins· loading · loading
AI Generated AI Theory Privacy 🏢 Google
Differentially private set representations achieve optimal privacy-utility tradeoffs with exponentially smaller error than prior histogram methods.
Differentially Private Reinforcement Learning with Self-Play
·347 words·2 mins· loading · loading
AI Theory Privacy 🏢 UC San Diego
This paper presents DP-Nash-VI, a novel algorithm ensuring trajectory-wise privacy in multi-agent reinforcement learning, achieving near-optimal regret bounds under both joint and local differential p…
Differentially Private Optimization with Sparse Gradients
·1282 words·7 mins· loading · loading
AI Theory Privacy 🏢 Google Research
This paper presents new, nearly optimal differentially private algorithms for handling sparse gradients, significantly improving efficiency and scalability in large embedding models.
Differentially Private Graph Diffusion with Applications in Personalized PageRanks
·1969 words·10 mins· loading · loading
AI Theory Privacy 🏢 Georgia Institute of Technology
This paper introduces a novel differentially private graph diffusion framework ensuring edge-level privacy, significantly improving utility-privacy trade-offs for personalized PageRank computation.
Differentially Private Equivalence Testing for Continuous Distributions and Applications
·583 words·3 mins· loading · loading
AI Generated AI Theory Privacy 🏢 Bar-Ilan University
First differentially private algorithm for testing equivalence between continuous distributions, enabling privacy-preserving comparisons of sensitive data.
Differential Privacy in Scalable General Kernel Learning via $K$-means Nystr{"o}m Random Features
·1468 words·7 mins· loading · loading
AI Generated AI Theory Privacy 🏢 KAIST
Differentially private scalable kernel learning is achieved via a novel DP K-means Nyström method, enabling efficient and accurate model training for general kernels while safeguarding privacy.
Differentiable Structure Learning with Partial Orders
·2110 words·10 mins· loading · loading
AI Theory Causality 🏢 University of Science and Technology of China
This research introduces a novel plug-and-play module that efficiently integrates prior partial order constraints into differentiable structure learning, significantly improving structure recovery qua…
Diffeomorphic interpolation for efficient persistence-based topological optimization
·2900 words·14 mins· loading · loading
AI Generated AI Theory Optimization 🏢 DataShape
Diffeomorphic interpolation boosts topological optimization by transforming sparse gradients into smooth vector fields, enabling efficient large-scale point cloud optimization and black-box autoencode…