Skip to main content

AI Theory

On the Ability of Developers' Training Data Preservation of Learnware
·449 words·3 mins· loading · loading
AI Theory Privacy 🏢 Nanjing University
Learnware systems enable model reuse; this paper proves RKME specifications protect developers’ training data while enabling effective model identification.
On Statistical Rates and Provably Efficient Criteria of Latent Diffusion Transformers (DiTs)
·402 words·2 mins· loading · loading
AI Theory Generalization 🏢 Northwestern University
Latent Diffusion Transformers (DiTs) achieve almost-linear time training and inference through low-rank gradient approximations and efficient criteria, overcoming high dimensionality challenges.
On Sparse Canonical Correlation Analysis
·2005 words·10 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of Tennessee
This paper presents novel, efficient algorithms and formulations for Sparse Canonical Correlation Analysis (SCCA), a method that improves the interpretability of traditional CCA. SCCA is especially us…
On Socially Fair Low-Rank Approximation and Column Subset Selection
·363 words·2 mins· loading · loading
AI Generated AI Theory Fairness 🏢 UC Berkeley
This paper reveals the surprising computational hardness of achieving fairness in low-rank approximation while offering efficient approximation algorithms.
On provable privacy vulnerabilities of graph representations
·4156 words·20 mins· loading · loading
AI Theory Privacy 🏢 Ant Group
Graph representation learning’s structural vulnerabilities are proven and mitigated via noisy aggregation, revealing crucial privacy-utility trade-offs.
On Neural Networks as Infinite Tree-Structured Probabilistic Graphical Models
·2116 words·10 mins· loading · loading
AI Theory Interpretability 🏢 Duke University
DNNs are powerful but lack the clear semantics of PGMs. This paper innovatively constructs infinite tree-structured PGMs that exactly correspond to DNNs, revealing that DNN forward propagation approxi…
On Mesa-Optimization in Autoregressively Trained Transformers: Emergence and Capability
·2131 words·11 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Gaoling School of Artificial Intelligence, Renmin University of China
Autoregressively trained transformers surprisingly learn algorithms during pretraining, enabling in-context learning; this paper reveals when and why this ‘mesa-optimization’ happens.
On Feature Learning in Structured State Space Models
·1624 words·8 mins· loading · loading
AI Theory Generalization 🏢 AGI Foundations
Unlocking the scaling secrets of structured state-space models, this research identifies novel scaling rules for improved stability, generalization, and hyperparameter transferability, revolutionizing…
On Differentially Private U Statistics
·403 words·2 mins· loading · loading
AI Theory Privacy 🏢 UC San Diego
New algorithms achieve near-optimal differentially private U-statistic estimation, significantly improving accuracy over existing methods.
On Differentially Private Subspace Estimation in a Distribution-Free Setting
·418 words·2 mins· loading · loading
AI Theory Privacy 🏢 Georgetown University
This paper presents novel measures quantifying data easiness for DP subspace estimation, supporting them with improved upper and lower bounds and a practical algorithm.
On Convergence of Adam for Stochastic Optimization under Relaxed Assumptions
·308 words·2 mins· loading · loading
AI Theory Optimization 🏢 Zhejiang University
Adam optimizer achieves near-optimal convergence in non-convex scenarios with unbounded gradients and relaxed noise assumptions, improving its theoretical understanding and practical application.
On Causal Discovery in the Presence of Deterministic Relations
·2932 words·14 mins· loading · loading
AI Generated AI Theory Causality 🏢 Mohamed Bin Zayed University of Artificial Intelligence
DGES, a novel framework, efficiently detects & handles deterministic relations in causal discovery, enhancing accuracy and scalability for real-world applications.
Off-policy estimation with adaptively collected data: the power of online learning
·240 words·2 mins· loading · loading
AI Generated AI Theory Causality 🏢 University of Chicago
This paper develops novel finite-sample bounds for off-policy linear treatment effect estimation with adaptively collected data, proposing online learning algorithms to improve estimation accuracy and…
Not so griddy: Internal representations of RNNs path integrating more than one agent
·2491 words·12 mins· loading · loading
AI Theory Representation Learning 🏢 Johns Hopkins Applied Physics Laboratory
RNNs trained on dual-agent path integration develop distinct internal representations compared to single-agent models, exhibiting weaker grid cell responses and enhanced border/band cell activity, wit…
Nonparametric Instrumental Variable Regression through Stochastic Approximate Gradients
·1348 words·7 mins· loading · loading
AI Theory Causality 🏢 Columbia University
SAGD-IV: a novel functional stochastic gradient descent algorithm for stable nonparametric instrumental variable regression, excelling in handling binary outcomes and various loss functions.
Non-geodesically-convex optimization in the Wasserstein space
·332 words·2 mins· loading · loading
AI Theory Optimization 🏢 Department of Computer Science, University of Helsinki
A novel semi Forward-Backward Euler scheme provides convergence guarantees for non-geodesically-convex optimization in Wasserstein space, advancing both sampling and optimization.
Non-asymptotic Convergence of Training Transformers for Next-token Prediction
·361 words·2 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Penn State University
This paper reveals how a one-layer transformer’s training converges for next-token prediction, showing sub-linear convergence for both layers and shedding light on its surprising generalization abilit…
Noisy Dual Mirror Descent: A Near Optimal Algorithm for Jointly-DP Convex Resource Allocation
·2148 words·11 mins· loading · loading
AI Generated AI Theory Privacy 🏢 Nanyang Business School, Nanyang Technological University
Near-optimal algorithm for private resource allocation is introduced, achieving improved accuracy and privacy guarantees.
Noise-Aware Differentially Private Regression via Meta-Learning
·3336 words·16 mins· loading · loading
AI Generated AI Theory Privacy 🏢 University of Helsinki
Meta-learning and differential privacy combine to enable accurate, well-calibrated private regression, even with limited data, via the novel DPConvCNP model.
No-Regret M${}^{ atural}$-Concave Function Maximization: Stochastic Bandit Algorithms and NP-Hardness of Adversarial Full-Information Setting
·1615 words·8 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Hokkaido University
This paper reveals efficient stochastic bandit algorithms for maximizing M-concave functions and proves NP-hardness for adversarial full-information settings.