Posters
2024
On the Adversarial Robustness of Benjamini Hochberg
·1747 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 Operations Research Department Naval Postgraduate School
Even a few data changes can break the Benjamini-Hochberg (BH) procedure, a widely used multiple testing method, highlighting a critical vulnerability.
On the Ability of Developers' Training Data Preservation of Learnware
·449 words·3 mins·
loading
·
loading
AI Theory
Privacy
🏢 Nanjing University
Learnware systems enable model reuse; this paper proves RKME specifications protect developers’ training data while enabling effective model identification.
On Statistical Rates and Provably Efficient Criteria of Latent Diffusion Transformers (DiTs)
·402 words·2 mins·
loading
·
loading
AI Theory
Generalization
🏢 Northwestern University
Latent Diffusion Transformers (DiTs) achieve almost-linear time training and inference through low-rank gradient approximations and efficient criteria, overcoming high dimensionality challenges.
On Sparse Canonical Correlation Analysis
·2005 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 University of Tennessee
This paper presents novel, efficient algorithms and formulations for Sparse Canonical Correlation Analysis (SCCA), a method that improves the interpretability of traditional CCA. SCCA is especially us…
On Softmax Direct Preference Optimization for Recommendation
·1530 words·8 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 National University of Singapore
Softmax-DPO boosts LM-based recommender performance by directly optimizing for personalized ranking using a novel loss function that incorporates multiple negative samples, significantly outperforming…
On Socially Fair Low-Rank Approximation and Column Subset Selection
·363 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
🏢 UC Berkeley
This paper reveals the surprising computational hardness of achieving fairness in low-rank approximation while offering efficient approximation algorithms.
On scalable oversight with weak LLMs judging strong LLMs
·5158 words·25 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Google DeepMind
Weak LLMs can accurately supervise strong LLMs via debate, outperforming simpler consultancy methods, especially in information-asymmetric tasks.
On Sampling Strategies for Spectral Model Sharding
·1797 words·9 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 Qualcomm AI Research
Two novel sampling strategies for spectral model sharding in federated learning minimize approximation error and create unbiased estimators, improving performance on various datasets.
On provable privacy vulnerabilities of graph representations
·4156 words·20 mins·
loading
·
loading
AI Theory
Privacy
🏢 Ant Group
Graph representation learning’s structural vulnerabilities are proven and mitigated via noisy aggregation, revealing crucial privacy-utility trade-offs.
On Neural Networks as Infinite Tree-Structured Probabilistic Graphical Models
·2116 words·10 mins·
loading
·
loading
AI Theory
Interpretability
🏢 Duke University
DNNs are powerful but lack the clear semantics of PGMs. This paper innovatively constructs infinite tree-structured PGMs that exactly correspond to DNNs, revealing that DNN forward propagation approxi…
On Mesa-Optimization in Autoregressively Trained Transformers: Emergence and Capability
·2131 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Gaoling School of Artificial Intelligence, Renmin University of China
Autoregressively trained transformers surprisingly learn algorithms during pretraining, enabling in-context learning; this paper reveals when and why this ‘mesa-optimization’ happens.
On Learning Multi-Modal Forgery Representation for Diffusion Generated Video Detection
·2133 words·11 mins·
loading
·
loading
Computer Vision
Video Understanding
🏢 Shanghai Jiao Tong University
MM-Det, a novel algorithm, uses multimodal learning and spatiotemporal attention to detect diffusion-generated videos, achieving state-of-the-art performance on the new DVF dataset.
On improved Conditioning Mechanisms and Pre-training Strategies for Diffusion Models
·3235 words·16 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 FAIR at Meta
Researchers achieve state-of-the-art image generation by disentangling semantic and control metadata in diffusion models and optimizing pre-training across resolutions.
On Giant's Shoulders: Effortless Weak to Strong by Dynamic Logits Fusion
·2220 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Huazhong University of Science and Technology
Effortlessly boost large language model performance by dynamically fusing knowledge from smaller, task-specific models – achieving near full fine-tuning results with minimal computational cost!
On Feature Learning in Structured State Space Models
·1624 words·8 mins·
loading
·
loading
AI Theory
Generalization
🏢 AGI Foundations
Unlocking the scaling secrets of structured state-space models, this research identifies novel scaling rules for improved stability, generalization, and hyperparameter transferability, revolutionizing…
On Divergence Measures for Training GFlowNets
·2110 words·10 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 School of Applied Mathematics
Researchers enhanced Generative Flow Network training by introducing variance-reducing control variates for divergence-based learning objectives, accelerating convergence and improving accuracy.
On Differentially Private U Statistics
·403 words·2 mins·
loading
·
loading
AI Theory
Privacy
🏢 UC San Diego
New algorithms achieve near-optimal differentially private U-statistic estimation, significantly improving accuracy over existing methods.
On Differentially Private Subspace Estimation in a Distribution-Free Setting
·418 words·2 mins·
loading
·
loading
AI Theory
Privacy
🏢 Georgetown University
This paper presents novel measures quantifying data easiness for DP subspace estimation, supporting them with improved upper and lower bounds and a practical algorithm.
On Convergence of Adam for Stochastic Optimization under Relaxed Assumptions
·308 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 Zhejiang University
Adam optimizer achieves near-optimal convergence in non-convex scenarios with unbounded gradients and relaxed noise assumptions, improving its theoretical understanding and practical application.
On conditional diffusion models for PDE simulations
·5766 words·28 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Cambridge
This paper introduces novel autoregressive sampling and hybrid training strategies for score-based diffusion models, significantly boosting PDE forecasting and assimilation accuracy.