Skip to main content

Machine Learning

On the Stability and Generalization of Meta-Learning
·1358 words·7 mins· loading · loading
Machine Learning Meta Learning 🏒 Johns Hopkins University
This paper introduces uniform meta-stability for meta-learning, providing tighter generalization bounds for convex and weakly-convex problems, addressing computational limitations of existing algorith…
On the Scalability of GNNs for Molecular Graphs
·2680 words·13 mins· loading · loading
Machine Learning Deep Learning 🏒 Valence Labs
Giant leap in molecular GNNs! MolGPS, a new foundation model, achieves state-of-the-art performance on molecular property prediction by leveraging massive datasets and demonstrating the scalability o…
On the Scalability of Certified Adversarial Robustness with Generated Data
·2448 words·12 mins· loading · loading
Machine Learning Deep Learning 🏒 Machine Learning and Data Analytics Lab, FAU Erlangen Nürnberg, Germany
Boosting certified robustness of machine learning models by 3-4% using generated data from diffusion models!
On the Role of Information Structure in Reinforcement Learning for Partially-Observable Sequential Teams and Games
·2014 words·10 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 Yale University
New reinforcement learning model clarifies the role of information structure in partially-observable sequential decision-making problems, proving an upper bound on learning complexity.
On the Optimal Time Complexities in Decentralized Stochastic Asynchronous Optimization
·2010 words·10 mins· loading · loading
Machine Learning Optimization 🏒 KAUST AIRI
Fragile SGD & Amelie SGD achieve near-optimal speed in decentralized asynchronous optimization, handling diverse worker & communication speeds.
On the Necessity of Collaboration for Online Model Selection with Decentralized Data
·1427 words·7 mins· loading · loading
Machine Learning Federated Learning 🏒 School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen
Federated online model selection needs collaboration only when clients have limited computing power; otherwise, independent learning suffices.
On the Minimax Regret for Contextual Linear Bandits and Multi-Armed Bandits with Expert Advice
·360 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Tokyo
This paper provides novel algorithms and matching lower bounds for multi-armed bandits with expert advice and contextual linear bandits, resolving open questions and advancing theoretical understandin…
On the Limitations of Fractal Dimension as a Measure of Generalization
·1917 words·9 mins· loading · loading
Machine Learning Deep Learning 🏒 University of Oxford
Fractal dimension, while showing promise, fails to consistently predict neural network generalization due to hyperparameter influence and adversarial initializations; prompting further research.
On the Identifiability of Hybrid Deep Generative Models: Meta-Learning as a Solution
·1881 words·9 mins· loading · loading
Machine Learning Meta Learning 🏒 Rochester Institute of Technology
Meta-learning solves hybrid deep generative model unidentifiability!
On the Efficiency of ERM in Feature Learning
·316 words·2 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Toronto
ERM’s efficiency in feature learning surprisingly remains high even with massive feature maps; its excess risk asymptotically matches an oracle procedure’s, implying potential for streamlined feature-…
On the Curses of Future and History in Future-dependent Value Functions for Off-policy Evaluation
·299 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Illinois Urbana-Champaign
This paper tackles the ‘curse of horizon’ in off-policy evaluation for partially observable Markov decision processes (POMDPs) by proposing novel coverage assumptions, enabling polynomial estimation e…
On the Convergence of Loss and Uncertainty-based Active Learning Algorithms
·2336 words·11 mins· loading · loading
AI Generated Machine Learning Active Learning 🏒 Meta
New active learning algorithm, Adaptive-Weight Sampling (AWS), achieves faster convergence with theoretical guarantees, improving data efficiency for machine learning.
On the Complexity of Teaching a Family of Linear Behavior Cloning Learners
·1819 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Washington
A novel algorithm, TIE, optimally teaches a family of linear behavior cloning learners, achieving instance-optimal teaching dimension while providing efficient approximation for larger action spaces.
On the Complexity of Learning Sparse Functions with Statistical and Gradient Queries
·484 words·3 mins· loading · loading
Machine Learning Deep Learning 🏒 Toyota Technological Institute at Chicago
Learning sparse functions efficiently with gradient methods is challenging; this paper introduces Differentiable Learning Queries (DLQ) to precisely characterize gradient query complexity, revealing s…
On Sampling Strategies for Spectral Model Sharding
·1797 words·9 mins· loading · loading
Machine Learning Federated Learning 🏒 Qualcomm AI Research
Two novel sampling strategies for spectral model sharding in federated learning minimize approximation error and create unbiased estimators, improving performance on various datasets.
On Divergence Measures for Training GFlowNets
·2110 words·10 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏒 School of Applied Mathematics
Researchers enhanced Generative Flow Network training by introducing variance-reducing control variates for divergence-based learning objectives, accelerating convergence and improving accuracy.
On conditional diffusion models for PDE simulations
·5766 words·28 mins· loading · loading
Machine Learning Deep Learning 🏒 University of Cambridge
This paper introduces novel autoregressive sampling and hybrid training strategies for score-based diffusion models, significantly boosting PDE forecasting and assimilation accuracy.
On $f$-Divergence Principled Domain Adaptation: An Improved Framework
·1963 words·10 mins· loading · loading
Machine Learning Transfer Learning 🏒 Tongji University
Improved unsupervised domain adaptation framework achieves superior performance via refined f-divergence and novel f-domain discrepancy, enabling faster algorithms and tighter generalization bounds.
Oja's Algorithm for Streaming Sparse PCA
·382 words·2 mins· loading · loading
Machine Learning Unsupervised Learning 🏒 University of Texas at Austin
Oja’s algorithm achieves minimax optimal error rates for streaming sparse PCA using a simple single-pass thresholding method, requiring only O(d) space and O(nd) time.
Offline Reinforcement Learning with OOD State Correction and OOD Action Suppression
·2487 words·12 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 Tsinghua University
Offline RL agents often fail in real-world scenarios due to unseen test states. SCAS, a novel method, simultaneously corrects OOD states to high-value, in-distribution states and suppresses risky OOD …