Machine Learning
Maximum Entropy Reinforcement Learning via Energy-Based Normalizing Flow
·2836 words·14 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 NVIDIA Corporation
MEow, a novel MaxEnt RL framework, achieves superior performance by unifying policy evaluation and improvement steps, enabling exact soft value function calculation without Monte Carlo approximation.
Markovian Flow Matching: Accelerating MCMC with Continuous Normalizing Flows
·2056 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Lancaster University
Adaptive MCMC with CNFs accelerates probabilistic inference by combining local and flow-informed transition kernels, achieving state-of-the-art results efficiently.
Making Offline RL Online: Collaborative World Models for Offline Visual Reinforcement Learning
·3140 words·15 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 MoE Key Lab of Artificial Intelligence
CoWorld: a novel model-based RL approach tackles offline visual RL challenges by using online simulators as testbeds, enabling flexible value estimation & mitigating overestimation bias for effective …
Make Continual Learning Stronger via C-Flat
·2055 words·10 mins·
loading
·
loading
Machine Learning
Continual Learning
🏢 Tsinghua University
Boost continual learning with C-Flat: a novel, one-line-code optimizer creating flatter loss landscapes for enhanced stability and generalization across various continual learning scenarios.
Maia-2: A Unified Model for Human-AI Alignment in Chess
·2577 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Toronto
Maia-2: A unified model for human-AI alignment in chess, coherently captures human play across skill levels, significantly improving AI-human alignment and paving the way for AI-guided teaching.
MADiff: Offline Multi-agent Learning with Diffusion Models
·2719 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Shanghai Jiao Tong University
MADIFF: Offline multi-agent learning uses attention-based diffusion models to achieve effective coordination and teammate modeling, outperforming existing methods.
Lower Bounds of Uniform Stability in Gradient-Based Bilevel Algorithms for Hyperparameter Optimization
·1822 words·9 mins·
loading
·
loading
Machine Learning
Optimization
🏢 Gaoling School of Artificial Intelligence, Renmin University of China
This paper establishes tight lower bounds for the uniform stability of gradient-based bilevel programming algorithms used for hyperparameter optimization, resolving a key open problem regarding the ti…
Lower Bounds and Optimal Algorithms for Non-Smooth Convex Decentralized Optimization over Time-Varying Networks
·1231 words·6 mins·
loading
·
loading
Machine Learning
Optimization
🏢 Yandex Research
First optimal algorithms matching lower bounds for non-smooth convex decentralized optimization over time-varying networks are presented, substantially improving theoretical performance.
Low-Rank Optimal Transport through Factor Relaxation with Latent Coupling
·2606 words·13 mins·
loading
·
loading
Machine Learning
Optimization
🏢 Princeton University
FRLC: a novel algorithm for low-rank optimal transport using latent coupling, enabling faster computation and better interpretability for diverse applications.
Low Precision Local Training is Enough for Federated Learning
·2011 words·10 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 Fudan University
Low-precision local training, surprisingly, is sufficient for accurate federated learning, significantly reducing communication and computation costs.
Lorentz-Equivariant Geometric Algebra Transformers for High-Energy Physics
·1502 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Heidelberg University
Lorentz Geometric Algebra Transformer (L-GATr): A novel, scalable architecture for high-energy physics, achieving high-precision, data-efficient learning and outperforming existing methods on regressi…
Long-range Meta-path Search on Large-scale Heterogeneous Graphs
·2383 words·12 mins·
loading
·
loading
Machine Learning
Representation Learning
🏢 Huazhong University of Science and Technology
LMSPS: a novel framework efficiently leverages long-range dependencies in large heterogeneous graphs by dynamically identifying effective meta-paths, mitigating computational costs and over-smoothing.
Localizing Memorization in SSL Vision Encoders
·4999 words·24 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
🏢 CISPA, Helmholtz Center for Information Security
SSL vision encoders, while trained on massive datasets, surprisingly memorize individual data points. This paper introduces novel methods to precisely pinpoint this memorization within encoders at bot…
Local Superior Soups: A Catalyst for Model Merging in Cross-Silo Federated Learning
·3305 words·16 mins·
loading
·
loading
AI Generated
Machine Learning
Federated Learning
🏢 University of British Columbia
Local Superior Soups (LSS) significantly accelerates federated learning by efficiently merging pre-trained models, drastically cutting communication rounds without sacrificing accuracy.
Local Linearity: the Key for No-regret Reinforcement Learning in Continuous MDPs
·451 words·3 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Politecnico Di Milano
CINDERELLA: a new algorithm achieves state-of-the-art no-regret bounds for continuous RL problems by exploiting local linearity.
Local Curvature Smoothing with Stein's Identity for Efficient Score Matching
·1604 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 LY Corporation
LCSS, a novel score-matching method, enables efficient and high-quality image generation in score-based diffusion models by using Stein’s identity to bypass the computationally expensive Jacobian trac…
Local Anti-Concentration Class: Logarithmic Regret for Greedy Linear Contextual Bandit
·2759 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Columbia University
Greedy algorithms for linear contextual bandits achieve poly-logarithmic regret under the novel Local Anti-Concentration condition, expanding applicable distributions beyond Gaussians and uniforms.
Local and Adaptive Mirror Descents in Extensive-Form Games
·321 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 CREST - FairPlay, ENSAE Paris
LocalOMD: Adaptive OMD in extensive-form games achieves near-optimal sample complexity by using fixed sampling and local updates, reducing variance and generalizing well.
LM-HT SNN: Enhancing the Performance of SNN to ANN Counterpart through Learnable Multi-hierarchical Threshold Model
·1826 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Peking University
LM-HT SNN: A learnable multi-hierarchical threshold model dramatically improves SNN performance, achieving near-ANN accuracy through dynamic current regulation and seamless ANN-SNN conversion.
LLM-AutoDA: Large Language Model-Driven Automatic Data Augmentation for Long-tailed Problems
·1891 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Science and Technology of China (USTC)
LLM-AutoDA: Automating data augmentation for long-tailed learning using large language models, significantly boosting model performance.