Skip to main content

Machine Learning

Task-Agnostic Machine-Learning-Assisted Inference
·1432 words·7 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏒 University of Wisconsin-Madison
PSPS: a novel task-agnostic framework enables valid and efficient ML-assisted statistical inference for virtually any task, simply using summary statistics from existing analysis routines!
Task Confusion and Catastrophic Forgetting in Class-Incremental Learning: A Mathematical Framework for Discriminative and Generative Modelings
·1620 words·8 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 Queen's University
Researchers unveil the Infeasibility Theorem, proving optimal class-incremental learning is impossible with discriminative models due to task confusion, and the Feasibility Theorem, showing generative…
Taming Heavy-Tailed Losses in Adversarial Bandits and the Best-of-Both-Worlds Setting
·418 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 Virginia Tech
This paper proposes novel algorithms achieving near-optimal regret in adversarial and logarithmic regret in stochastic multi-armed bandit settings with heavy-tailed losses, relaxing strong assumptions…
Taming Cross-Domain Representation Variance in Federated Prototype Learning with Heterogeneous Data Domains
·3467 words·17 mins· loading · loading
AI Generated Machine Learning Federated Learning 🏒 University of Florida
FedPLVM tames cross-domain variance in federated prototype learning using dual-level clustering and an a-sparsity loss, achieving superior performance.
Taming 'data-hungry' reinforcement learning? Stability in continuous state-action spaces
·358 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 New York University
Reinforcement learning achieves unprecedented fast convergence rates in continuous state-action spaces by leveraging novel stability properties of Markov Decision Processes.
Take A Shortcut Back: Mitigating the Gradient Vanishing for Training Spiking Neural Networks
·1272 words·6 mins· loading · loading
Machine Learning Deep Learning 🏒 Peking University
Shortcut back-propagation and an evolutionary training framework conquer gradient vanishing in spiking neural networks, drastically improving training and achieving state-of-the-art accuracy.
TabEBM: A Tabular Data Augmentation Method with Distinct Class-Specific Energy-Based Models
·8456 words·40 mins· loading · loading
AI Generated Machine Learning Generative Models 🏒 University of Cambridge
TabEBM: Class-specific EBMs boost tabular data augmentation, improving classification accuracy, especially on small datasets, by generating high-quality synthetic data.
Symmetry-Informed Governing Equation Discovery
·2189 words·11 mins· loading · loading
Machine Learning Deep Learning 🏒 UC San Diego
Leveraging symmetry in automated equation discovery improves accuracy and simplicity of learned governing equations, enhancing robustness against noise and achieving higher success rates across divers…
Symmetric Linear Bandits with Hidden Symmetry
·1466 words·7 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Warwick
Researchers unveil a novel algorithm for high-dimensional symmetric linear bandits, achieving a regret bound of O(d^(2/3)T^(2/3)log(d)), surpassing limitations of existing approaches that assume expli…
Swift Sampler: Efficient Learning of Sampler by 10 Parameters
·2624 words·13 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Washington
Swift Sampler (SS) automates the learning of efficient data samplers for deep learning, achieving significant performance gains (e.g., 1.5% on ImageNet) with minimal computational cost using only 10 p…
Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling
·2206 words·11 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 Tencent Hunyuan
Deep learning’s Adam-style optimizers exhibit a surprising surge phenomenon: optimal learning rates initially increase, then decrease, before converging to a non-zero value as batch size grows.
Supra-Laplacian Encoding for Transformer on Dynamic Graphs
·2790 words·14 mins· loading · loading
Machine Learning Deep Learning 🏒 Conservatoire National Des Arts Et Métiers
SLATE: Supra-Laplacian encoding for spatio-temporal Transformers achieves state-of-the-art dynamic link prediction by innovatively using a multi-layer graph representation and a unique cross-attention…
Supervised Kernel Thinning
·1753 words·9 mins· loading · loading
AI Generated Machine Learning Supervised Learning 🏒 Cornell University
Supervised Kernel Thinning accelerates kernel regression by cleverly compressing data, achieving quadratic speedups in training and inference with minimal accuracy loss.
Super Consistency of Neural Network Landscapes and Learning Rate Transfer
·3859 words·19 mins· loading · loading
Machine Learning Deep Learning 🏒 ETH Zurich
Neural network hyperparameter transferability across vastly different model sizes is achieved via a newly discovered property called ‘Super Consistency’ of loss landscapes.
Subwords as Skills: Tokenization for Sparse-Reward Reinforcement Learning
·2209 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 TTI-Chicago
This paper introduces Subwords as Skills (SaS), a fast and efficient skill extraction method for sparse-reward reinforcement learning that uses tokenization. SaS enables 1000x faster skill extraction…
SubgDiff: A Subgraph Diffusion Model to Improve Molecular Representation Learning
·2691 words·13 mins· loading · loading
Machine Learning Representation Learning 🏒 Hong Kong University of Science and Technology
SubgDiff enhances molecular representation learning by incorporating substructural information into a diffusion model framework, achieving superior performance in molecular force predictions.
Sub-optimal Experts mitigate Ambiguity in Inverse Reinforcement Learning
·2049 words·10 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏒 Politecnico Di Milano
Sub-optimal expert data improves Inverse Reinforcement Learning by significantly reducing ambiguity in reward function estimation.
Style Adaptation and Uncertainty Estimation for Multi-Source Blended-Target Domain Adaptation
·1973 words·10 mins· loading · loading
Machine Learning Transfer Learning 🏒 South China Normal University
SAUE: A novel multi-source blended-target domain adaptation approach using style adaptation and uncertainty estimation to improve model robustness and accuracy.
Structured Matrix Basis for Multivariate Time Series Forecasting with Interpretable Dynamics
·2313 words·11 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 Harbin Institute of Technology
Sumba: a novel forecasting model achieves up to 8.5% improvement by using a structured matrix basis to generate dynamic spatial structures with lower variance and better interpretability.
Structural Inference of Dynamical Systems with Conjoined State Space Models
·2686 words·13 mins· loading · loading
Machine Learning Deep Learning 🏒 University of Luxembourg
SICSM, a novel framework, integrates selective SSMs and GFNs to accurately infer complex dynamical system structures from irregularly sampled, partially observed trajectories.