Machine Learning
The Map Equation Goes Neural: Mapping Network Flows with Graph Neural Networks
·3312 words·16 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 University of Zurich
Neuromap leverages graph neural networks to optimize the map equation for community detection, achieving competitive performance and automatically determining the optimal number of clusters.
The Many Faces of Optimal Weak-to-Strong Learning
·1344 words·7 mins·
loading
·
loading
Machine Learning
Optimization
🏢 Aarhus University
A new, surprisingly simple boosting algorithm achieves provably optimal sample complexity and outperforms existing algorithms on large datasets.
The Limits of Transfer Reinforcement Learning with Latent Low-rank Structure
·1762 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Cornell University
This paper presents computationally efficient transfer reinforcement learning algorithms that remove the dependence on state/action space sizes while achieving minimax optimality.
The Ladder in Chaos: Improving Policy Learning by Harnessing the Parameter Evolving Path in A Low-dimensional Space
·2918 words·14 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 College of Intelligence and Computing, Tianjin University
Deep RL policy learning is improved by identifying and boosting key parameter update directions using a novel temporal SVD analysis, leading to more efficient and effective learning.
The Iterative Optimal Brain Surgeon: Faster Sparse Recovery by Leveraging Second-Order Information
·1778 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Institute of Science and Technology Austria
I-OBS, a novel family of sparse recovery algorithms leveraging second-order information, achieves faster convergence rates for sparse DNNs, validated by large-scale experiments.
The Importance of Being Scalable: Improving the Speed and Accuracy of Neural Network Interatomic Potentials Across Chemical Domains
·1885 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 UC Berkeley
ESCAIP, a novel neural network architecture, dramatically boosts the speed and accuracy of atomic simulations by leveraging attention mechanisms, enabling efficient large-scale modeling across diverse…
The Implicit Bias of Gradient Descent on Separable Multiclass Data
·1300 words·7 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Michigan
Researchers extended implicit bias theory to multiclass classification using a novel framework, proving that gradient descent prefers simple solutions even with complex alternatives.
The Impact of Geometric Complexity on Neural Collapse in Transfer Learning
·1870 words·9 mins·
loading
·
loading
Machine Learning
Transfer Learning
🏢 Google Research
Lowering a neural network’s geometric complexity during pre-training enhances neural collapse and improves transfer learning, especially in few-shot scenarios.
The High Line: Exact Risk and Learning Rate Curves of Stochastic Adaptive Learning Rate Algorithms
·1819 words·9 mins·
loading
·
loading
Machine Learning
Optimization
🏢 McGill University
Researchers developed a framework for analyzing stochastic adaptive learning rate algorithms, providing exact risk and learning rate curves, revealing the importance of data covariance and uncovering …
The Feature Speed Formula: a flexible approach to scale hyper-parameters of deep neural networks
·1538 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Institute of Mathematics, EPFL
New ‘Feature Speed Formula’ predicts & controls deep learning’s hierarchical feature learning by linking hyperparameter tuning to the angle between feature updates and backward pass.
The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning
·2452 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Oxford
Offline model-based RL methods fail as dynamics models improve; this paper reveals the ’edge-of-reach’ problem causing this and introduces RAVL, a simple solution ensuring robust performance.
The Dormant Neuron Phenomenon in Multi-Agent Reinforcement Learning Value Factorization
·2766 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Xiamen University
ReBorn revitalizes multi-agent reinforcement learning by tackling dormant neurons, boosting network expressivity and learning efficiency.
The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection
·2465 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Tencent AI Lab
Researchers found that superior OOD detection performance comes at the cost of reduced generalization. Their novel Decoupled Uncertainty Learning (DUL) algorithm harmonizes OOD detection and generali…
The Benefits of Balance: From Information Projections to Variance Reduction
·1859 words·9 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
🏢 University of Washington
Data balancing in foundation models surprisingly reduces variance, improving model training and performance.
TFGDA: Exploring Topology and Feature Alignment in Semi-supervised Graph Domain Adaptation through Robust Clustering
·1822 words·9 mins·
loading
·
loading
Machine Learning
Transfer Learning
🏢 Zhejiang University
TFGDA: Leveraging graph topology and feature alignment for superior semi-supervised domain adaptation.
Test-Time Adaptation Induces Stronger Accuracy and Agreement-on-the-Line
·2874 words·14 mins·
loading
·
loading
Machine Learning
Few-Shot Learning
🏢 Carnegie Mellon University
Test-time adaptation strengthens the linear correlation between in- and out-of-distribution accuracy, enabling precise OOD performance prediction and hyperparameter optimization without labeled OOD da…
Test-time Adaptation in Non-stationary Environments via Adaptive Representation Alignment
·2451 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Representation Learning
🏢 Stanford University
Ada-ReAlign: a novel algorithm for continual test-time adaptation that leverages non-stationary representation learning to effectively align unlabeled data streams with source data, enhancing model ad…
Test Where Decisions Matter: Importance-driven Testing for Deep Reinforcement Learning
·3658 words·18 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Graz University of Technology
Prioritize crucial decisions in deep RL policy testing with a novel model-based method for rigorous state importance ranking, enabling efficient safety and performance verification.
Temporal-Difference Learning Using Distributed Error Signals
·2668 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Toronto
Artificial Dopamine (AD) algorithm achieves comparable performance to backpropagation methods in complex RL tasks by using only synchronously distributed per-layer TD errors, demonstrating the suffici…
Task-recency bias strikes back: Adapting covariances in Exemplar-Free Class Incremental Learning
·2323 words·11 mins·
loading
·
loading
Machine Learning
Continual Learning
🏢 Warsaw University of Technology
AdaGauss tackles task-recency bias in exemplar-free class incremental learning by adapting class covariances and introducing an anti-collapse loss, achieving state-of-the-art results.