Machine Learning
Zero-Shot Reinforcement Learning from Low Quality Data
·4722 words·23 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ University of Cambridge
Zero-shot RL struggles with low-quality data; this paper introduces conservative algorithms that significantly boost performance on such data without sacrificing performance on high-quality data.
Your Diffusion Model is Secretly a Noise Classifier and Benefits from Contrastive Training
·2424 words·12 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
π’ UC Riverside
Diffusion models benefit from contrastive training, improving sample quality and speed by addressing poor denoiser estimation in out-of-distribution regions.
Your contrastive learning problem is secretly a distribution alignment problem
·381 words·2 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
π’ University of Toronto
Contrastive learning is reframed as a distribution alignment problem, leading to a flexible framework (GCA) that improves representation learning with unbalanced optimal transport.
You Donβt Need Domain-Specific Data Augmentations When Scaling Self-Supervised Learning
·2133 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Self-Supervised Learning
π’ FAIR at Meta
Self-supervised learning’s reliance on complex data augmentations is challenged; a large-scale study shows comparable performance using only cropping, suggesting dataset size is more important than au…
Worst-Case Offline Reinforcement Learning with Arbitrary Data Support
·450 words·3 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ IBM Research
Worst-case offline RL guarantees near-optimal policy performance without data support assumptions, achieving a sample complexity bound of O(Ξ΅β»Β²).
Why Go Full? Elevating Federated Learning Through Partial Network Updates
·3064 words·15 mins·
loading
·
loading
AI Generated
Machine Learning
Federated Learning
π’ Beihang University
FedPart boosts federated learning by updating only parts of the network, solving the layer mismatch problem, and achieving faster convergence with higher accuracy.
When Your AIs Deceive You: Challenges of Partial Observability in Reinforcement Learning from Human Feedback
·2699 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ University of Amsterdam
RLHF’s reliance on fully observable environments is challenged: human feedback, often partial, leads to deceptive AI behavior (inflation & overjustification).
When to Sense and Control? A Time-adaptive Approach for Continuous-Time RL
·2003 words·10 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ ETH Zurich
TACOS: A novel time-adaptive RL framework drastically reduces interactions in continuous-time systems while improving performance, offering both model-free and model-based algorithms.
What Matters in Graph Class Incremental Learning? An Information Preservation Perspective
·3421 words·17 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ College of Intelligence and Computing, Tianjin University
GSIP framework mitigates catastrophic forgetting in graph class incremental learning by preserving crucial graph information, achieving a 10% improvement in forgetting metrics.
What Makes Partial-Label Learning Algorithms Effective?
·1666 words·8 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
π’ Southeast University
Unlocking Partial-Label Learning: A new study reveals surprisingly simple design principles for highly accurate algorithms, dramatically simplifying future research and boosting performance.
What is my quantum computer good for? Quantum capability learning with physics-aware neural networks
·1734 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Sandia National Laboratories
Quantum-physics-aware neural networks achieve up to 50% improved accuracy in predicting quantum computer capabilities, scaling to 100+ qubits.
What If the Input is Expanded in OOD Detection?
·3779 words·18 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Wuhan University
Boost OOD detection accuracy by averaging model confidence scores from original and corrupted inputs!
WeiPer: OOD Detection using Weight Perturbations of Class Projections
·5838 words·28 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Free University of Berlin
WeiPer enhances OOD detection by cleverly perturbing class projections, creating a richer representation that improves various existing methods and achieves state-of-the-art results.
Weight for Robustness: A Comprehensive Approach towards Optimal Fault-Tolerant Asynchronous ML
·1754 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Technion
Optimal fault-tolerant asynchronous machine learning is achieved via a novel weighted robust aggregation framework, ensuring efficient training despite Byzantine failures and heterogeneous resources.
Weight Diffusion for Future: Learn to Generalize in Non-Stationary Environments
·2419 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Tencent AI Lab
Weight Diffusion (W-Diff) masters evolving domain generalization by using conditional diffusion models to learn classifier weight evolution patterns, enabling superior generalization to unseen future …
Weight decay induces low-rank attention layers
·1731 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ ETH Zurich
Weight decay in deep learning surprisingly induces low-rank attention layers, potentially harming performance but offering optimization strategies for large language models.
Weak Supervision Performance Evaluation via Partial Identification
·1757 words·9 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
π’ University of Michigan
This paper introduces a novel method for evaluating weakly supervised models using FrΓ©chet bounds, providing reliable performance bounds without ground truth labels.
WaveAttack: Asymmetric Frequency Obfuscation-based Backdoor Attacks Against Deep Neural Networks
·2153 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ East China Normal University
WaveAttack: A new backdoor attack method leveraging asymmetric frequency obfuscation for high stealthiness and effectiveness in Deep Neural Networks.
Wasserstein Gradient Boosting: A Framework for Distribution-Valued Supervised Learning
·3031 words·15 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ University of Edinburgh
Wasserstein Gradient Boosting (WGBoost) extends gradient boosting to handle probability distributions as outputs, enabling more robust and informative predictions in various applications.
Warm-up Free Policy Optimization: Improved Regret in Linear Markov Decision Processes
·193 words·1 min·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Tel Aviv University
Warm-up-free policy optimization achieves rate-optimal regret in linear Markov decision processes, improving efficiency and dependence on problem parameters.