Skip to main content

Machine Learning

Zero-Shot Reinforcement Learning from Low Quality Data
·4722 words·23 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏒 University of Cambridge
Zero-shot RL struggles with low-quality data; this paper introduces conservative algorithms that significantly boost performance on such data without sacrificing performance on high-quality data.
Your Diffusion Model is Secretly a Noise Classifier and Benefits from Contrastive Training
·2424 words·12 mins· loading · loading
Machine Learning Self-Supervised Learning 🏒 UC Riverside
Diffusion models benefit from contrastive training, improving sample quality and speed by addressing poor denoiser estimation in out-of-distribution regions.
Your contrastive learning problem is secretly a distribution alignment problem
·381 words·2 mins· loading · loading
Machine Learning Self-Supervised Learning 🏒 University of Toronto
Contrastive learning is reframed as a distribution alignment problem, leading to a flexible framework (GCA) that improves representation learning with unbalanced optimal transport.
You Don’t Need Domain-Specific Data Augmentations When Scaling Self-Supervised Learning
·2133 words·11 mins· loading · loading
AI Generated Machine Learning Self-Supervised Learning 🏒 FAIR at Meta
Self-supervised learning’s reliance on complex data augmentations is challenged; a large-scale study shows comparable performance using only cropping, suggesting dataset size is more important than au…
Worst-Case Offline Reinforcement Learning with Arbitrary Data Support
·450 words·3 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏒 IBM Research
Worst-case offline RL guarantees near-optimal policy performance without data support assumptions, achieving a sample complexity bound of O(Ρ⁻²).
Why Go Full? Elevating Federated Learning Through Partial Network Updates
·3064 words·15 mins· loading · loading
AI Generated Machine Learning Federated Learning 🏒 Beihang University
FedPart boosts federated learning by updating only parts of the network, solving the layer mismatch problem, and achieving faster convergence with higher accuracy.
When Your AIs Deceive You: Challenges of Partial Observability in Reinforcement Learning from Human Feedback
·2699 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Amsterdam
RLHF’s reliance on fully observable environments is challenged: human feedback, often partial, leads to deceptive AI behavior (inflation & overjustification).
When to Sense and Control? A Time-adaptive Approach for Continuous-Time RL
·2003 words·10 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏒 ETH Zurich
TACOS: A novel time-adaptive RL framework drastically reduces interactions in continuous-time systems while improving performance, offering both model-free and model-based algorithms.
What Matters in Graph Class Incremental Learning? An Information Preservation Perspective
·3421 words·17 mins· loading · loading
Machine Learning Deep Learning 🏒 College of Intelligence and Computing, Tianjin University
GSIP framework mitigates catastrophic forgetting in graph class incremental learning by preserving crucial graph information, achieving a 10% improvement in forgetting metrics.
What Makes Partial-Label Learning Algorithms Effective?
·1666 words·8 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏒 Southeast University
Unlocking Partial-Label Learning: A new study reveals surprisingly simple design principles for highly accurate algorithms, dramatically simplifying future research and boosting performance.
What is my quantum computer good for? Quantum capability learning with physics-aware neural networks
·1734 words·9 mins· loading · loading
Machine Learning Deep Learning 🏒 Sandia National Laboratories
Quantum-physics-aware neural networks achieve up to 50% improved accuracy in predicting quantum computer capabilities, scaling to 100+ qubits.
What If the Input is Expanded in OOD Detection?
·3779 words·18 mins· loading · loading
Machine Learning Deep Learning 🏒 Wuhan University
Boost OOD detection accuracy by averaging model confidence scores from original and corrupted inputs!
WeiPer: OOD Detection using Weight Perturbations of Class Projections
·5838 words·28 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 Free University of Berlin
WeiPer enhances OOD detection by cleverly perturbing class projections, creating a richer representation that improves various existing methods and achieves state-of-the-art results.
Weight for Robustness: A Comprehensive Approach towards Optimal Fault-Tolerant Asynchronous ML
·1754 words·9 mins· loading · loading
Machine Learning Deep Learning 🏒 Technion
Optimal fault-tolerant asynchronous machine learning is achieved via a novel weighted robust aggregation framework, ensuring efficient training despite Byzantine failures and heterogeneous resources.
Weight Diffusion for Future: Learn to Generalize in Non-Stationary Environments
·2419 words·12 mins· loading · loading
Machine Learning Deep Learning 🏒 Tencent AI Lab
Weight Diffusion (W-Diff) masters evolving domain generalization by using conditional diffusion models to learn classifier weight evolution patterns, enabling superior generalization to unseen future …
Weight decay induces low-rank attention layers
·1731 words·9 mins· loading · loading
Machine Learning Deep Learning 🏒 ETH Zurich
Weight decay in deep learning surprisingly induces low-rank attention layers, potentially harming performance but offering optimization strategies for large language models.
Weak Supervision Performance Evaluation via Partial Identification
·1757 words·9 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏒 University of Michigan
This paper introduces a novel method for evaluating weakly supervised models using FrΓ©chet bounds, providing reliable performance bounds without ground truth labels.
WaveAttack: Asymmetric Frequency Obfuscation-based Backdoor Attacks Against Deep Neural Networks
·2153 words·11 mins· loading · loading
Machine Learning Deep Learning 🏒 East China Normal University
WaveAttack: A new backdoor attack method leveraging asymmetric frequency obfuscation for high stealthiness and effectiveness in Deep Neural Networks.
Wasserstein Gradient Boosting: A Framework for Distribution-Valued Supervised Learning
·3031 words·15 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Edinburgh
Wasserstein Gradient Boosting (WGBoost) extends gradient boosting to handle probability distributions as outputs, enabling more robust and informative predictions in various applications.
Warm-up Free Policy Optimization: Improved Regret in Linear Markov Decision Processes
·193 words·1 min· loading · loading
Machine Learning Reinforcement Learning 🏒 Tel Aviv University
Warm-up-free policy optimization achieves rate-optimal regret in linear Markov decision processes, improving efficiency and dependence on problem parameters.