Machine Learning
Computing the Bias of Constant-step Stochastic Approximation with Markovian Noise
·2187 words·11 mins·
loading
·
loading
Machine Learning
Stochastic Approximation
🏢 Univ. Grenoble Alpes and Inria
New method quantifies & reduces bias in constant-step stochastic approximation algorithms with Markovian noise, improving accuracy and efficiency.
Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference
·1804 words·9 mins·
loading
·
loading
Machine Learning
Gaussian Processes
🏢 Columbia University
Computation-Aware Gaussian Processes (CaGP) achieve linear-time inference and model selection, enabling efficient training of GPs on large datasets without compromising uncertainty quantification, sur…
Compositional Automata Embeddings for Goal-Conditioned Reinforcement Learning
·3934 words·19 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 UC Berkeley
Goal-conditioned RL gets a temporal upgrade with compositional DFAs (cDFAs), enabling zero-shot generalization and faster policy specialization via novel graph neural network embeddings and reach-avoi…
Communication-Efficient Federated Group Distributionally Robust Optimization
·1930 words·10 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 Texas A&M University
Communication-efficient algorithms for federated group distributionally robust optimization (FGDRO) are introduced, achieving lower communication complexity and superior performance on real-world task…
Communication Efficient Distributed Training with Distributed Lion
·1698 words·8 mins·
loading
·
loading
Machine Learning
Optimization
🏢 University of Texas at Austin
Distributed Lion: Training large AI models efficiently by communicating only binary or low-precision vectors between workers and a server, significantly reducing communication costs and maintaining co…
Combining Statistical Depth and Fermat Distance for Uncertainty Quantification
·2374 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Institut De Recherche en Informatique De Toulouse
Boosting neural network prediction reliability, this research ingeniously combines statistical depth and Fermat distance for superior uncertainty quantification, eliminating the need for distributiona…
Collaborative Refining for Learning from Inaccurate Labels
·1752 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Ant Group
Collaborative Refining for Learning from Inaccurate Labels (CRL) refines data using annotator agreement, improving model accuracy with noisy labels.
CODA: A Correlation-Oriented Disentanglement and Augmentation Modeling Scheme for Better Resisting Subpopulation Shifts
·1907 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 City University of Hong Kong
CODA: A novel modeling scheme tackles subpopulation shifts in machine learning by disentangling spurious correlations, augmenting data strategically, and using reweighted consistency loss for improved…
CoBo: Collaborative Learning via Bilevel Optimization
·1628 words·8 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 EPFL
CoBo: A novel bilevel optimization algorithm for collaborative learning surpasses existing methods by efficiently selecting helpful clients, resulting in superior performance and scalability.
Clustering with Non-adaptive Subset Queries
·407 words·2 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 UC San Diego
This paper introduces novel non-adaptive algorithms for clustering using subset queries, achieving near-linear query complexity and improving upon existing limitations of pairwise query methods.
Clustering then Propagation: Select Better Anchors for Knowledge Graph Embedding
·1993 words·10 mins·
loading
·
loading
Machine Learning
Knowledge Graph Embedding
🏢 National University of Defense Technology
RecPiece: Relational Clustering for Better Knowledge Graph Embedding Anchors
Changing the Training Data Distribution to Reduce Simplicity Bias Improves In-distribution Generalization
·3113 words·15 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 UC Los Angeles
Boosting in-distribution generalization is achieved by strategically altering the training data distribution to reduce simplicity bias and promote uniform feature learning.
CE-NAS: An End-to-End Carbon-Efficient Neural Architecture Search Framework
·2041 words·10 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Worcester Polytechnic Institute
CE-NAS: A novel framework minimizes the carbon footprint of Neural Architecture Search by dynamically allocating GPU resources based on predicted carbon intensity, achieving state-of-the-art results w…
Causal Imitation for Markov Decision Processes: a Partial Identification Approach
·1601 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Columbia University
This paper presents novel causal imitation learning algorithms using partial identification to achieve expert performance even when unobserved confounders affect Markov Decision Processes.
Causal Contrastive Learning for Counterfactual Regression Over Time
·3424 words·17 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
🏢 Paris-Saclay University
Causal CPC: a novel method for accurate and efficient counterfactual regression over time using RNNs, CPC, and InfoMax, achieving state-of-the-art performance.
Categorical Flow Matching on Statistical Manifolds
·2341 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Generative Models
🏢 Peking University
Statistical Flow Matching (SFM) uses information geometry to create a new flow-matching framework for generating discrete data, achieving superior sampling quality and likelihood compared to existing …
Catastrophic Goodhart: regularizing RLHF with KL divergence does not mitigate heavy-tailed reward misspecification
·2257 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Independent / FAR Labs
RLHF’s KL regularization fails to prevent ‘catastrophic Goodhart’—policies achieving high proxy reward but low actual utility—when reward errors have heavy tails.
Cascade of phase transitions in the training of energy-based models
·1743 words·9 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Université PSL
Energy-based models’ training reveals a cascade of phase transitions, progressively learning data features, offering new insights into deep learning dynamics.
Carrot and Stick: Eliciting Comparison Data and Beyond
·1825 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Harvard University
Truthful comparison data is hard to obtain without ground truth. This paper presents novel peer prediction mechanisms using bonus-penalty payments that incentivize truthful comparisons, even in networ…
Cardinality-Aware Set Prediction and Top-$k$ Classification
·1676 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Google Research
This paper proposes cardinality-aware top-k classification, improving accuracy and efficiency by dynamically adjusting prediction set sizes.