🏢 University of Washington
Visual Sketchpad: Sketching as a Visual Chain of Thought for Multimodal Language Models
·3551 words·17 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 University of Washington
Visual SKETCHPAD empowers multimodal language models (LLMs) with visual reasoning abilities by allowing them to generate intermediate sketches. This innovative framework substantially enhances LLM per…
Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback
·2380 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Washington
This study disentangles best practices for learning from preference feedback in LLMs, revealing that data quality, algorithm choice, and reward model significantly impact performance.
Uniform Last-Iterate Guarantee for Bandits and Reinforcement Learning
·285 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Washington
This paper introduces the Uniform Last-Iterate (ULI) guarantee, a novel metric for evaluating reinforcement learning algorithms that considers both cumulative and instantaneous performance. Unlike ex…
Understanding the Gains from Repeated Self-Distillation
·2009 words·10 mins·
loading
·
loading
Machine Learning
Optimization
🏢 University of Washington
Repeated self-distillation significantly reduces excess risk in linear regression, achieving up to a ’d’ factor improvement over single-step methods.
Toward Global Convergence of Gradient EM for Over-Paramterized Gaussian Mixture Models
·345 words·2 mins·
loading
·
loading
Machine Learning
Optimization
🏢 University of Washington
Gradient EM for over-parameterized Gaussian Mixture Models globally converges with a sublinear rate, solving a longstanding open problem in machine learning.
The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
·2874 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
Image Classification
🏢 University of Washington
Using real images retrieved from a generator’s training data outperforms using synthetic images generated by that same model for image classification.
The Benefits of Balance: From Information Projections to Variance Reduction
·1859 words·9 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
🏢 University of Washington
Data balancing in foundation models surprisingly reduces variance, improving model training and performance.
Tell What You Hear From What You See - Video to Audio Generation Through Text
·2349 words·12 mins·
loading
·
loading
Multimodal Learning
Audio-Visual Learning
🏢 University of Washington
VATT: Text-guided video-to-audio generation, enabling refined audio control via text prompts and improved compatibility.
Swift Sampler: Efficient Learning of Sampler by 10 Parameters
·2624 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University of Washington
Swift Sampler (SS) automates the learning of efficient data samplers for deep learning, achieving significant performance gains (e.g., 1.5% on ImageNet) with minimal computational cost using only 10 p…
Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
·2903 words·14 mins·
loading
·
loading
Natural Language Processing
Text Generation
🏢 University of Washington
Generate multiple text drafts from a single language model pass with Superposed Decoding, significantly boosting efficiency!
Smoothed Energy Guidance: Guiding Diffusion Models with Reduced Energy Curvature of Attention
·2786 words·14 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Washington
Smoothed Energy Guidance (SEG) improves unconditional image generation by reducing self-attention’s energy curvature, leading to higher-quality outputs with fewer artifacts.
Self-Calibrating Conformal Prediction
·2092 words·10 mins·
loading
·
loading
AI Applications
Healthcare
🏢 University of Washington
Self-Calibrating Conformal Prediction (SC-CP) marries model calibration and conformal prediction for more efficient and interpretable prediction intervals with prediction-conditional validity.
Scaling Retrieval-Based Language Models with a Trillion-Token Datastore
·4019 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Washington
Massive language models improve with bigger datastores at inference time. A 1.4 trillion-token datastore, MASSIVEDS, shows that retrieval-based LMs outperform larger, solely-trained models on knowled…
Sample Complexity Reduction via Policy Difference Estimation in Tabular Reinforcement Learning
·406 words·2 mins·
loading
·
loading
Reinforcement Learning
🏢 University of Washington
This paper reveals that estimating only policy differences, while effective in bandits, is insufficient for tabular reinforcement learning. However, it introduces a novel algorithm achieving near-opti…
QUEST: Quality-Aware Metropolis-Hastings Sampling for Machine Translation
·2191 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Machine Translation
🏢 University of Washington
QUEST, a novel Metropolis-Hastings sampling method, generates high-quality & diverse machine translations by using quality metrics as energy functions, overcoming limitations of likelihood-based and r…
Query-Based Adversarial Prompt Generation
·1773 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Washington
Researchers developed a query-based attack that generates adversarial prompts, fooling language models into producing harmful outputs with significantly higher success rates than previous methods, eff…
Personalizing Reinforcement Learning from Human Feedback with Variational Preference Learning
·2411 words·12 mins·
loading
·
loading
Reinforcement Learning
🏢 University of Washington
VPL: a novel multimodal RLHF personalizes AI by inferring user-specific latent preferences, enabling accurate reward modeling and improved policy alignment for diverse populations.
On the Complexity of Teaching a Family of Linear Behavior Cloning Learners
·1819 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Washington
A novel algorithm, TIE, optimally teaches a family of linear behavior cloning learners, achieving instance-optimal teaching dimension while providing efficient approximation for larger action spaces.
Nearly Optimal Approximation of Matrix Functions by the Lanczos Method
·1646 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Washington
Lanczos-FA, a simple algorithm for approximating matrix functions, surprisingly outperforms newer methods; this paper proves its near-optimality for rational functions, explaining its practical succes…
Nearly Minimax Optimal Submodular Maximization with Bandit Feedback
·384 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Washington
This research establishes the first minimax optimal algorithm for submodular maximization with bandit feedback, achieving a regret bound matching the lower bound.