🏢 University of Washington
Multilingual Diversity Improves Vision-Language Representations
·2777 words·14 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Washington
Boosting vision-language models: Multilingual data improves performance on English-centric benchmarks.
Meta-Diffu$B$: A Contextualized Sequence-to-Sequence Text Diffusion Model with Meta-Exploration
·3164 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Text Generation
🏢 University of Washington
Meta-DiffuB enhances sequence-to-sequence text diffusion models by using meta-exploration to learn a contextualized noise schedule, resulting in state-of-the-art performance.
MediQ: Question-Asking LLMs and a Benchmark for Reliable Interactive Clinical Reasoning
·2561 words·13 mins·
loading
·
loading
Natural Language Processing
Question Answering
🏢 University of Washington
MEDIQ benchmark revolutionizes LLM evaluation by shifting from static to interactive clinical reasoning, revealing LLMs’ struggles with proactive information-seeking and highlighting the importance of…
MAGNET: Improving the Multilingual Fairness of Language Models with Adaptive Gradient-Based Tokenization
·2236 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Washington
MAGNET, a novel adaptive gradient-based tokenization method, tackles multilingual language model bias by employing language-specific boundary predictors to achieve equitable segmentation across divers…
Learning to Price Homogeneous Data
·1192 words·6 mins·
loading
·
loading
AI Applications
Finance
🏢 University of Washington
This paper develops efficient algorithms for pricing homogeneous data in online settings, achieving low regret using novel discretization schemes that scale well with data size and number of buyer typ…
Learning to Cooperate with Humans using Generative Agents
·2752 words·13 mins·
loading
·
loading
AI Applications
Gaming
🏢 University of Washington
Generative Agent Modeling for Multi-agent Adaptation (GAMMA) improves human-AI cooperation by training AI agents against diverse partners generated from a latent model, enhancing zero-shot coordinatio…
Learning Optimal Tax Design in Nonatomic Congestion Games
·442 words·3 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Washington
AI learns optimal taxes for congestion games, maximizing social welfare with limited feedback, via a novel algorithm.
Large Scale Transfer Learning for Tabular Data via Language Modeling
·2834 words·14 mins·
loading
·
loading
Machine Learning
Transfer Learning
🏢 University of Washington
TABULA-8B, a novel language model for tabular prediction, achieves state-of-the-art zero-shot and few-shot performance across various benchmarks, exceeding existing methods by 5-15 percentage points.
Is O(log N) practical? Near-Equivalence Between Delay Robustness and Bounded Regret in Bandits and RL
·403 words·2 mins·
loading
·
loading
AI Theory
Robustness
🏢 University of Washington
Zero Graves-Lai constant ensures both bounded regret and delay robustness in online decision-making, particularly for linear models.
Initializing Services in Interactive ML Systems for Diverse Users
·1498 words·8 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 University of Washington
Adaptively initializing multi-service ML systems for diverse users using minimal data, this paper introduces a randomized algorithm achieving near-optimal loss with provable guarantees.
How does Gradient Descent Learn Features --- A Local Analysis for Regularized Two-Layer Neural Networks
·1326 words·7 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Washington
Neural networks learn features effectively through gradient descent, not just at the beginning, but also at the end of training, even with carefully regularized objectives.
From an Image to a Scene: Learning to Imagine the World from a Million 360° Videos
·2541 words·12 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 University of Washington
ODIN, trained on a million 360° videos (360-1M), generates realistic novel views and reconstructs 3D scenes from single images.
Drago: Primal-Dual Coupled Variance Reduction for Faster Distributionally Robust Optimization
·1908 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Washington
DRAGO: A novel primal-dual algorithm delivers faster, state-of-the-art convergence for distributionally robust optimization.
Distributional Successor Features Enable Zero-Shot Policy Optimization
·2834 words·14 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Washington
DiSPOs: a novel model for zero-shot policy optimization in reinforcement learning, enabling quick adaptation to new tasks by learning a distribution of successor features and avoiding compounding erro…
Discovering plasticity rules that organize and maintain neural circuits
·1657 words·8 mins·
loading
·
loading
Machine Learning
Meta Learning
🏢 University of Washington
AI discovers robust, biologically-plausible plasticity rules that self-organize and maintain neural circuits’ sequential activity, even with synaptic turnover.
Deep Submodular Peripteral Networks
·2349 words·12 mins·
loading
·
loading
🏢 University of Washington
Deep Submodular Peripteral Networks (DSPNs) learn submodular functions efficiently using graded pairwise comparisons, surpassing traditional methods and demonstrating superiority in experimental desig…
Data Mixture Inference Attack: BPE Tokenizers Reveal Training Data Compositions
·3904 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Washington
Researchers uncover hidden training data secrets of large language models by analyzing their byte-pair encoding tokenizers, revealing the proportions of different languages and domains.
Customized Multiple Clustering via Multi-Modal Subspace Proxy Learning
·1867 words·9 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Washington
Multi-Sub leverages multi-modal learning to achieve customized multiple clustering, aligning user-defined textual preferences with visual representations via a subspace proxy learning framework.
Cryptographic Hardness of Score Estimation
·386 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 University of Washington
Score estimation, crucial for diffusion models, is computationally hard even with polynomial sample complexity unless strong distributional assumptions are made.
CLIPLoss and Norm-Based Data Selection Methods for Multimodal Contrastive Learning
·2694 words·13 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Washington
Boosting multimodal contrastive learning, this research introduces negCLIPLoss and NormSim, novel data selection methods surpassing existing techniques by improving data quality and task relevance. Th…