🏢 University of Wisconsin-Madison
Yo'LLaVA: Your Personalized Language and Vision Assistant
·4272 words·21 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Wisconsin-Madison
Yo’LLaVA personalizes Large Multimodal Models (LMMs) to converse about specific subjects using just a few images, embedding concepts into latent tokens for efficient and effective personalized convers…
TSDS: Data Selection for Task-Specific Model Finetuning
·2005 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Wisconsin-Madison
TSDS: A novel framework selects optimal training data for efficient large language model finetuning using only a few examples, boosting performance.
Towards Calibrated Robust Fine-Tuning of Vision-Language Models
·3938 words·19 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 University of Wisconsin-Madison
Calibrated robust fine-tuning boosts vision-language model accuracy and confidence in out-of-distribution scenarios by using a constrained multimodal contrastive loss and self-distillation.
Tighter Convergence Bounds for Shuffled SGD via Primal-Dual Perspective
·1717 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 University of Wisconsin-Madison
Shuffled SGD’s convergence is now better understood through a primal-dual analysis, yielding tighter bounds that align with its superior empirical performance.
The ALCHEmist: Automated Labeling 500x CHEaper than LLM Data Annotators
·2284 words·11 mins·
loading
·
loading
🏢 University of Wisconsin-Madison
Alchemist, a novel automated labeling system, reduces data annotation costs by 500x compared to LLMs while improving accuracy by an average of 12.9%.
Task-Agnostic Machine-Learning-Assisted Inference
·1432 words·7 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
🏢 University of Wisconsin-Madison
PSPS: a novel task-agnostic framework enables valid and efficient ML-assisted statistical inference for virtually any task, simply using summary statistics from existing analysis routines!
Span-Based Optimal Sample Complexity for Weakly Communicating and General Average Reward MDPs
·1956 words·10 mins·
loading
·
loading
Reinforcement Learning
🏢 University of Wisconsin-Madison
This paper achieves minimax-optimal bounds for learning near-optimal policies in average-reward MDPs, addressing a long-standing open problem in reinforcement learning.
RL in Latent MDPs is Tractable: Online Guarantees via Off-Policy Evaluation
·435 words·3 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Wisconsin-Madison
First sample-efficient algorithm for LMDPs without separation assumptions, achieving near-optimal guarantees via novel off-policy evaluation.
Reliable Learning of Halfspaces under Gaussian Marginals
·265 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Wisconsin-Madison
New algorithm reliably learns Gaussian halfspaces with significantly improved sample and computational complexity compared to existing methods, offering strong computational separation from standard a…
Pearls from Pebbles: Improved Confidence Functions for Auto-labeling
·5535 words·26 mins·
loading
·
loading
AI Generated
Machine Learning
Semi-Supervised Learning
🏢 University of Wisconsin-Madison
Colander: a novel auto-labeling technique boosts data efficiency by 60%, optimizing confidence functions for maximum coverage with minimal error.
Optimal Design for Human Preference Elicitation
·1485 words·7 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Wisconsin-Madison
Dope: Efficient algorithms optimize human preference elicitation for learning to rank, minimizing ranking loss and prediction error with absolute and ranking feedback models.
Learning a Single Neuron Robustly to Distributional Shifts and Adversarial Label Noise
·235 words·2 mins·
loading
·
loading
AI Theory
Robustness
🏢 University of Wisconsin-Madison
This work presents a computationally efficient algorithm that robustly learns a single neuron despite adversarial label noise and distributional shifts, providing provable approximation guarantees.
HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection
·2304 words·11 mins·
loading
·
loading
Large Language Models
🏢 University of Wisconsin-Madison
HaloScope leverages unlabeled LLM outputs to accurately detect AI hallucinations without human annotation, significantly outperforming existing methods.
Grammar-Aligned Decoding
·7195 words·34 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Wisconsin-Madison
Adaptive Sampling with Approximate Expected Futures (ASAp) ensures LLMs generate grammatically correct outputs that closely match the model’s original probability distribution.
Faster Algorithms for User-Level Private Stochastic Convex Optimization
·1097 words·6 mins·
loading
·
loading
AI Theory
Privacy
🏢 University of Wisconsin-Madison
Faster algorithms achieve optimal excess risk in user-level private stochastic convex optimization, overcoming limitations of prior methods without restrictive assumptions.
Deterministic Policies for Constrained Reinforcement Learning in Polynomial Time
·1494 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Wisconsin-Madison
This paper presents an efficient algorithm to compute near-optimal deterministic policies for constrained reinforcement learning problems, solving a 25-year-old computational complexity challenge.
Consistency Purification: Effective and Efficient Diffusion Purification towards Certified Robustness
·1893 words·9 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Wisconsin-Madison
Consistency Purification boosts certified robustness by efficiently purifying noisy images using a one-step generative model, achieving state-of-the-art results while maintaining semantic alignment.
Coherence-free Entrywise Estimation of Eigenvectors in Low-rank Signal-plus-noise Matrix Models
·1535 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Wisconsin-Madison
New method for eigenvector estimation achieves optimal rates without coherence dependence, improving low-rank matrix denoising and related tasks.
BackdoorAlign: Mitigating Fine-tuning based Jailbreak Attack with Backdoor Enhanced Safety Alignment
·2859 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Wisconsin-Madison
BackdoorAlign defends against fine-tuning-based LLM jailbreaks using a ‘backdoor trigger’ to enforce safety alignment during inference, effectively mitigating risks with minimal additional safety exam…
An effective framework for estimating individualized treatment rules
·3345 words·16 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 University of Wisconsin-Madison
This paper introduces a unified ITR estimation framework using covariate balancing weights, achieving significant gains in robustness and effectiveness compared to existing methods.