Skip to main content

🏢 University of Wisconsin-Madison

Yo'LLaVA: Your Personalized Language and Vision Assistant
·4272 words·21 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 University of Wisconsin-Madison
Yo’LLaVA personalizes Large Multimodal Models (LMMs) to converse about specific subjects using just a few images, embedding concepts into latent tokens for efficient and effective personalized convers…
TSDS: Data Selection for Task-Specific Model Finetuning
·2005 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Wisconsin-Madison
TSDS: A novel framework selects optimal training data for efficient large language model finetuning using only a few examples, boosting performance.
Towards Calibrated Robust Fine-Tuning of Vision-Language Models
·3938 words·19 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 University of Wisconsin-Madison
Calibrated robust fine-tuning boosts vision-language model accuracy and confidence in out-of-distribution scenarios by using a constrained multimodal contrastive loss and self-distillation.
Tighter Convergence Bounds for Shuffled SGD via Primal-Dual Perspective
·1717 words·9 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of Wisconsin-Madison
Shuffled SGD’s convergence is now better understood through a primal-dual analysis, yielding tighter bounds that align with its superior empirical performance.
The ALCHEmist: Automated Labeling 500x CHEaper than LLM Data Annotators
·2284 words·11 mins· loading · loading
🏢 University of Wisconsin-Madison
Alchemist, a novel automated labeling system, reduces data annotation costs by 500x compared to LLMs while improving accuracy by an average of 12.9%.
Task-Agnostic Machine-Learning-Assisted Inference
·1432 words·7 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏢 University of Wisconsin-Madison
PSPS: a novel task-agnostic framework enables valid and efficient ML-assisted statistical inference for virtually any task, simply using summary statistics from existing analysis routines!
Span-Based Optimal Sample Complexity for Weakly Communicating and General Average Reward MDPs
·1956 words·10 mins· loading · loading
Reinforcement Learning 🏢 University of Wisconsin-Madison
This paper achieves minimax-optimal bounds for learning near-optimal policies in average-reward MDPs, addressing a long-standing open problem in reinforcement learning.
RL in Latent MDPs is Tractable: Online Guarantees via Off-Policy Evaluation
·435 words·3 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Wisconsin-Madison
First sample-efficient algorithm for LMDPs without separation assumptions, achieving near-optimal guarantees via novel off-policy evaluation.
Reliable Learning of Halfspaces under Gaussian Marginals
·265 words·2 mins· loading · loading
AI Theory Optimization 🏢 University of Wisconsin-Madison
New algorithm reliably learns Gaussian halfspaces with significantly improved sample and computational complexity compared to existing methods, offering strong computational separation from standard a…
Pearls from Pebbles: Improved Confidence Functions for Auto-labeling
·5535 words·26 mins· loading · loading
AI Generated Machine Learning Semi-Supervised Learning 🏢 University of Wisconsin-Madison
Colander: a novel auto-labeling technique boosts data efficiency by 60%, optimizing confidence functions for maximum coverage with minimal error.
Optimal Design for Human Preference Elicitation
·1485 words·7 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Wisconsin-Madison
Dope: Efficient algorithms optimize human preference elicitation for learning to rank, minimizing ranking loss and prediction error with absolute and ranking feedback models.
Learning a Single Neuron Robustly to Distributional Shifts and Adversarial Label Noise
·235 words·2 mins· loading · loading
AI Theory Robustness 🏢 University of Wisconsin-Madison
This work presents a computationally efficient algorithm that robustly learns a single neuron despite adversarial label noise and distributional shifts, providing provable approximation guarantees.
HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection
·2304 words·11 mins· loading · loading
Large Language Models 🏢 University of Wisconsin-Madison
HaloScope leverages unlabeled LLM outputs to accurately detect AI hallucinations without human annotation, significantly outperforming existing methods.
Grammar-Aligned Decoding
·7195 words·34 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Wisconsin-Madison
Adaptive Sampling with Approximate Expected Futures (ASAp) ensures LLMs generate grammatically correct outputs that closely match the model’s original probability distribution.
Faster Algorithms for User-Level Private Stochastic Convex Optimization
·1097 words·6 mins· loading · loading
AI Theory Privacy 🏢 University of Wisconsin-Madison
Faster algorithms achieve optimal excess risk in user-level private stochastic convex optimization, overcoming limitations of prior methods without restrictive assumptions.
Deterministic Policies for Constrained Reinforcement Learning in Polynomial Time
·1494 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Wisconsin-Madison
This paper presents an efficient algorithm to compute near-optimal deterministic policies for constrained reinforcement learning problems, solving a 25-year-old computational complexity challenge.
Consistency Purification: Effective and Efficient Diffusion Purification towards Certified Robustness
·1893 words·9 mins· loading · loading
Computer Vision Image Generation 🏢 University of Wisconsin-Madison
Consistency Purification boosts certified robustness by efficiently purifying noisy images using a one-step generative model, achieving state-of-the-art results while maintaining semantic alignment.
Coherence-free Entrywise Estimation of Eigenvectors in Low-rank Signal-plus-noise Matrix Models
·1535 words·8 mins· loading · loading
AI Theory Optimization 🏢 University of Wisconsin-Madison
New method for eigenvector estimation achieves optimal rates without coherence dependence, improving low-rank matrix denoising and related tasks.
BackdoorAlign: Mitigating Fine-tuning based Jailbreak Attack with Backdoor Enhanced Safety Alignment
·2859 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Wisconsin-Madison
BackdoorAlign defends against fine-tuning-based LLM jailbreaks using a ‘backdoor trigger’ to enforce safety alignment during inference, effectively mitigating risks with minimal additional safety exam…
An effective framework for estimating individualized treatment rules
·3345 words·16 mins· loading · loading
AI Generated AI Theory Causality 🏢 University of Wisconsin-Madison
This paper introduces a unified ITR estimation framework using covariate balancing weights, achieving significant gains in robustness and effectiveness compared to existing methods.