Skip to main content

🏢 UC San Diego

VLG-CBM: Training Concept Bottleneck Models with Vision-Language Guidance
·3011 words·15 mins· loading · loading
Computer Vision Visual Question Answering 🏢 UC San Diego
VLG-CBM enhances concept bottleneck models with vision-language guidance for faithful interpretability and improved accuracy.
UniMTS: Unified Pre-training for Motion Time Series
·2033 words·10 mins· loading · loading
AI Applications Healthcare 🏢 UC San Diego
UniMTS, a novel unified pre-training model, achieves state-of-the-art performance in motion time series classification by generalizing across diverse device factors and activities.
Towards Diverse Device Heterogeneous Federated Learning via Task Arithmetic Knowledge Integration
·2932 words·14 mins· loading · loading
Machine Learning Federated Learning 🏢 UC San Diego
TAKFL, a novel federated learning framework, tackles device heterogeneity by independently distilling knowledge from diverse devices and integrating it adaptively, achieving state-of-the-art performan…
Symmetry-Informed Governing Equation Discovery
·2189 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 UC San Diego
Leveraging symmetry in automated equation discovery improves accuracy and simplicity of learned governing equations, enhancing robustness against noise and achieving higher success rates across divers…
SpatialRGPT: Grounded Spatial Reasoning in Vision-Language Models
·3123 words·15 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 UC San Diego
SpatialRGPT enhances Vision-Language Models’ spatial reasoning by integrating 3D scene graphs and depth information, achieving significant performance gains on spatial reasoning tasks.
Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
·2661 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 UC San Diego
SharpBalance, a novel training approach, effectively improves deep ensemble performance by addressing the sharpness-diversity trade-off, leading to significant improvements in both in-distribution and…
Replicable Uniformity Testing
·268 words·2 mins· loading · loading
AI Generated AI Theory Optimization 🏢 UC San Diego
This paper presents the first replicable uniformity tester with nearly linear dependence on the replicability parameter, enhancing the reliability of scientific studies using distribution testing algo…
RA-PbRL: Provably Efficient Risk-Aware Preference-Based Reinforcement Learning
·1598 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC San Diego
RA-PbRL introduces a provably efficient algorithm for risk-aware preference-based reinforcement learning, addressing the limitations of existing risk-neutral methods in applications demanding heighten…
Quantifying and Optimizing Global Faithfulness in Persona-driven Role-playing
·2570 words·13 mins· loading · loading
Natural Language Processing Dialogue Systems 🏢 UC San Diego
New APC metric precisely quantifies & optimizes global faithfulness in persona-driven role-playing, offering a fine-grained, explainable evaluation and improving AI character consistency.
Provable and Efficient Dataset Distillation for Kernel Ridge Regression
·1601 words·8 mins· loading · loading
Machine Learning Deep Learning 🏢 UC San Diego
One data point per class suffices for efficient and provable dataset distillation in kernel ridge regression, significantly reducing computational costs.
Online Feature Updates Improve Online (Generalized) Label Shift Adaptation
·1991 words·10 mins· loading · loading
Machine Learning Self-Supervised Learning 🏢 UC San Diego
Online Label Shift adaptation with Online Feature Updates (OLS-OFU) significantly boosts online label shift adaptation by dynamically refining feature extractors using self-supervised learning, achiev…
Online Consistency of the Nearest Neighbor Rule
·1388 words·7 mins· loading · loading
AI Theory Optimization 🏢 UC San Diego
The 1-nearest neighbor rule achieves online consistency under surprisingly broad conditions: measurable label functions and mild assumptions on instance generation in doubling metric spaces.
On Differentially Private U Statistics
·403 words·2 mins· loading · loading
AI Theory Privacy 🏢 UC San Diego
New algorithms achieve near-optimal differentially private U-statistic estimation, significantly improving accuracy over existing methods.
Linking In-context Learning in Transformers to Human Episodic Memory
·3883 words·19 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC San Diego
Transformers’ in-context learning mirrors human episodic memory, with specific attention heads acting like the brain’s contextual maintenance and retrieval system.
Learning from Uncertain Data: From Possible Worlds to Possible Models
·2895 words·14 mins· loading · loading
AI Theory Robustness 🏢 UC San Diego
ZORRO: A new method for learning linear models from uncertain data, providing sound over-approximations of all possible models and prediction ranges.
Iteratively Refined Early Interaction Alignment for Subgraph Matching based Graph Retrieval
·3157 words·15 mins· loading · loading
Machine Learning Deep Learning 🏢 UC San Diego
IsoNet++ iteratively refines subgraph matching via early interaction GNNs and node-pair partner interactions, significantly boosting graph retrieval accuracy.
Inexact Augmented Lagrangian Methods for Conic Optimization: Quadratic Growth and Linear Convergence
·1589 words·8 mins· loading · loading
AI Theory Optimization 🏢 UC San Diego
This paper proves that inexact ALMs applied to SDPs achieve linear convergence for both primal and dual iterates, contingent solely on strict complementarity and a bounded solution set, thus resol…
HYSYNTH: Context-Free LLM Approximation for Guiding Program Synthesis
·2870 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC San Diego
HYSYNTH: A hybrid approach uses LLMs to create context-free surrogate models that guide efficient program synthesis, outperforming LLMs alone and existing synthesizers across multiple domains.
Efficient LLM Scheduling by Learning to Rank
·2254 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC San Diego
Learning to rank request outputs improves LLM scheduling, resulting in 2.8x lower chatbot latency and 6.5x higher synthetic data generation throughput.
Distribution Learning with Valid Outputs Beyond the Worst-Case
·320 words·2 mins· loading · loading
AI Theory Optimization 🏢 UC San Diego
Generative models often produce invalid outputs; this work shows that ensuring validity is easier than expected when using log-loss and carefully selecting model classes and data distributions.