๐ข UC San Diego
VLG-CBM: Training Concept Bottleneck Models with Vision-Language Guidance
ยท3011 wordsยท15 minsยท
loading
ยท
loading
Computer Vision
Visual Question Answering
๐ข UC San Diego
VLG-CBM enhances concept bottleneck models with vision-language guidance for faithful interpretability and improved accuracy.
UniMTS: Unified Pre-training for Motion Time Series
ยท2033 wordsยท10 minsยท
loading
ยท
loading
AI Applications
Healthcare
๐ข UC San Diego
UniMTS, a novel unified pre-training model, achieves state-of-the-art performance in motion time series classification by generalizing across diverse device factors and activities.
Towards Diverse Device Heterogeneous Federated Learning via Task Arithmetic Knowledge Integration
ยท2932 wordsยท14 minsยท
loading
ยท
loading
Machine Learning
Federated Learning
๐ข UC San Diego
TAKFL, a novel federated learning framework, tackles device heterogeneity by independently distilling knowledge from diverse devices and integrating it adaptively, achieving state-of-the-art performanโฆ
Symmetry-Informed Governing Equation Discovery
ยท2189 wordsยท11 minsยท
loading
ยท
loading
Machine Learning
Deep Learning
๐ข UC San Diego
Leveraging symmetry in automated equation discovery improves accuracy and simplicity of learned governing equations, enhancing robustness against noise and achieving higher success rates across diversโฆ
SpatialRGPT: Grounded Spatial Reasoning in Vision-Language Models
ยท3123 wordsยท15 minsยท
loading
ยท
loading
Multimodal Learning
Vision-Language Models
๐ข UC San Diego
SpatialRGPT enhances Vision-Language Modelsโ spatial reasoning by integrating 3D scene graphs and depth information, achieving significant performance gains on spatial reasoning tasks.
Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
ยท2661 wordsยท13 minsยท
loading
ยท
loading
Machine Learning
Deep Learning
๐ข UC San Diego
SharpBalance, a novel training approach, effectively improves deep ensemble performance by addressing the sharpness-diversity trade-off, leading to significant improvements in both in-distribution andโฆ
Replicable Uniformity Testing
ยท268 wordsยท2 minsยท
loading
ยท
loading
AI Generated
AI Theory
Optimization
๐ข UC San Diego
This paper presents the first replicable uniformity tester with nearly linear dependence on the replicability parameter, enhancing the reliability of scientific studies using distribution testing algoโฆ
RA-PbRL: Provably Efficient Risk-Aware Preference-Based Reinforcement Learning
ยท1598 wordsยท8 minsยท
loading
ยท
loading
Machine Learning
Reinforcement Learning
๐ข UC San Diego
RA-PbRL introduces a provably efficient algorithm for risk-aware preference-based reinforcement learning, addressing the limitations of existing risk-neutral methods in applications demanding heightenโฆ
Quantifying and Optimizing Global Faithfulness in Persona-driven Role-playing
ยท2570 wordsยท13 minsยท
loading
ยท
loading
Natural Language Processing
Dialogue Systems
๐ข UC San Diego
New APC metric precisely quantifies & optimizes global faithfulness in persona-driven role-playing, offering a fine-grained, explainable evaluation and improving AI character consistency.
Provable and Efficient Dataset Distillation for Kernel Ridge Regression
ยท1601 wordsยท8 minsยท
loading
ยท
loading
Machine Learning
Deep Learning
๐ข UC San Diego
One data point per class suffices for efficient and provable dataset distillation in kernel ridge regression, significantly reducing computational costs.
Online Feature Updates Improve Online (Generalized) Label Shift Adaptation
ยท1991 wordsยท10 minsยท
loading
ยท
loading
Machine Learning
Self-Supervised Learning
๐ข UC San Diego
Online Label Shift adaptation with Online Feature Updates (OLS-OFU) significantly boosts online label shift adaptation by dynamically refining feature extractors using self-supervised learning, achievโฆ
Online Consistency of the Nearest Neighbor Rule
ยท1388 wordsยท7 minsยท
loading
ยท
loading
AI Theory
Optimization
๐ข UC San Diego
The 1-nearest neighbor rule achieves online consistency under surprisingly broad conditions: measurable label functions and mild assumptions on instance generation in doubling metric spaces.
On Differentially Private U Statistics
ยท403 wordsยท2 minsยท
loading
ยท
loading
AI Theory
Privacy
๐ข UC San Diego
New algorithms achieve near-optimal differentially private U-statistic estimation, significantly improving accuracy over existing methods.
Linking In-context Learning in Transformers to Human Episodic Memory
ยท3883 wordsยท19 minsยท
loading
ยท
loading
AI Generated
Natural Language Processing
Large Language Models
๐ข UC San Diego
Transformersโ in-context learning mirrors human episodic memory, with specific attention heads acting like the brainโs contextual maintenance and retrieval system.
Learning from Uncertain Data: From Possible Worlds to Possible Models
ยท2895 wordsยท14 minsยท
loading
ยท
loading
AI Theory
Robustness
๐ข UC San Diego
ZORRO: A new method for learning linear models from uncertain data, providing sound over-approximations of all possible models and prediction ranges.
Iteratively Refined Early Interaction Alignment for Subgraph Matching based Graph Retrieval
ยท3157 wordsยท15 minsยท
loading
ยท
loading
Machine Learning
Deep Learning
๐ข UC San Diego
IsoNet++ iteratively refines subgraph matching via early interaction GNNs and node-pair partner interactions, significantly boosting graph retrieval accuracy.
Inexact Augmented Lagrangian Methods for Conic Optimization: Quadratic Growth and Linear Convergence
ยท1589 wordsยท8 minsยท
loading
ยท
loading
AI Theory
Optimization
๐ข UC San Diego
This paper proves that inexact ALMs applied to SDPs achieve linear convergence for both primal and dual iterates, contingent solely on strict complementarity and a bounded solution set, thus resolโฆ
HYSYNTH: Context-Free LLM Approximation for Guiding Program Synthesis
ยท2870 wordsยท14 minsยท
loading
ยท
loading
AI Generated
Natural Language Processing
Large Language Models
๐ข UC San Diego
HYSYNTH: A hybrid approach uses LLMs to create context-free surrogate models that guide efficient program synthesis, outperforming LLMs alone and existing synthesizers across multiple domains.
Efficient LLM Scheduling by Learning to Rank
ยท2254 wordsยท11 minsยท
loading
ยท
loading
Natural Language Processing
Large Language Models
๐ข UC San Diego
Learning to rank request outputs improves LLM scheduling, resulting in 2.8x lower chatbot latency and 6.5x higher synthetic data generation throughput.
Distribution Learning with Valid Outputs Beyond the Worst-Case
ยท320 wordsยท2 minsยท
loading
ยท
loading
AI Theory
Optimization
๐ข UC San Diego
Generative models often produce invalid outputs; this work shows that ensuring validity is easier than expected when using log-loss and carefully selecting model classes and data distributions.