🏢 UC Los Angeles
SmallToLarge (S2L): Scalable Data Selection for Fine-tuning Large Language Models by Summarizing Training Trajectories of Small Models
·2599 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
SMALLTOLARGE (S2L) revolutionizes large language model (LLM) fine-tuning by using a small model to summarize training loss trajectories, enabling efficient data selection for larger models.
SimGen: Simulator-conditioned Driving Scene Generation
·2751 words·13 mins·
loading
·
loading
AI Applications
Autonomous Vehicles
🏢 UC Los Angeles
SimGen: Simulator-conditioned driving scene generation, uses a novel cascade diffusion pipeline to generate diverse driving scenes by mixing real-world and simulator data, addressing Sim2Real gaps.
Shared Autonomy with IDA: Interventional Diffusion Assistance
·1793 words·9 mins·
loading
·
loading
AI Applications
Robotics
🏢 UC Los Angeles
IDA, a novel intervention assistance, dynamically shares control between human and AI copilots by intervening only when the AI’s action is superior across all goals, maximizing performance and preserv…
Self-Play Fine-tuning of Diffusion Models for Text-to-image Generation
·4025 words·19 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 UC Los Angeles
Self-Play Fine-Tuning (SPIN-Diffusion) revolutionizes diffusion model training, achieving superior text-to-image results with less data via iterative self-improvement, surpassing supervised and RLHF m…
Scaling transformer neural networks for skillful and reliable medium-range weather forecasting
·2398 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 UC Los Angeles
Stormer, a simple transformer model, achieves state-of-the-art medium-range weather forecasting accuracy by using weather-specific embedding, randomized dynamics forecasting, and a pressure-weighted l…
SafeWorld: Geo-Diverse Safety Alignment
·3977 words·19 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
SAFEWORLD: a new benchmark reveals and fixes LLMs’ struggle with diverse safety standards.
Reciprocal Reward Influence Encourages Cooperation From Self-Interested Agents
·1896 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 UC Los Angeles
Reciprocators: AI agents that learn to cooperate by reciprocating influence, achieving prosocial outcomes in complex scenarios.
PureGen: Universal Data Purification for Train-Time Poison Defense via Generative Model Dynamics
·3344 words·16 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 UC Los Angeles
PUREGEN uses generative model dynamics to purify poisoned training data, providing a universal, effective, and efficient train-time defense against various data poisoning attacks.
Probing the Decision Boundaries of In-context Learning in Large Language Models
·3963 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
LLMs’ in-context learning, though effective, exhibits surprisingly irregular decision boundaries, hindering generalization; this paper reveals this issue and proposes methods to improve smoothness via…
Pre-trained Large Language Models Use Fourier Features to Compute Addition
·7726 words·37 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
Pre-trained LLMs surprisingly use Fourier features to perform addition, with MLP layers approximating magnitude and attention layers handling modular arithmetic; this mechanism requires pre-training.
Physics-Informed Regularization for Domain-Agnostic Dynamical System Modeling
·1913 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 UC Los Angeles
TREAT: a novel framework boosting dynamical systems modeling accuracy by enforcing Time-Reversal Symmetry (TRS) via a regularization term. High-precision modeling is achieved across diverse systems, …
Non-Euclidean Mixture Model for Social Network Embedding
·2185 words·11 mins·
loading
·
loading
Machine Learning
Representation Learning
🏢 UC Los Angeles
Non-Euclidean Mixture Model (NMM-GNN) outperforms existing methods by using spherical and hyperbolic spaces to model homophily and social influence in social network embedding, improving link predicti…
Matryoshka Query Transformer for Large Vision-Language Models
·1913 words·9 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 UC Los Angeles
Matryoshka Query Transformer (MQT) empowers large vision-language models with flexible visual token encoding, drastically reducing inference costs while maintaining high accuracy across multiple bench…
Matching the Statistical Query Lower Bound for $k$-Sparse Parity Problems with Sign Stochastic Gradient Descent
·2323 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 UC Los Angeles
Sign Stochastic Gradient Descent (SGD) achieves optimal sample complexity for solving k-sparse parity problems, matching Statistical Query lower bounds.
Latent Plan Transformer for Trajectory Abstraction: Planning as Latent Space Inference
·2609 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 UC Los Angeles
Latent Plan Transformer (LPT) solves long-term planning challenges in reinforcement learning by using latent variables to connect trajectory generation with final returns, achieving competitive result…
Integrating Deep Metric Learning with Coreset for Active Learning in 3D Segmentation
·2210 words·11 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 UC Los Angeles
Deep metric learning and Coreset integration enables efficient slice-based active learning for 3D medical segmentation, surpassing existing methods in performance with low annotation budgets.
GraphVis: Boosting LLMs with Visual Knowledge Graph Integration
·2376 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
GraphVis boosts LLMs by visualizing knowledge graphs, improving accuracy in textual and visual question answering.
Flow Priors for Linear Inverse Problems via Iterative Corrupted Trajectory Matching
·2171 words·11 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 UC Los Angeles
ICTM efficiently solves linear inverse problems using flow priors by iteratively optimizing local MAP objectives, outperforming other flow-based methods.
Fast Sampling via Discrete Non-Markov Diffusion Models with Predetermined Transition Time
·2326 words·11 mins·
loading
·
loading
Natural Language Processing
Text Generation
🏢 UC Los Angeles
Accelerated discrete diffusion model sampling is achieved via novel discrete non-Markov diffusion models (DNDM) with predetermined transition times, enabling a training-free algorithm that significant…
Enhancing Large Vision Language Models with Self-Training on Image Comprehension
·3514 words·17 mins·
loading
·
loading
AI Generated
Natural Language Processing
Vision-Language Models
🏢 UC Los Angeles
Self-Training on Image Comprehension (STIC) significantly boosts Large Vision Language Model (LVLM) performance using unlabeled image data. STIC generates a preference dataset for image descriptions …