Skip to main content

🏢 UC Berkeley

Learning to Assist Humans without Inferring Rewards
·1833 words·9 mins· loading · loading
AI Applications Human-AI Interaction 🏢 UC Berkeley
AI agents trained with Empowerment via Successor Representations (ESR) empower humans by maximizing their control over environmental outcomes, eliminating the need for human intention inference, unlik…
Latent Learning Progress Drives Autonomous Goal Selection in Human Reinforcement Learning
·2787 words·14 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 UC Berkeley
Humans autonomously select goals based on both observed and latent learning progress, impacting goal-conditioned policy learning.
Large Stepsize Gradient Descent for Non-Homogeneous Two-Layer Networks: Margin Improvement and Fast Optimization
·2166 words·11 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Berkeley
Large stepsize GD on non-homogeneous neural networks shows monotonic risk reduction after an initial oscillating phase, demonstrating implicit bias and optimization gains.
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
·5270 words·25 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC Berkeley
KVQuant achieves <0.1 perplexity degradation with 3-bit quantization in LLMs by using per-channel key quantization, pre-RoPE quantization, and non-uniform quantization, enabling 10M context length inf…
Is Value Learning Really the Main Bottleneck in Offline RL?
·2601 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC Berkeley
Offline RL’s performance often lags behind imitation learning, but this paper reveals that policy learning and generalization, not value function learning, are often the main bottlenecks.
Is Knowledge Power? On the (Im)possibility of Learning from Strategic Interactions
·385 words·2 mins· loading · loading
AI Theory Optimization 🏢 UC Berkeley
In strategic settings, repeated interactions alone may not enable uninformed players to achieve optimal outcomes, highlighting the persistent impact of information asymmetry.
Interpreting the Weight Space of Customized Diffusion Models
·3822 words·18 mins· loading · loading
Computer Vision Image Generation 🏢 UC Berkeley
Researchers model a manifold of customized diffusion models as a subspace of weights, enabling controllable creation of new models via sampling, editing, and inversion from a single image.
In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization
·436 words·3 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC Berkeley
Linear Transformer Blocks (LTBs) achieve near-optimal in-context learning (ICL) for linear regression by effectively implementing one-step gradient descent with learnable initialization, a significant…
Immiscible Diffusion: Accelerating Diffusion Training with Noise Assignment
·3506 words·17 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 UC Berkeley
Immiscible Diffusion boosts diffusion model training efficiency up to 3x by cleverly assigning noise to images, preventing the mixing of data in noise space and thus improving optimization.
How many classifiers do we need?
·1821 words·9 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Berkeley
Boost ensemble accuracy by predicting performance with fewer classifiers using a novel polarization law and refined error bounds.
Gorilla: Large Language Model Connected with Massive APIs
·2454 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
Gorilla: a fine-tuned LLaMA model surpasses GPT-4 in generating accurate API calls by using Retriever Aware Training (RAT) to adapt to changing APIs and reduce hallucinations.
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning
·3674 words·18 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 UC Berkeley
This paper presents a novel RL framework that fine-tunes large vision-language models (VLMs) to become effective decision-making agents. By incorporating chain-of-thought reasoning, the framework enab…
Fair Allocation in Dynamic Mechanism Design
·2010 words·10 mins· loading · loading
AI Theory Fairness 🏢 UC Berkeley
This paper presents optimal fair mechanisms for dynamic auction design, maximizing seller revenue while guaranteeing minimum allocations to multiple buyer groups.
Explaining Datasets in Words: Statistical Models with Natural Language Parameters
·2281 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
This paper introduces a model-agnostic algorithm that uses natural language predicates to make statistical model parameters directly interpretable, significantly improving explainability.
Evidence of Learned Look-Ahead in a Chess-Playing Neural Network
·4142 words·20 mins· loading · loading
AI Generated AI Theory Interpretability 🏢 UC Berkeley
Chess AI Leela Zero surprisingly uses learned look-ahead, internally representing future optimal moves, significantly improving its strategic decision-making.
Evaluating the design space of diffusion-based generative models
·378 words·2 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Berkeley
This paper provides the first complete error analysis for diffusion models, theoretically justifying optimal training and sampling strategies and design choices for enhanced generative capabilities.
Enhancing Efficiency of Safe Reinforcement Learning via Sample Manipulation
·1946 words·10 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC Berkeley
ESPO enhances safe RL efficiency by dynamically manipulating sample size based on reward-safety gradient conflicts, ensuring faster training and superior performance.
Dimension-free Private Mean Estimation for Anisotropic Distributions
·233 words·2 mins· loading · loading
AI Generated AI Theory Privacy 🏢 UC Berkeley
Dimension-free private mean estimation is achieved for anisotropic data, breaking the curse of dimensionality in privacy-preserving high-dimensional analysis.
DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning
·3604 words·17 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 UC Berkeley
DigiRL: Autonomous RL trains robust in-the-wild device-control agents by offline-to-online RL, surpassing prior methods.
Designing Cell-Type-Specific Promoter Sequences Using Conservative Model-Based Optimization
·2381 words·12 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 UC Berkeley
Researchers developed a data-efficient method using conservative model-based optimization to design cell-type-specific promoters for gene therapy, significantly improving cell-type specificity.