🏢 University of California, Berkeley
Toxicity Detection for Free
·2767 words·13 mins·
loading
·
loading
Large Language Models
🏢 University of California, Berkeley
Moderation Using LLM Introspection (MULI) leverages the first response token’s logits from LLMs to create a highly accurate toxicity detector, surpassing state-of-the-art methods with minimal overhead…
Stylus: Automatic Adapter Selection for Diffusion Models
·3022 words·15 mins·
loading
·
loading
Image Generation
🏢 University of California, Berkeley
Stylus: an automatic adapter selection system for diffusion models, boosts image quality and diversity by intelligently composing task-specific adapters based on prompt keywords.
Pretrained Transformer Efficiently Learns Low-Dimensional Target Functions In-Context
·1358 words·7 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of California, Berkeley
Pretrained transformers surprisingly learn low-dimensional nonlinear functions efficiently from few in-context examples, outperforming baseline algorithms.
Poisson Variational Autoencoder
·2239 words·11 mins·
loading
·
loading
🏢 University of California, Berkeley
Poisson Variational Autoencoder (P-VAE) improves deep learning by encoding inputs as discrete spike counts, enhancing biological realism and interpretability while avoiding posterior collapse and achi…
Overcoming the Sim-to-Real Gap: Leveraging Simulation to Learn to Explore for Real-World RL
·2419 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of California, Berkeley
Leveraging simulation for real-world RL is often hampered by the sim-to-real gap. This paper shows that instead of directly transferring policies, transferring exploratory policies from simulation d…
Metric Transforms and Low Rank Representations of Kernels for Fast Attention
·275 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of California, Berkeley
Researchers unveil novel linear-algebraic tools revealing the limits of fast attention, classifying positive definite kernels for Manhattan distance, and fully characterizing metric transforms for Man…
Learning to Mitigate Externalities: the Coase Theorem with Hindsight Rationality
·314 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of California, Berkeley
Economists learn to resolve externalities efficiently even when players lack perfect information, maximizing social welfare by leveraging bargaining and online learning.
Humanoid Locomotion as Next Token Prediction
·1485 words·7 mins·
loading
·
loading
AI Applications
Robotics
🏢 University of California, Berkeley
Humanoid robots now walk in San Francisco zero-shot, thanks to a novel ’next token prediction’ approach trained on diverse sensorimotor data, enabling real-world generalization and data efficiency.
BPQP: A Differentiable Convex Optimization Framework for Efficient End-to-End Learning
·1651 words·8 mins·
loading
·
loading
AI Applications
Finance
🏢 University of California, Berkeley
BPQP: A new differentiable convex optimization framework accelerates end-to-end learning by an order of magnitude, achieving significant efficiency gains over existing methods.
An Analysis of Tokenization: Transformers under Markov Data
·2141 words·11 mins·
loading
·
loading
Large Language Models
🏢 University of California, Berkeley
Tokenization’s crucial role in transformer language models is revealed: Transformers struggle on simple Markov data without tokenization, but achieve near-optimal performance with appropriate tok…