Skip to main content

🏢 UC Berkeley

Warm-starting Push-Relabel
·1936 words·10 mins· loading · loading
AI Theory Optimization 🏢 UC Berkeley
This research introduces the first theoretical guarantees for warm-starting the celebrated Push-Relabel network flow algorithm, improving its speed using a predicted flow, while maintaining worst-case…
Verified Code Transpilation with LLMs
·2009 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
LLMLIFT: An LLM-powered approach builds verified lifting tools for DSLs, outperforming prior symbolic methods in benchmark transpilation and requiring less development effort.
Using Surrogates in Covariate-adjusted Response-adaptive Randomization Experiments with Delayed Outcomes
·1375 words·7 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 UC Berkeley
Boosting clinical trial efficiency, this research introduces a covariate-adjusted response-adaptive randomization (CARA) design that effectively leverages surrogate outcomes to handle delayed primary …
Ultrafast classical phylogenetic method beats large protein language models on variant effect prediction
·2536 words·12 mins· loading · loading
AI Generated AI Theory Optimization 🏢 UC Berkeley
A revolutionary ultrafast phylogenetic method outperforms protein language models in variant effect prediction by efficiently estimating amino acid substitution rates from massive datasets.
Truthfulness of Calibration Measures
·337 words·2 mins· loading · loading
AI Theory Optimization 🏢 UC Berkeley
Researchers developed Subsampled Smooth Calibration Error (SSCE), a new truthful calibration measure for sequential prediction, solving the problem of existing measures being easily gamed.
Towards a Theoretical Understanding of the 'Reversal Curse' via Training Dynamics
·2572 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
LLMs struggle with simple logical reasoning due to the ‘reversal curse.’ This paper reveals that weight asymmetry during training is the culprit, offering a new theoretical perspective and potential s…
The Importance of Being Scalable: Improving the Speed and Accuracy of Neural Network Interatomic Potentials Across Chemical Domains
·1885 words·9 mins· loading · loading
Machine Learning Deep Learning 🏢 UC Berkeley
ESCAIP, a novel neural network architecture, dramatically boosts the speed and accuracy of atomic simulations by leveraging attention mechanisms, enabling efficient large-scale modeling across diverse…
The Impact of Initialization on LoRA Finetuning Dynamics
·2220 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
LoRA’s initialization significantly impacts finetuning; initializing matrix A randomly and B to zero yields better performance than vice-versa due to enabling larger learning rates.
Synthetic Programming Elicitation for Text-to-Code in Very Low-Resource Programming and Formal Languages
·1817 words·9 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC Berkeley
LLMs struggle with very low-resource programming languages. SPEAC, a novel synthetic programming elicitation and compilation approach, uses an intermediate language to enable LLMs to generate syntact…
SSDM: Scalable Speech Dysfluency Modeling
·2807 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
SSDM: Scalable Speech Dysfluency Modeling tackles challenges in speech dysfluency analysis by using articulatory gestures for scalable alignment, a connectionist subsequence aligner for efficient dysf…
SGLang: Efficient Execution of Structured Language Model Programs
·1898 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Berkeley
SGLang: A new system boosts LLM program execution speed by up to 6.4x, simplifying complex LLM application programming.
Segment Anything without Supervision
·1959 words·10 mins· loading · loading
Computer Vision Image Segmentation 🏢 UC Berkeley
Unsupervised SAM (UnSAM) achieves competitive image segmentation results without human annotation, surpassing previous unsupervised methods and even improving supervised SAM’s accuracy.
Secret Collusion among AI Agents: Multi-Agent Deception via Steganography
·5189 words·25 mins· loading · loading
AI Generated AI Theory Safety 🏢 UC Berkeley
AI agents can secretly collude using steganography, hiding their interactions from oversight. This research formalizes this threat, analyzes LLMs’ capabilities, and proposes mitigation strategies.
Scaling Laws in Linear Regression: Compute, Parameters, and Data
·1352 words·7 mins· loading · loading
AI Theory Optimization 🏢 UC Berkeley
Deep learning’s neural scaling laws defy conventional wisdom; this paper uses infinite-dimensional linear regression to theoretically explain this phenomenon, showing that implicit regularization of S…
Rethinking Score Distillation as a Bridge Between Image Distributions
·2251 words·11 mins· loading · loading
Computer Vision Image Generation 🏢 UC Berkeley
Researchers enhanced image generation by improving score distillation sampling via a novel Schrödinger Bridge framework, improving realism without computational overhead.
On Socially Fair Low-Rank Approximation and Column Subset Selection
·363 words·2 mins· loading · loading
AI Generated AI Theory Fairness 🏢 UC Berkeley
This paper reveals the surprising computational hardness of achieving fairness in low-rank approximation while offering efficient approximation algorithms.
Mutli-Armed Bandits with Network Interference
·1421 words·7 mins· loading · loading
AI Theory Causality 🏢 UC Berkeley
New algorithms conquer regret in multi-armed bandits challenged by network interference, achieving provably low regret with both known and unknown network structures.
Mitigating Partial Observability in Decision Processes via the Lambda Discrepancy
·2495 words·12 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC Berkeley
New metric, λ-discrepancy, precisely detects & mitigates partial observability in sequential decision processes, significantly boosting reinforcement learning agent performance.
Med-Real2Sim: Non-Invasive Medical Digital Twins using Physics-Informed Self-Supervised Learning
·2852 words·14 mins· loading · loading
AI Applications Healthcare 🏢 UC Berkeley
Med-Real2Sim uses physics-informed self-supervised learning to build non-invasive medical digital twins, enabling in-silico clinical trials and unsupervised disease detection.
Learning to Understand: Identifying Interactions via the Möbius Transform
·2143 words·11 mins· loading · loading
AI Theory Interpretability 🏢 UC Berkeley
Unlocking complex models’ secrets: New algorithm identifies input interactions using the Möbius Transform, boosting interpretability with surprising speed and accuracy.