🏢 Harvard University
Generalized Protein Pocket Generation with Prior-Informed Flow Matching
·1970 words·10 mins·
loading
·
loading
🏢 Harvard University
PocketFlow: a novel generative model designs high-affinity protein pockets using prior-informed flow matching, outperforming existing methods.
FinCon: A Synthesized LLM Multi-Agent System with Conceptual Verbal Reinforcement for Enhanced Financial Decision Making
·3232 words·16 mins·
loading
·
loading
AI Generated
AI Applications
Finance
🏢 Harvard University
FINCON: an LLM-based multi-agent system uses conceptual verbal reinforcement for superior financial decision-making, generalizing well across various tasks.
Fast TRAC: A Parameter-Free Optimizer for Lifelong Reinforcement Learning
·2979 words·14 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Harvard University
TRAC: a parameter-free optimizer conquering lifelong RL’s plasticity loss!
Eye-gaze Guided Multi-modal Alignment for Medical Representation Learning
·2845 words·14 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 Harvard University
Eye-gaze data boosts medical image-text alignment!
Evaluating the World Model Implicit in a Generative Model
·4059 words·20 mins·
loading
·
loading
Large Language Models
🏢 Harvard University
New metrics reveal that generative models often possess surprisingly incoherent world models, despite seemingly accurate next-token predictions. This incoherence leads to fragility in solving related …
Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space
·3166 words·15 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Harvard University
Generative models learn hidden capabilities suddenly during training, which can be explained and predicted using a novel ‘concept space’ framework that analyzes learning dynamics and concept signal.
Dissecting the Interplay of Attention Paths in a Statistical Mechanics Theory of Transformers
·2710 words·13 mins·
loading
·
loading
AI Generated
AI Theory
Interpretability
🏢 Harvard University
Researchers dissected attention paths in Transformers using statistical mechanics, revealing a task-relevant kernel combination mechanism boosting generalization performance.
Covariate Shift Corrected Conditional Randomization Test
·2259 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 Harvard University
A new Covariate Shift Corrected Pearson Chi-squared Conditional Randomization (csPCR) test accurately assesses conditional independence even when data distributions vary between source and target popu…
Carrot and Stick: Eliciting Comparison Data and Beyond
·1825 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Harvard University
Truthful comparison data is hard to obtain without ground truth. This paper presents novel peer prediction mechanisms using bonus-penalty payments that incentivize truthful comparisons, even in networ…
Bias Detection via Signaling
·295 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 Harvard University
This paper presents efficient algorithms to detect whether an agent updates beliefs optimally (Bayesian) or exhibits bias towards their prior beliefs, using information design and signaling schemes.
Axioms for AI Alignment from Human Feedback
·1869 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Harvard University
This paper revolutionizes AI alignment by applying social choice theory axioms to RLHF, exposing flaws in existing methods and proposing novel, axiomatically guaranteed reward learning rules.
Approximating mutual information of high-dimensional variables using learned representations
·2528 words·12 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Harvard University
Latent Mutual Information (LMI) approximation accurately estimates mutual information in high-dimensional data using low-dimensional learned representations, solving a critical problem in various scie…
A teacher-teacher framework for clinical language representation learning
·1643 words·8 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Harvard University
A lightweight knowledge alignment module enables two pre-trained LLMs to mutually learn and improve clinical language representation, exceeding individual model performance on various downstream tasks…
A Label is Worth A Thousand Images in Dataset Distillation
·2824 words·14 mins·
loading
·
loading
Computer Vision
Image Classification
🏢 Harvard University
Soft labels, not sophisticated data synthesis, are the key to successful dataset distillation, significantly improving data-efficient learning and challenging existing methods.
A Closer Look at AUROC and AUPRC under Class Imbalance
·2353 words·12 mins·
loading
·
loading
AI Theory
Fairness
🏢 Harvard University
Debunking a common myth, this paper proves that AUPRC is not superior to AUROC for imbalanced datasets, and in fact, can worsen algorithmic bias.