Skip to main content

🏢 Cornell University

WorldCoder, a Model-Based LLM Agent: Building World Models by Writing Code and Interacting with the Environment
·3322 words·16 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
WorldCoder: an LLM agent builds world models via code generation and interaction, proving highly sample-efficient and enabling knowledge transfer.
User-item fairness tradeoffs in recommendations
·3047 words·15 mins· loading · loading
AI Theory Fairness 🏢 Cornell University
Recommendation systems must balance user satisfaction with fair item exposure. This research provides a theoretical model and empirical validation showing that user preference diversity can significan…
Unified Insights: Harnessing Multi-modal Data for Phenotype Imputation via View Decoupling
·1912 words·9 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 Cornell University
MPI: A novel framework harnesses multi-modal biological data via view decoupling for superior phenotype imputation.
The Poisson Midpoint Method for Langevin Dynamics: Provably Efficient Discretization for Diffusion Models
·2297 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 Cornell University
Poisson Midpoint Method quadratically accelerates Langevin Monte Carlo for diffusion models, achieving high-quality image generation with significantly fewer computations.
The Mamba in the Llama: Distilling and Accelerating Hybrid Models
·2037 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
This research dramatically accelerates and improves hybrid language models by distilling large Transformers into linear RNNs, achieving performance comparable to the original Transformer with signific…
The Limits of Transfer Reinforcement Learning with Latent Low-rank Structure
·1762 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Cornell University
This paper presents computationally efficient transfer reinforcement learning algorithms that remove the dependence on state/action space sizes while achieving minimax optimality.
The Collusion of Memory and Nonlinearity in Stochastic Approximation With Constant Stepsize
·1378 words·7 mins· loading · loading
Reinforcement Learning 🏢 Cornell University
Unlocking the mysteries of stochastic approximation with constant stepsize, this paper reveals how memory and nonlinearity interact to create bias, providing novel analysis and solutions for more accu…
Supervised Kernel Thinning
·1753 words·9 mins· loading · loading
AI Generated Machine Learning Supervised Learning 🏢 Cornell University
Supervised Kernel Thinning accelerates kernel regression by cleverly compressing data, achieving quadratic speedups in training and inference with minimal accuracy loss.
Sample Complexity of Posted Pricing for a Single Item
·273 words·2 mins· loading · loading
AI Theory Optimization 🏢 Cornell University
This paper reveals how many buyer samples are needed to set near-optimal posted prices for a single item, resolving a fundamental problem in online markets and offering both theoretical and practical …
REBEL: Reinforcement Learning via Regressing Relative Rewards
·2652 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Cornell University
REBEL, a novel reinforcement learning algorithm, simplifies policy optimization by regressing relative rewards, achieving strong performance in language and image generation tasks with increased effic…
Qualitative Mechanism Independence
·1560 words·8 mins· loading · loading
AI Theory Causality 🏢 Cornell University
Researchers introduce QIM-compatibility, a novel framework for modeling qualitative relationships in probability distributions using directed hypergraphs, significantly expanding beyond standard condi…
QTIP: Quantization with Trellises and Incoherence Processing
·2586 words·13 mins· loading · loading
Large Language Models 🏢 Cornell University
QTIP: Ultra-high dimensional LLM quantization using trellis codes for faster, higher-quality inference.
Neural Gaffer: Relighting Any Object via Diffusion
·2042 words·10 mins· loading · loading
Computer Vision Image Generation 🏢 Cornell University
Neural Gaffer: Relighting any object via diffusion using a single image and an environment map to produce high-quality, realistic relit images.
Navigating Chemical Space with Latent Flows
·2900 words·14 mins· loading · loading
Machine Learning Deep Learning 🏢 Cornell University
ChemFlow: a new framework efficiently explores chemical space using latent flows, unifying existing methods & incorporating physical priors for molecule manipulation and optimization.
Multivariate Stochastic Dominance via Optimal Transport and Applications to Models Benchmarking
·1726 words·9 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Cornell University
This paper introduces an efficient multivariate stochastic dominance test using optimal transport, enabling robust model benchmarking by considering metric dependencies.
Microstructures and Accuracy of Graph Recall by Large Language Models
·2595 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
LLMs struggle with graph recall, exhibiting biases like favoring triangles and underperforming compared to humans; advanced models show striking domain dependence.
Language Generation in the Limit
·244 words·2 mins· loading · loading
🏢 Cornell University
This paper proves that language generation in the limit is always possible, even with an adversarial setting, contrasting with the impossibility of language identification in the limit.
Is Programming by Example solved by LLMs?
·2523 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
Large Language Models (LLMs) surprisingly improve the challenging task of Programming by Example (PBE) when fine-tuned on problem-specific data, outperforming classic symbolic methods and even surpass…
Intervention and Conditioning in Causal Bayesian Networks
·296 words·2 mins· loading · loading
AI Theory Causality 🏢 Cornell University
Researchers uniquely estimate probabilities in Causal Bayesian Networks using simple independence assumptions, enabling analysis from observational data and simplifying counterfactual probability calc…
Hybrid Top-Down Global Causal Discovery with Local Search for Linear and Nonlinear Additive Noise Models
·2421 words·12 mins· loading · loading
AI Theory Causality 🏢 Cornell University
Hybrid causal discovery algorithm efficiently learns unique causal graphs from observational data by leveraging local substructures and topological sorting, outperforming existing methods in accuracy …