🏢 MIT
NeuralFluid: Nueral Fluidic System Design and Control with Differentiable Simulation
·1858 words·9 mins·
loading
·
loading
AI Applications
Robotics
🏢 MIT
NeuralFluid: Design & control complex fluidic systems with dynamic boundaries using differentiable simulation, achieving superior results in benchmark tasks.
Model-Based Transfer Learning for Contextual Reinforcement Learning
·3972 words·19 mins·
loading
·
loading
AI Applications
Smart Cities
🏢 MIT
Model-Based Transfer Learning (MBTL) boosts deep RL sample efficiency by strategically selecting training tasks, achieving up to 50x improvement over traditional methods.
MeMo: Meaningful, Modular Controllers via Noise Injection
·3100 words·15 mins·
loading
·
loading
AI Applications
Robotics
🏢 MIT
MeMo: a novel framework for pretraining meaningful, modular robot controllers via noise injection, enabling efficient transfer learning across different robot morphologies and tasks.
Measuring Progress in Dictionary Learning for Language Model Interpretability with Board Game Models
·2097 words·10 mins·
loading
·
loading
Natural Language Processing
Interpretability
🏢 MIT
New metrics and p-annealing improve sparse autoencoder training for better language model interpretability.
Mean-Field Analysis for Learning Subspace-Sparse Polynomials with Gaussian Input
·272 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 MIT
Researchers establish basis-free conditions for SGD learnability in two-layer neural networks learning subspace-sparse polynomials with Gaussian input, offering insights into training dynamics.
Maximizing utility in multi-agent environments by anticipating the behavior of other learners
·1732 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 MIT
Optimizing against learning agents: New algorithms and computational limits revealed!
Long-Horizon Planning for Multi-Agent Robots in Partially Observable Environments
·2334 words·11 mins·
loading
·
loading
AI Applications
Robotics
🏢 MIT
LLaMAR: LM-based planner for multi-agent robots excels in long-horizon, partially observable tasks, achieving 30% higher success than existing methods.
Learning Multimodal Behaviors from Scratch with Diffusion Policy Gradient
·3397 words·16 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 MIT
DDiffPG: A novel actor-critic algorithm learns multimodal policies from scratch using diffusion models, enabling agents to master versatile behaviors in complex tasks.
Learning Mixtures of Unknown Causal Interventions
·2082 words·10 mins·
loading
·
loading
AI Theory
Causality
🏢 MIT
Researchers developed an efficient algorithm to uniquely identify causal relationships from mixed interventional and observational data with noisy interventions.
Interpolating Item and User Fairness in Multi-Sided Recommendations
·1620 words·8 mins·
loading
·
loading
AI Theory
Fairness
🏢 MIT
Problem (FAIR) framework and FORM algorithm achieve flexible multi-stakeholder fairness in online recommendation systems, balancing platform revenue with user and item fairness.
Improving Subgroup Robustness via Data Selection
·1691 words·8 mins·
loading
·
loading
AI Theory
Robustness
🏢 MIT
Data Debiasing with Datamodels (D3M) efficiently improves machine learning model robustness by identifying and removing specific training examples that disproportionately harm minority groups’ accurac…
Identifiability Guarantees for Causal Disentanglement from Purely Observational Data
·1394 words·7 mins·
loading
·
loading
AI Theory
Causality
🏢 MIT
This paper provides identifiability guarantees for causal disentanglement from purely observational data using nonlinear additive Gaussian noise models, addressing a major challenge in causal represen…
How Does Variance Shape the Regret in Contextual Bandits?
·334 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 MIT
Low reward variance drastically improves contextual bandit regret, defying minimax assumptions and highlighting the crucial role of eluder dimension.
How Diffusion Models Learn to Factorize and Compose
·3926 words·19 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 MIT
Diffusion models surprisingly learn factorized representations, enabling compositional generalization, but struggle with interpolation; training with independent factors drastically improves data effi…
Hamiltonian Score Matching and Generative Flows
·1465 words·7 mins·
loading
·
loading
Machine Learning
Generative Modeling
🏢 MIT
Hamiltonian Generative Flows (HGFs) revolutionize generative modeling by leveraging Hamiltonian dynamics, offering enhanced score matching and generative capabilities.
Generative Modeling of Molecular Dynamics Trajectories
·2510 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 MIT
MDGEN: Generative modeling unlocks MD data for diverse tasks, achieving significant speedups via flexible multi-task surrogate models.
Flexible mapping of abstract domains by grid cells via self-supervised extraction and projection of generalized velocity signals
·2121 words·10 mins·
loading
·
loading
Machine Learning
Self-Supervised Learning
🏢 MIT
Brain’s flexible mapping of abstract domains is achieved via self-supervised extraction and projection of generalized velocity signals by grid cells, enabling efficient map generation.
Flexible Context-Driven Sensory Processing in Dynamical Vision Models
·2040 words·10 mins·
loading
·
loading
Computer Vision
Vision-Language Models
🏢 MIT
Biologically-inspired DCnet neural network flexibly modulates visual processing based on context, outperforming existing models on visual search and attention tasks.
Few-Shot Task Learning through Inverse Generative Modeling
·2587 words·13 mins·
loading
·
loading
Machine Learning
Few-Shot Learning
🏢 MIT
Few-shot task learning through inverse generative modeling (FTL-IGM) enables AI agents to quickly master new tasks from minimal data by leveraging invertible generative models.
Fair Wasserstein Coresets
·2137 words·11 mins·
loading
·
loading
AI Theory
Fairness
🏢 MIT
Fair Wasserstein Coresets (FWC) efficiently generates fair, representative subsets of large datasets for downstream machine learning tasks, improving fairness and utility.