🏢 Stanford University
Higher-Order Causal Message Passing for Experimentation with Complex Interference
·1660 words·8 mins·
loading
·
loading
AI Theory
Causality
🏢 Stanford University
Higher-Order Causal Message Passing (HO-CMP) accurately estimates treatment effects in complex systems with unknown interference by using observed data to learn the system’s dynamics over time.
Grasp as You Say: Language-guided Dexterous Grasp Generation
·2373 words·12 mins·
loading
·
loading
AI Applications
Robotics
🏢 Stanford University
Robots can now dexterously grasp objects based on natural language commands thanks to DexGYS, a new language-guided dexterous grasp generation framework and dataset.
GraphMETRO: Mitigating Complex Graph Distribution Shifts via Mixture of Aligned Experts
·2217 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Stanford University
GraphMETRO tackles complex graph distribution shifts by using a Mixture-of-Experts model to decompose shifts into interpretable components, achieving state-of-the-art results.
Graph-based Uncertainty Metrics for Long-form Language Model Generations
·2055 words·10 mins·
loading
·
loading
Large Language Models
🏢 Stanford University
Graph Uncertainty boosts LLM factuality by 6.8% using graph centrality to estimate claim-level uncertainty and a novel uncertainty-aware decoding process.
Get rich quick: exact solutions reveal how unbalanced initializations promote rapid feature learning
·2253 words·11 mins·
loading
·
loading
AI Theory
Optimization
🏢 Stanford University
Unbalanced initializations dramatically accelerate neural network feature learning by modifying the geometry of learning trajectories, enabling faster feature extraction and improved generalization.
Geometric Trajectory Diffusion Models
·3387 words·16 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Stanford University
GeoTDM: First diffusion model generating realistic 3D geometric trajectories, capturing complex spatial interactions and temporal correspondence, significantly improving generation quality.
Generalized Linear Bandits with Limited Adaptivity
·341 words·2 mins·
loading
·
loading
Reinforcement Learning
🏢 Stanford University
This paper introduces two novel algorithms, achieving optimal regret in generalized linear contextual bandits despite limited policy updates, a significant advancement for real-world applications.
FactorSim: Generative Simulation via Factorized Representation
·2722 words·13 mins·
loading
·
loading
AI Applications
Robotics
🏢 Stanford University
FACTORSim generates full, coded simulations from natural language descriptions, outperforming existing methods in accuracy and zero-shot transfer learning by using a factored POMDP representation.
Distributionally Robust Reinforcement Learning with Interactive Data Collection: Fundamental Hardness and Near-Optimal Algorithms
·518 words·3 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Stanford University
Provably sample-efficient robust RL via interactive data collection is achieved by introducing the vanishing minimal value assumption to mitigate the curse of support shift, enabling near-optimal algo…
Directional Smoothness and Gradient Methods: Convergence and Adaptivity
·1502 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Stanford University
New sub-optimality bounds for gradient descent leverage directional smoothness, a localized gradient variation measure, achieving tighter convergence guarantees and adapting to optimization paths.
Depth Anywhere: Enhancing 360 Monocular Depth Estimation via Perspective Distillation and Unlabeled Data Augmentation
·2664 words·13 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Stanford University
Depth Anywhere enhances 360-degree monocular depth estimation by cleverly using perspective models to label unlabeled 360-degree data, significantly improving accuracy.
Deep Learning for Computing Convergence Rates of Markov Chains
·1326 words·7 mins·
loading
·
loading
🏢 Stanford University
Deep learning tackles Markov chain convergence rate analysis! Deep Contractive Drift Calculator (DCDC) provides sample-based bounds in Wasserstein distance, surpassing traditional methods’ limitations…
CRONOS: Enhancing Deep Learning with Scalable GPU Accelerated Convex Neural Networks
·2029 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Stanford University
CRONOS: Scaling convex neural network training to ImageNet!
Convolutional Differentiable Logic Gate Networks
·2283 words·11 mins·
loading
·
loading
🏢 Stanford University
Convolutional Differentiable Logic Gate Networks achieve state-of-the-art accuracy on CIFAR-10 with 29x fewer gates than existing models, demonstrating highly efficient deep learning inference.
Constrained Diffusion with Trust Sampling
·2019 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Stanford University
Trust Sampling enhances guided diffusion by iteratively optimizing constrained generation at each step, improving efficiency and accuracy in image and 3D motion generation.
Consistency of Neural Causal Partial Identification
·2971 words·14 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 Stanford University
Neural causal models consistently estimate partial causal effects, even with continuous/categorical variables, thanks to Lipschitz regularization.
Compressing Large Language Models using Low Rank and Low Precision Decomposition
·2393 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Stanford University
CALDERA: a new post-training LLM compression algorithm achieving state-of-the-art zero-shot performance using low-rank, low-precision decomposition.
Collaborative Video Diffusion: Consistent Multi-video Generation with Camera Control
·2588 words·13 mins·
loading
·
loading
Computer Vision
Video Understanding
🏢 Stanford University
Collaborative Video Diffusion (CVD) generates multiple consistent videos from various camera angles using a novel cross-video synchronization module, significantly improving consistency compared to ex…
Boosted Conformal Prediction Intervals
·4433 words·21 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Stanford University
Boosting conformal prediction intervals improves accuracy and precision by tailoring them to specific desired properties via machine learning.
Bayesian Strategic Classification
·315 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 Stanford University
Learners can improve accuracy in strategic classification by selectively revealing partial classifier information to agents, strategically guiding agent behavior and maximizing accuracy.