🏢 Carnegie Mellon University
Provably Robust Score-Based Diffusion Posterior Sampling for Plug-and-Play Image Reconstruction
·1596 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Carnegie Mellon University
Provably robust diffusion posterior sampling for plug-and-play image reconstruction is achieved via a novel algorithmic framework, DPnP, offering both asymptotic and non-asymptotic performance guarant…
Protein-Nucleic Acid Complex Modeling with Frame Averaging Transformer
·1968 words·10 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Carnegie Mellon University
Unsupervised learning predicts protein-nucleic acid binding using contact map prediction, significantly improving aptamer screening via FAFormer, a novel equivariant transformer.
Private and Personalized Frequency Estimation in a Federated Setting
·1856 words·9 mins·
loading
·
loading
AI Generated
Machine Learning
Federated Learning
🏢 Carnegie Mellon University
This paper introduces a novel privacy-preserving algorithm for personalized frequency estimation in federated settings, significantly improving accuracy and efficiency over existing methods.
Predicting the Performance of Foundation Models via Agreement-on-the-Line
·4845 words·23 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
Foundation model OOD performance prediction is reliably achieved via ensemble diversity, especially through random linear head initialization, enabling precise estimations without extensive OOD labels…
Pard: Permutation-Invariant Autoregressive Diffusion for Graph Generation
·2350 words·12 mins·
loading
·
loading
Machine Learning
Graph Generation
🏢 Carnegie Mellon University
PARD: a novel permutation-invariant autoregressive diffusion model for efficient and high-quality graph generation, achieving state-of-the-art results.
On the Surprising Effectiveness of Attention Transfer for Vision Transformers
·2971 words·14 mins·
loading
·
loading
Computer Vision
Image Classification
🏢 Carnegie Mellon University
Vision Transformers achieve surprisingly high accuracy by transferring only pre-training attention maps, challenging the conventional belief that feature learning is crucial.
On the Parameter Identifiability of Partially Observed Linear Causal Models
·3769 words·18 mins·
loading
·
loading
AI Generated
AI Theory
Causality
🏢 Carnegie Mellon University
Researchers achieve full parameter identifiability in partially observed linear causal models using novel graphical conditions and a likelihood-based estimation method, addressing previous limitations…
On the Benefits of Public Representations for Private Transfer Learning under Distribution Shift
·1927 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Privacy
🏢 Carnegie Mellon University
Public data boosts private AI accuracy even with extreme distribution shifts, improving private model training by up to 67% in three tasks. This is due to shared low-dimensional representations betwe…
Omnigrasp: Simulated Humanoid Grasping on Diverse Objects
·2619 words·13 mins·
loading
·
loading
AI Generated
AI Applications
Robotics
🏢 Carnegie Mellon University
Omnigrasp: A novel RL-based method enables simulated humanoids to grasp diverse objects and precisely follow complex trajectories, advancing realistic human-object interaction in virtual environments.
OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
·2351 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Carnegie Mellon University
OASIS, a novel data-centric approach, shapes offline data distributions toward safer, higher-reward policies using a conditional diffusion model, outperforming existing offline safe RL methods.
No Free Lunch in LLM Watermarking: Trade-offs in Watermarking Design Choices
·3353 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
LLM watermarking faces inherent trade-offs; this paper reveals simple attacks exploiting common design choices, proposing guidelines and defenses for more secure systems.
Multi-Agent Imitation Learning: Value is Easy, Regret is Hard
·1706 words·9 mins·
loading
·
loading
AI Theory
Reinforcement Learning
🏢 Carnegie Mellon University
In multi-agent imitation learning, achieving regret equivalence is harder than value equivalence; this paper introduces novel algorithms that efficiently minimize the regret gap under various assumpti…
Model-based Diffusion for Trajectory Optimization
·2272 words·11 mins·
loading
·
loading
AI Applications
Robotics
🏢 Carnegie Mellon University
Model-Based Diffusion (MBD) uses diffusion processes and model information for data-free trajectory optimization, outperforming existing methods on complex tasks.
MiSO: Optimizing brain stimulation to create neural activity states
·2684 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Carnegie Mellon University
MiSO: a novel closed-loop brain stimulation framework optimizes stimulation parameters to achieve desired neural population activity states, overcoming limitations of current methods by merging data a…
MGF: Mixed Gaussian Flow for Diverse Trajectory Prediction
·2012 words·10 mins·
loading
·
loading
AI Applications
Robotics
🏢 Carnegie Mellon University
MGF: Mixed Gaussian Flow enhances trajectory prediction by using a mixed Gaussian prior, achieving state-of-the-art diversity and alignment accuracy.
Metric from Human: Zero-shot Monocular Metric Depth Estimation via Test-time Adaptation
·4145 words·20 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 Carnegie Mellon University
Humans as landmarks: A novel zero-shot monocular metric depth estimation method leverages generative models and human mesh recovery to transfer metric scale information, achieving superior generalizat…
Markov Equivalence and Consistency in Differentiable Structure Learning
·2350 words·12 mins·
loading
·
loading
AI Theory
Causality
🏢 Carnegie Mellon University
Researchers developed a new, differentiable score function for learning causal relationships from data that reliably recovers the simplest causal model, even with complex data.
MAmmoTH2: Scaling Instructions from the Web
·2418 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
MAmmoTH2: Harvesting 10M web instructions for enhanced LLM reasoning!
LLM Dataset Inference: Did you train on my dataset?
·4983 words·24 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
LLM dataset inference reliably detects if a dataset was used in training, overcoming limitations of existing membership inference attacks.
Lips Are Lying: Spotting the Temporal Inconsistency between Audio and Visual in Lip-Syncing DeepFakes
·2315 words·11 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Carnegie Mellon University
LipFD: a novel method leverages audio-visual inconsistencies to accurately spot lip-syncing deepfakes, outperforming existing methods and introducing a high-quality dataset for future research.