Skip to main content

🏢 MIT

Edit Distance Robust Watermarks via Indexing Pseudorandom Codes
·245 words·2 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 MIT
This paper presents a novel watermarking scheme for language models that is both undetectable and robust to a constant fraction of adversarial edits (insertions, deletions, substitutions).
Dynamic Service Fee Pricing under Strategic Behavior: Actions as Instruments and Phase Transition
·2204 words·11 mins· loading · loading
AI Generated AI Theory Optimization 🏢 MIT
This research introduces novel algorithms to dynamically price third-party platform service fees under strategic buyer behavior, achieving optimal revenue with a theoretically proven regret bound.
Designs for Enabling Collaboration in Human-Machine Teaming via Interactive and Explainable Systems
·1914 words·9 mins· loading · loading
AI Applications Robotics 🏢 MIT
Boosting Human-AI teamwork via interactive, explainable AI!
Data Acquisition via Experimental Design for Data Markets
·2343 words·11 mins· loading · loading
Machine Learning Federated Learning 🏢 MIT
Federated data acquisition via experimental design (DAVED) achieves lower prediction error without labeled validation data, optimizing cost-effectively for test-set predictions in decentralized market…
ContextCite: Attributing Model Generation to Context
·7666 words·36 mins· loading · loading
AI Generated Natural Language Processing Text Generation 🏢 MIT
CONTEXTCITE pinpoints which parts of a given context led a language model to generate a specific statement, improving model verification and response quality.
Compact Proofs of Model Performance via Mechanistic Interpretability
·4006 words·19 mins· loading · loading
AI Theory Interpretability 🏢 MIT
Researchers developed a novel method using mechanistic interpretability to create compact formal proofs for AI model performance, improving AI safety and reliability.
Clustering in Causal Attention Masking
·1455 words·7 mins· loading · loading
AI Theory Causality 🏢 MIT
Researchers strengthen understanding of transformer self-attention by proving asymptotic convergence to single clusters under causal masking, linking it to the Rényi parking problem.
BrainBits: How Much of the Brain are Generative Reconstruction Methods Using?
·2562 words·13 mins· loading · loading
Computer Vision Image Generation 🏢 MIT
BrainBits reveals that surprisingly little brain information is needed for high-fidelity image & text reconstruction, highlighting the dominance of generative model priors over neural signal extractio…
BitDelta: Your Fine-Tune May Only Be Worth One Bit
·2156 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 MIT
BitDelta drastically shrinks fine-tuned LLMs by quantizing their weight deltas to just one bit, achieving 10x memory reduction and latency improvements without sacrificing performance.
BendVLM: Test-Time Debiasing of Vision-Language Embeddings
·2604 words·13 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 MIT
BEND-VLM: A novel, efficient test-time debiasing method for vision-language models, resolving bias without retraining.
Are Uncertainty Quantification Capabilities of Evidential Deep Learning a Mirage?
·2360 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 MIT
Evidential deep learning’s uncertainty quantification is unreliable; this paper reveals its limitations, proposes model uncertainty incorporation for improved performance.
Architect: Generating Vivid and Interactive 3D Scenes with Hierarchical 2D Inpainting
·2150 words·11 mins· loading · loading
Multimodal Learning Embodied AI 🏢 MIT
ARCHITECT: Generating realistic 3D scenes using hierarchical 2D inpainting!
ALPS: Improved Optimization for Highly Sparse One-Shot Pruning for Large Language Models
·3567 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 MIT
ALPS: An optimization-based framework achieves state-of-the-art one-shot LLM pruning, significantly reducing test perplexity and improving zero-shot performance.
Algorithmic Capabilities of Random Transformers
·3079 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 MIT
Randomly initialized transformers, with only embedding layers optimized, surprisingly excel at various algorithmic tasks, revealing inherent capabilities even before training.
Achieving Constant Regret in Linear Markov Decision Processes
·1852 words·9 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 MIT
Cert-LSVI-UCB achieves constant regret in RL with linear function approximation, even under model misspecification, using a novel certified estimator.
A Decision-Language Model (DLM) for Dynamic Restless Multi-Armed Bandit Tasks in Public Health
·2747 words·13 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 MIT
LLMs dynamically adjust restless multi-armed bandit (RMAB) resource allocation policies in public health via human-language commands.