Posters
2024
Treeffuser: probabilistic prediction via conditional diffusions with gradient-boosted trees
·2082 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Department of Computer Science, Columbia University
Treeffuser: Accurate probabilistic predictions from tabular data using conditional diffusion models and gradient-boosted trees!
Tree of Attacks: Jailbreaking Black-Box LLMs Automatically
·1948 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Yale University
TAP: automated jailbreaking of black-box LLMs with high success rates, using fewer queries than previous methods.
Treatment of Statistical Estimation Problems in Randomized Smoothing for Adversarial Robustness
·1858 words·9 mins·
loading
·
loading
AI Theory
Robustness
🏢 Tübingen AI Center, University of Tübingen
This paper optimizes randomized smoothing, a crucial certified defense against adversarial attacks, by introducing novel statistical methods that drastically reduce the computational cost, leading to …
Trap-MID: Trapdoor-based Defense against Model Inversion Attacks
·3599 words·17 mins·
loading
·
loading
AI Generated
AI Theory
Privacy
🏢 National Taiwan University
Trap-MID: Outsmarting model inversion attacks with cleverly placed ’trapdoors'!
TransVIP: Speech to Speech Translation System with Voice and Isochrony Preservation
·1866 words·9 mins·
loading
·
loading
Natural Language Processing
Machine Translation
🏢 Microsoft
TransVIP: groundbreaking speech-to-speech translation system preserving voice & isochrony, outperforming current state-of-the-art models!
Transition Constrained Bayesian Optimization via Markov Decision Processes
·2420 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Imperial College London
This paper presents a novel BayesOpt framework that incorporates Markov Decision Processes to optimize black-box functions with transition constraints, overcoming limitations of traditional methods.
Transforming Vision Transformer: Towards Efficient Multi-Task Asynchronous Learner
·2048 words·10 mins·
loading
·
loading
Computer Vision
Scene Understanding
🏢 String
Efficient Multi-Task Learning (EMTAL) transforms pre-trained Vision Transformers into efficient multi-task learners by using a MoEfied LoRA structure, a Quality Retaining optimization, and a router fa…
Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Models
·2720 words·13 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
MOHAWK: Distilling Transformers’ quadratic knowledge into faster subquadratic SSMs, achieving state-of-the-art performance with <1% of training data!
Transformers Represent Belief State Geometry in their Residual Stream
·1739 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Simplex
Transformers encode information beyond next-token prediction by linearly representing belief state geometry in their residual stream, even with complex fractal structures.
Transformers need glasses! Information over-squashing in language tasks
·3003 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Oxford
Large language models (LLMs) suffer from information loss due to representational collapse and over-squashing, causing failures in simple tasks; this paper provides theoretical analysis and practical …
Transformers Learn to Achieve Second-Order Convergence Rates for In-Context Linear Regression
·3338 words·16 mins·
loading
·
loading
Machine Learning
Few-Shot Learning
🏢 University of Southern California
Transformers surprisingly learn second-order optimization methods for in-context linear regression, achieving exponentially faster convergence than gradient descent!
Transformers Can Do Arithmetic with the Right Embeddings
·3154 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Maryland
Researchers enhanced transformer performance on arithmetic tasks by introducing Abacus Embeddings, which encode each digit’s position, enabling improved generalization and unlocking multi-step reasoni…
Transformers as Game Players: Provable In-context Game-playing Capabilities of Pre-trained Models
·502 words·3 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Virginia
Pre-trained transformers can provably learn to play games near-optimally using in-context learning, offering theoretical guarantees for both decentralized and centralized settings.
Transformers are Minimax Optimal Nonparametric In-Context Learners
·1461 words·7 mins·
loading
·
loading
AI Generated
Machine Learning
Meta Learning
🏢 University of Tokyo
Transformers excel at in-context learning by leveraging minimax-optimal nonparametric learning, achieving near-optimal risk with sufficient pretraining data diversity.
Transformer Doctor: Diagnosing and Treating Vision Transformers
·3080 words·15 mins·
loading
·
loading
AI Generated
Computer Vision
Image Classification
🏢 College of Computer Science and Technology, Zhejiang University
Transformer Doctor diagnoses and treats vision transformer errors by identifying and correcting information integration issues, improving model performance and interpretability.
Transformation-Invariant Learning and Theoretical Guarantees for OOD Generalization
·541 words·3 mins·
loading
·
loading
AI Theory
Generalization
🏢 Yale University
This paper introduces a novel theoretical framework for robust machine learning under distribution shifts, offering learning rules and guarantees, highlighting the game-theoretic viewpoint of distribu…
Transferring disentangled representations: bridging the gap between synthetic and real images
·3866 words·19 mins·
loading
·
loading
Machine Learning
Representation Learning
🏢 Università Degli Studi Di Genova
This paper bridges the gap between synthetic and real image disentanglement by proposing a novel transfer learning approach. The method leverages weakly supervised learning on synthetic data to train…
Transferable Boltzmann Generators
·4942 words·24 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Freie Universität Berlin
Transferable Boltzmann Generators enable efficient, zero-shot sampling of unseen molecular systems’ equilibrium distributions, boosting molecular simulations.
Transferable Adversarial Attacks on SAM and Its Downstream Models
·2130 words·10 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 Nanyang Technological University
UMI-GRAT: A universal meta-initialized and gradient robust adversarial attack effectively exploits vulnerabilities in the Segment Anything Model (SAM) and its fine-tuned downstream models, even withou…
Transferability Bound Theory: Exploring Relationship between Adversarial Transferability and Flatness
·1926 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 East China Normal University
Challenging common assumptions, researchers prove that flatter adversarial examples don’t guarantee better transferability and introduce TPA, a theoretically-grounded attack creating more transferable…