Skip to main content

Posters

2024

Treeffuser: probabilistic prediction via conditional diffusions with gradient-boosted trees
·2082 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 Department of Computer Science, Columbia University
Treeffuser: Accurate probabilistic predictions from tabular data using conditional diffusion models and gradient-boosted trees!
Tree of Attacks: Jailbreaking Black-Box LLMs Automatically
·1948 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Yale University
TAP: automated jailbreaking of black-box LLMs with high success rates, using fewer queries than previous methods.
Treatment of Statistical Estimation Problems in Randomized Smoothing for Adversarial Robustness
·1858 words·9 mins· loading · loading
AI Theory Robustness 🏢 Tübingen AI Center, University of Tübingen
This paper optimizes randomized smoothing, a crucial certified defense against adversarial attacks, by introducing novel statistical methods that drastically reduce the computational cost, leading to …
Trap-MID: Trapdoor-based Defense against Model Inversion Attacks
·3599 words·17 mins· loading · loading
AI Generated AI Theory Privacy 🏢 National Taiwan University
Trap-MID: Outsmarting model inversion attacks with cleverly placed ’trapdoors'!
TransVIP: Speech to Speech Translation System with Voice and Isochrony Preservation
·1866 words·9 mins· loading · loading
Natural Language Processing Machine Translation 🏢 Microsoft
TransVIP: groundbreaking speech-to-speech translation system preserving voice & isochrony, outperforming current state-of-the-art models!
Transition Constrained Bayesian Optimization via Markov Decision Processes
·2420 words·12 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Imperial College London
This paper presents a novel BayesOpt framework that incorporates Markov Decision Processes to optimize black-box functions with transition constraints, overcoming limitations of traditional methods.
Transforming Vision Transformer: Towards Efficient Multi-Task Asynchronous Learner
·2048 words·10 mins· loading · loading
Computer Vision Scene Understanding 🏢 String
Efficient Multi-Task Learning (EMTAL) transforms pre-trained Vision Transformers into efficient multi-task learners by using a MoEfied LoRA structure, a Quality Retaining optimization, and a router fa…
Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Models
·2720 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
MOHAWK: Distilling Transformers’ quadratic knowledge into faster subquadratic SSMs, achieving state-of-the-art performance with <1% of training data!
Transformers Represent Belief State Geometry in their Residual Stream
·1739 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Simplex
Transformers encode information beyond next-token prediction by linearly representing belief state geometry in their residual stream, even with complex fractal structures.
Transformers need glasses! Information over-squashing in language tasks
·3003 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Oxford
Large language models (LLMs) suffer from information loss due to representational collapse and over-squashing, causing failures in simple tasks; this paper provides theoretical analysis and practical …
Transformers Learn to Achieve Second-Order Convergence Rates for In-Context Linear Regression
·3338 words·16 mins· loading · loading
Machine Learning Few-Shot Learning 🏢 University of Southern California
Transformers surprisingly learn second-order optimization methods for in-context linear regression, achieving exponentially faster convergence than gradient descent!
Transformers Can Do Arithmetic with the Right Embeddings
·3154 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Maryland
Researchers enhanced transformer performance on arithmetic tasks by introducing Abacus Embeddings, which encode each digit’s position, enabling improved generalization and unlocking multi-step reasoni…
Transformers as Game Players: Provable In-context Game-playing Capabilities of Pre-trained Models
·502 words·3 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 University of Virginia
Pre-trained transformers can provably learn to play games near-optimally using in-context learning, offering theoretical guarantees for both decentralized and centralized settings.
Transformers are Minimax Optimal Nonparametric In-Context Learners
·1461 words·7 mins· loading · loading
AI Generated Machine Learning Meta Learning 🏢 University of Tokyo
Transformers excel at in-context learning by leveraging minimax-optimal nonparametric learning, achieving near-optimal risk with sufficient pretraining data diversity.
Transformer Doctor: Diagnosing and Treating Vision Transformers
·3080 words·15 mins· loading · loading
AI Generated Computer Vision Image Classification 🏢 College of Computer Science and Technology, Zhejiang University
Transformer Doctor diagnoses and treats vision transformer errors by identifying and correcting information integration issues, improving model performance and interpretability.
Transformation-Invariant Learning and Theoretical Guarantees for OOD Generalization
·541 words·3 mins· loading · loading
AI Theory Generalization 🏢 Yale University
This paper introduces a novel theoretical framework for robust machine learning under distribution shifts, offering learning rules and guarantees, highlighting the game-theoretic viewpoint of distribu…
Transferring disentangled representations: bridging the gap between synthetic and real images
·3866 words·19 mins· loading · loading
Machine Learning Representation Learning 🏢 Università Degli Studi Di Genova
This paper bridges the gap between synthetic and real image disentanglement by proposing a novel transfer learning approach. The method leverages weakly supervised learning on synthetic data to train…
Transferable Boltzmann Generators
·4942 words·24 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Freie Universität Berlin
Transferable Boltzmann Generators enable efficient, zero-shot sampling of unseen molecular systems’ equilibrium distributions, boosting molecular simulations.
Transferable Adversarial Attacks on SAM and Its Downstream Models
·2130 words·10 mins· loading · loading
Computer Vision Image Segmentation 🏢 Nanyang Technological University
UMI-GRAT: A universal meta-initialized and gradient robust adversarial attack effectively exploits vulnerabilities in the Segment Anything Model (SAM) and its fine-tuned downstream models, even withou…
Transferability Bound Theory: Exploring Relationship between Adversarial Transferability and Flatness
·1926 words·10 mins· loading · loading
AI Generated AI Theory Robustness 🏢 East China Normal University
Challenging common assumptions, researchers prove that flatter adversarial examples don’t guarantee better transferability and introduce TPA, a theoretically-grounded attack creating more transferable…