Posters
2024
Accelerating Greedy Coordinate Gradient and General Prompt Optimization via Probe Sampling
·2272 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 National University of Singapore
Probe sampling accelerates Greedy Coordinate Gradient (GCG) and other prompt optimization methods by up to 5.6x, achieving equal or better attack success rates, making LLM safety research faster and m…
Accelerating ERM for data-driven algorithm design using output-sensitive techniques
·366 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 Carnegie Mellon University
Accelerating ERM for data-driven algorithm design using output-sensitive techniques achieves computationally efficient learning by scaling with the actual number of pieces in the dual loss function, n…
Accelerating Blockwise Parallel Language Models with Draft Refinement
·2883 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 KAIST AI
Boost LLM inference speed by 3x! This paper refines blockwise parallel decoding (BPD) by cleverly refining draft predictions, resulting in faster text generation for large language models.
Accelerating Augmentation Invariance Pretraining
·1854 words·9 mins·
loading
·
loading
Computer Vision
Self-Supervised Learning
🏢 University of Wisconsin-Madison
Boost Vision Transformer pretraining speed by 4x with novel sequence compression techniques!
Accelerated Regularized Learning in Finite N-Person Games
·1352 words·7 mins·
loading
·
loading
AI Theory
Optimization
🏢 Stanford University
Accelerated learning in games achieved! FTXL algorithm exponentially speeds up convergence to Nash equilibria in finite N-person games, even under limited feedback.
Abstracted Shapes as Tokens - A Generalizable and Interpretable Model for Time-series Classification
·3172 words·15 mins·
loading
·
loading
AI Generated
Machine Learning
Self-Supervised Learning
🏢 Rensselaer Polytechnic Institute
VQShape: a pre-trained model uses abstracted shapes as interpretable tokens for generalizable time-series classification, achieving comparable performance to black-box models and excelling in zero-sho…
Abstract Reward Processes: Leveraging State Abstraction for Consistent Off-Policy Evaluation
·1668 words·8 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Massachusetts
STAR framework leverages state abstraction for consistent, low-variance off-policy evaluation in reinforcement learning, outperforming existing methods.
Absorb & Escape: Overcoming Single Model Limitations in Generating Heterogeneous Genomic Sequences
·3759 words·18 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Imperial College London
Absorb & Escape: a novel post-training sampling method that overcomes single model limitations by combining Autoregressive (AR) and Diffusion Models (DMs), generating high-quality heterogeneous genomi…
Abrupt Learning in Transformers: A Case Study on Matrix Completion
·5285 words·25 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Michigan
Transformers exhibit abrupt learning: training loss plateaus, then suddenly drops. This study uses matrix completion to demonstrate this phenomenon, providing insights into the model’s algorithmic sh…
Abductive Reasoning in Logical Credal Networks
·1676 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 IBM Research
This paper presents efficient algorithms for abductive reasoning in Logical Credal Networks (LCNs), addressing the MAP and Marginal MAP inference tasks to enable scalable solutions for complex real-wo…
A2PO: Towards Effective Offline Reinforcement Learning from an Advantage-aware Perspective
·2287 words·11 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Zhejiang University
A2PO: A novel offline RL method tackles constraint conflicts in mixed-quality datasets by disentangling behavior policies with a conditional VAE and optimizing advantage-aware constraints, achieving s…
A Walsh Hadamard Derived Linear Vector Symbolic Architecture
·1922 words·10 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 University of Maryland, Baltimore County
Hadamard-derived Linear Binding (HLB): A novel, efficient vector symbolic architecture surpassing existing methods in classical AI tasks and deep learning applications.
A versatile informative diffusion model for single-cell ATAC-seq data generation and analysis
·1769 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 City University of Hong Kong
ATAC-Diff: A versatile diffusion model for high-quality single-cell ATAC-seq data generation and analysis, surpassing state-of-the-art.
A Versatile Diffusion Transformer with Mixture of Noise Levels for Audiovisual Generation
·2816 words·14 mins·
loading
·
loading
Multimodal Learning
Audio-Visual Learning
🏢 Seoul National University
A single model tackles diverse audiovisual generation tasks using a novel Mixture of Noise Levels approach, resulting in temporally consistent and high-quality outputs.
A Universal Growth Rate for Learning with Smooth Surrogate Losses
·1364 words·7 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Courant Institute
This paper reveals a universal square-root growth rate for H-consistency bounds of smooth surrogate losses in classification, significantly advancing our understanding of loss function selection.
A Unifying Post-Processing Framework for Multi-Objective Learn-to-Defer Problems
·1692 words·8 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
🏢 Max Planck Institute for Intelligent Systems
A novel post-processing framework, based on a d-dimensional generalization of the Neyman-Pearson lemma, optimally solves multi-objective learn-to-defer problems under various constraints, improving co…
A Unifying Normative Framework of Decision Confidence
·1353 words·7 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Washington
New normative framework for decision confidence models diverse tasks by incorporating rewards, priors, and uncertainty, outperforming existing methods.
A Unified Principle of Pessimism for Offline Reinforcement Learning under Model Mismatch
·1838 words·9 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Department of Electrical and Computer Engineering University of Central Florida
Unified pessimism principle in offline RL conquers data sparsity & model mismatch, achieving near-optimal performance across various divergence models.
A Unified Framework for 3D Scene Understanding
·2347 words·12 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Huazhong University of Science and Technology
UniSeg3D: One model to rule them all! This unified framework masters six 3D segmentation tasks (panoptic, semantic, instance, interactive, referring, and open-vocabulary) simultaneously, outperforming…
A Unified Confidence Sequence for Generalized Linear Models, with Applications to Bandits
·1965 words·10 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 KAIST
A unified confidence sequence (CS) construction for generalized linear models (GLMs) achieves state-of-the-art regret bounds for contextual bandits, notably a poly(S)-free regret for logistic bandits.