Posters
2024
Data Mixture Inference Attack: BPE Tokenizers Reveal Training Data Compositions
·3904 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Washington
Researchers uncover hidden training data secrets of large language models by analyzing their byte-pair encoding tokenizers, revealing the proportions of different languages and domains.
Data Free Backdoor Attacks
·2377 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 the Pennsylvania State University
Data-Free Backdoor Attacks (DFBA) injects undetectable backdoors into pre-trained classifiers without retraining or architectural changes, bypassing existing defenses.
Data Distribution Valuation
·3717 words·18 mins·
loading
·
loading
AI Theory
Valuation
🏢 Carnegie Mellon University
This paper proposes a novel MMD-based method for data distribution valuation, enabling theoretically-principled comparison of data distributions from limited samples, outperforming existing methods in…
Data Augmentation with Diffusion for Open-Set Semi-Supervised Learning
·3101 words·15 mins·
loading
·
loading
AI Generated
Machine Learning
Semi-Supervised Learning
🏢 Kim Jaechul Graduate School of AI, KAIST
Boosting semi-supervised learning, a new data augmentation method using diffusion models significantly improves model accuracy, especially with mismatched data distributions.
Data Attribution for Text-to-Image Models by Unlearning Synthesized Images
·4461 words·21 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 Carnegie Mellon University
Unlearning synthesized images efficiently reveals influential training data for text-to-image models, improving data attribution accuracy and facilitating better model understanding.
Data Acquisition via Experimental Design for Data Markets
·2343 words·11 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 MIT
Federated data acquisition via experimental design (DAVED) achieves lower prediction error without labeled validation data, optimizing cost-effectively for test-set predictions in decentralized market…
DAT: Improving Adversarial Robustness via Generative Amplitude Mix-up in Frequency Domain
·3721 words·18 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 State Key Laboratory of Internet of Things for Smart City, University of Macau
Boost AI model robustness against adversarial attacks by creatively mixing training sample’s frequency amplitude with distractor images, focusing model learning on phase patterns, thus enhancing accur…
DASH: Warm-Starting Neural Network Training in Stationary Settings without Loss of Plasticity
·4111 words·20 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Graduate School of AI, KAIST
DASH combats neural network training’s plasticity loss during warm-starting by selectively forgetting memorized noise while preserving features, improving accuracy and efficiency.
DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving
·1939 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Tsinghua University
DART-Math tackles LLM limitations in mathematical problem-solving by introducing Difficulty-Aware Rejection Tuning, a novel method that generates high-quality, bias-reduced datasets, resulting in supe…
DARNet: Dual Attention Refinement Network with Spatiotemporal Construction for Auditory Attention Detection
·1673 words·8 mins·
loading
·
loading
Multimodal Learning
Audio-Visual Learning
🏢 Tsinghua University
DARNet: a dual attention network for auditory attention detection surpasses current state-of-the-art models, especially in short decision windows, achieving this with a 91% reduction in parameters.
DarkSAM: Fooling Segment Anything Model to Segment Nothing
·3441 words·17 mins·
loading
·
loading
AI Generated
Computer Vision
Image Segmentation
🏢 Huazhong University of Science and Technology
DarkSAM, a novel prompt-free attack, renders the Segment Anything Model incapable of segmenting objects across diverse images, highlighting its vulnerability to universal adversarial perturbations.
DARG: Dynamic Evaluation of Large Language Models via Adaptive Reasoning Graph
·3306 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Dartmouth College
DARG dynamically evaluates LLMs via adaptive reasoning graphs, revealing performance drops with increased complexity and exposing model biases.
DAPE: Data-Adaptive Positional Encoding for Length Extrapolation
·3365 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 CUHK
DAPE: A novel data-adaptive positional encoding method dynamically adjusts positional information based on input context, improving transformer performance and length generalization.
DAGER: Exact Gradient Inversion for Large Language Models
·2286 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 INSAIT
DAGER: Exact gradient inversion for LLMs; recovers full input text batches precisely.
DA-Ada: Learning Domain-Aware Adapter for Domain Adaptive Object Detection
·2460 words·12 mins·
loading
·
loading
Computer Vision
Object Detection
🏢 Intelligent Software Research Center, Institute of Software, CAS, Beijing, China
DA-Ada enhances domain adaptive object detection by using a novel domain-aware adapter that leverages both domain-invariant and domain-specific knowledge for improved accuracy and generalization acros…
D2R2: Diffusion-based Representation with Random Distance Matching for Tabular Few-shot Learning
·1776 words·9 mins·
loading
·
loading
Machine Learning
Few-Shot Learning
🏢 Hong Kong University of Science and Technology
D2R2: A novel diffusion-based model for tabular few-shot learning, achieves state-of-the-art results by leveraging semantic knowledge and distance matching.
D-LLM: A Token Adaptive Computing Resource Allocation Strategy for Large Language Models
·2704 words·13 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Huawei Technologies Co., Ltd.
D-LLM dynamically allocates computing resources during LLM token processing, reducing computational costs and memory usage by up to 50% without sacrificing accuracy.
D-CPT Law: Domain-specific Continual Pre-Training Scaling Law for Large Language Models
·3930 words·19 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Taobao & Tmall Group of Alibaba
New D-CPT Law optimizes continual pre-training for LLMs by predicting optimal data mixture ratios, drastically cutting training costs.
CYCLO: Cyclic Graph Transformer Approach to Multi-Object Relationship Modeling in Aerial Videos
·2627 words·13 mins·
loading
·
loading
Computer Vision
Scene Understanding
🏢 University of Arkansas
CYCLO: A novel cyclic graph transformer excels at multi-object relationship modeling in aerial videos.
CV-VAE: A Compatible Video VAE for Latent Generative Video Models
·3396 words·16 mins·
loading
·
loading
AI Generated
Computer Vision
Video Understanding
🏢 Tencent AI Lab
CV-VAE: A compatible video VAE enabling efficient, high-quality latent video generation by bridging the gap between image and video latent spaces.