Posters
2024
DenseFormer: Enhancing Information Flow in Transformers via Depth Weighted Averaging
·3222 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 EPFL
DenseFormer enhances transformers by adding a depth-weighted averaging step, improving data efficiency and outperforming baselines in memory and inference time without increasing model size.
Dense Connector for MLLMs
·3198 words·16 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Tsinghua University
Boosting multimodal LLMs, the Dense Connector efficiently integrates multi-layer visual features for significantly enhanced performance.
Dense Associative Memory Through the Lens of Random Features
·1742 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 IBM Research
Boost associative memory capacity without extra parameters! DrDAM uses random features to approximate Dense Associative Memories, enabling efficient memory addition and retrieval.
Denoising Diffusion Path: Attribution Noise Reduction with An Auxiliary Diffusion Model
·2911 words·14 mins·
loading
·
loading
AI Generated
AI Theory
Interpretability
🏢 School of Computer Science, Fudan University
Denoising Diffusion Path (DDPath) uses diffusion models to dramatically reduce noise in attribution methods for deep neural networks, leading to clearer explanations and improved quantitative results.
DeNetDM: Debiasing by Network Depth Modulation
·2848 words·14 mins·
loading
·
loading
AI Generated
AI Theory
Fairness
🏢 University of Surrey
DeNetDM uses network depth modulation to automatically debiase image classifiers without bias annotations or data augmentation, improving accuracy by 5%.
Dendritic Integration Inspired Artificial Neural Networks Capture Data Correlation
·1801 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 School of Mathematical Sciences, Shanghai Jiao Tong University
Biologically-inspired Dit-CNNs leverage quadratic neuron integration to capture data correlation, achieving state-of-the-art performance on image classification benchmarks.
Demystify Mamba in Vision: A Linear Attention Perspective
·2184 words·11 mins·
loading
·
loading
Computer Vision
Image Classification
🏢 Tsinghua University
Vision’s Mamba model demystified: Researchers unveil its surprising link to linear attention, improving efficiency and accuracy through design enhancements.
DeMo: Decoupling Motion Forecasting into Directional Intentions and Dynamic States
·2266 words·11 mins·
loading
·
loading
AI Applications
Autonomous Vehicles
🏢 School of Data Science, Fudan University
DeMo: Decoupling motion forecasting into directional intentions and dynamic states for improved autonomous driving.
Delving into the Reversal Curse: How Far Can Large Language Models Generalize?
·3631 words·18 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Zhejiang University
Large language models struggle to generalize knowledge when facing seemingly simple reversals, a phenomenon termed the ‘reversal curse.’ This study reveals that this limitation is strongly linked to t…
DeltaDock: A Unified Framework for Accurate, Efficient, and Physically Reliable Molecular Docking
·4235 words·20 mins·
loading
·
loading
AI Generated
AI Applications
Healthcare
🏢 Peking University
DeltaDock: a novel two-stage framework revolutionizes molecular docking by improving accuracy and reliability, achieving a 300% increase in success rate compared to the prior state-of-the-art in blind…
DeltaDEQ: Exploiting Heterogeneous Convergence for Accelerating Deep Equilibrium Iterations
·2888 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
Video Understanding
🏢 ETH Zurich
DeltaDEQ accelerates deep equilibrium model inference by 73-84% via a novel ‘heterogeneous convergence’ exploitation technique, maintaining accuracy.
Delta-CoMe: Training-Free Delta-Compression with Mixed-Precision for Large Language Models
·2535 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Peking University
Delta-CoMe: Training-free mixed-precision delta compression boosts LLM deployment efficiency.
DEL: Discrete Element Learner for Learning 3D Particle Dynamics with Neural Rendering
·3655 words·18 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Hong Kong University of Science and Technology
DEL: Learns 3D particle dynamics from 2D images via physics-informed neural rendering, exceeding existing methods’ accuracy and robustness.
Déjà Vu Memorization in Vision–Language Models
·2200 words·11 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Meta
Vision-language models (VLMs) memorize training data, impacting generalization. This paper introduces ‘déjà vu memorization,’ a novel method measuring this, revealing significant memorization even in…
DeiSAM: Segment Anything with Deictic Prompting
·3865 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Vision-Language Models
🏢 Technical University of Darmstadt
DeiSAM uses large language models and differentiable logic to achieve highly accurate image segmentation using complex, context-dependent descriptions.
DEFT: Efficient Fine-tuning of Diffusion Models by Learning the Generalised $h$-transform
·3837 words·19 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University College London
DEFT: A novel method efficiently fine-tunes diffusion models for conditional generation via a generalized h-transform, achieving state-of-the-art performance with significant speed improvements.
DeformableTST: Transformer for Time Series Forecasting without Over-reliance on Patching
·3798 words·18 mins·
loading
·
loading
AI Applications
Forecasting
🏢 Tsinghua University
DeformableTST: a new Transformer model for time series forecasting that surpasses existing methods by reducing over-reliance on patching, enhancing performance and adaptability.
Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models
·3247 words·16 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Michigan State University
AdvUnlearn enhances diffusion model robustness against adversarial attacks during concept erasure by integrating adversarial training, improving the trade-off between robustness and model utility.
DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effective for LMMs
·2495 words·12 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Microsoft Research
DeepStack: Stacking visual tokens boosts LMMs efficiency and performance!
DeepLag: Discovering Deep Lagrangian Dynamics for Intuitive Fluid Prediction
·3777 words·18 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Tsinghua University
DeepLag improves fluid prediction by uniquely combining Lagrangian and Eulerian perspectives, tracking key particles to reveal hidden dynamics and improve prediction accuracy.