Image Generation
ECMamba: Consolidating Selective State Space Model with Retinex Guidance for Efficient Multiple Exposure Correction
·1754 words·9 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 McMaster University
ECMamba: A novel dual-branch framework efficiently corrects multiple exposure images by integrating Retinex theory and an innovative 2D selective state-space layer, achieving state-of-the-art performa…
DreamSteerer: Enhancing Source Image Conditioned Editability using Personalized Diffusion Models
·5101 words·24 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 Australian National University
DreamSteerer enhances source image-conditioned editability in personalized diffusion models via a novel Editability Driven Score Distillation objective and mode shifting regularization, achieving sign…
DRACO: A Denoising-Reconstruction Autoencoder for Cryo-EM
·1954 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 School of Information Science and Technology, ShanghaiTech University
DRACO, a denoising-reconstruction autoencoder, revolutionizes cryo-EM by leveraging a large-scale dataset and hybrid training for superior image denoising and downstream task performance.
Doubly Hierarchical Geometric Representations for Strand-based Human Hairstyle Generation
·2527 words·12 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Carnegie Mellon University
Doubly hierarchical geometric representations enable realistic human hairstyle generation by separating low and high-frequency details in hair strands, resulting in high-quality, detailed virtual hair…
DomainGallery: Few-shot Domain-driven Image Generation by Attribute-centric Finetuning
·1917 words·9 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Shanghai Jiao Tong University
DomainGallery: Few-shot domain-driven image generation via attribute-centric finetuning, solving key issues of previous works by introducing attribute erasure, disentanglement, regularization, and enh…
DiTFastAttn: Attention Compression for Diffusion Transformer Models
·2788 words·14 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Tsinghua University
DiTFastAttn: A post-training compression method drastically speeds up diffusion transformer models by cleverly reducing redundancy in attention calculations, leading to up to a 1.8x speedup at high re…
Disentangled Style Domain for Implicit $z$-Watermark Towards Copyright Protection
·1999 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Fudan University
This paper introduces a novel implicit Zero-Watermarking scheme using disentangled style domains to detect unauthorized dataset usage in text-to-image models, offering robust copyright protection via …
Direct Unlearning Optimization for Robust and Safe Text-to-Image Models
·4016 words·19 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 NAVER AI Lab
Direct Unlearning Optimization (DUO) robustly removes unsafe content from text-to-image models by using paired image data and output-preserving regularization, effectively defending against adversaria…
Direct Consistency Optimization for Robust Customization of Text-to-Image Diffusion models
·3011 words·15 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 KAIST
Boosting personalized image generation! Direct Consistency Optimization (DCO) fine-tunes text-to-image models, ensuring subject consistency and prompt fidelity, even when merging separately customized…
DiP-GO: A Diffusion Pruner via Few-step Gradient Optimization
·2302 words·11 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Advanced Micro Devices, Inc.
DiP-GO: A novel pruning method accelerates diffusion models via few-step gradient optimization, achieving a 4.4x speedup on Stable Diffusion 1.5 without accuracy loss.
DiMSUM: Diffusion Mamba - A Scalable and Unified Spatial-Frequency Method for Image Generation
·3655 words·18 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 VinAI Research
DiMSUM: A novel diffusion model boosts image generation by unifying spatial and frequency information, achieving superior results and faster training.
Diffusion4D: Fast Spatial-temporal Consistent 4D generation via Video Diffusion Models
·1559 words·8 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Toronto
Diffusion4D: Fast, consistent 4D content generation via a novel 4D-aware video diffusion model, surpassing existing methods in efficiency and 4D geometry consistency.
Diffusion Priors for Variational Likelihood Estimation and Image Denoising
·2184 words·11 mins·
loading
·
loading
Image Generation
🏢 Huazhong University of Science and Technology
Adaptive likelihood estimation and MAP inference during reverse diffusion tackles real-world image noise.
Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models
·3247 words·16 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Michigan State University
AdvUnlearn enhances diffusion model robustness against adversarial attacks during concept erasure by integrating adversarial training, improving the trade-off between robustness and model utility.
Dealing with Synthetic Data Contamination in Online Continual Learning
·2977 words·14 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Tokyo
AI-generated images contaminate online continual learning datasets, hindering performance. A new method, ESRM, leverages entropy and real/synthetic similarity maximization to select high-quality data…
DDR: Exploiting Deep Degradation Response as Flexible Image Descriptor
·2459 words·12 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 School of Computer Science and Technology, Tongji University, China
Deep Degradation Response (DDR) uses image deep feature changes under degradation to create a flexible image descriptor, excelling in blind image quality assessment and unsupervised image restoration.
Data Attribution for Text-to-Image Models by Unlearning Synthesized Images
·4461 words·21 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 Carnegie Mellon University
Unlearning synthesized images efficiently reveals influential training data for text-to-image models, improving data attribution accuracy and facilitating better model understanding.
Ctrl-X: Controlling Structure and Appearance for Text-To-Image Generation Without Guidance
·2880 words·14 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 UC Los Angeles
Ctrl-X: Zero-shot text-to-image generation with training-free structure & appearance control!
CryoGEM: Physics-Informed Generative Cryo-Electron Microscopy
·2131 words·11 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 ShanghaiTech University
CryoGEM: Physics-informed generative model creates realistic synthetic cryo-EM datasets, boosting particle picking and pose estimation accuracy for higher-resolution protein structure determination.
Cross-Scale Self-Supervised Blind Image Deblurring via Implicit Neural Representation
·3186 words·15 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 National University of Singapore
Self-supervised blind image deblurring (BID) breakthrough! A novel cross-scale consistency loss and progressive training scheme using implicit neural representations achieves superior performance wit…