3D Vision
Effective Rank Analysis and Regularization for Enhanced 3D Gaussian Splatting
·2803 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ KAIST
Effective rank regularization enhances 3D Gaussian splatting, resolving needle-like artifacts and improving 3D model quality.
Dynamic 3D Gaussian Fields for Urban Areas
·2544 words·12 mins·
loading
·
loading
3D Vision
π’ ETH Zurich
4DGF, a novel neural scene representation, achieves interactive-speed novel view synthesis for large-scale dynamic urban areas by efficiently combining 3D Gaussians and neural fields.
Dual-frame Fluid Motion Estimation with Test-time Optimization and Zero-divergence Loss
·2477 words·12 mins·
loading
·
loading
Computer Vision
3D Vision
π’ University of Chinese Academy of Sciences
Self-supervised dual-frame fluid motion estimation achieves superior accuracy with 99% less training data, using a novel zero-divergence loss and dynamic velocimetry enhancement.
Dual-Diffusion for Binocular 3D Human Pose Estimation
·3829 words·18 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ Shanghai Jiao Tong University
Dual-Diffusion boosts binocular 3D human pose estimation accuracy by simultaneously denoising 2D and 3D pose uncertainties using a diffusion model.
Dual Encoder GAN Inversion for High-Fidelity 3D Head Reconstruction from Single Images
·3653 words·18 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Bilkent University
Dual encoder GAN inversion achieves high-fidelity 3D head reconstruction from single images by cleverly combining outputs from encoders specialized for visible and invisible regions, surpassing existi…
DreamMesh4D: Video-to-4D Generation with Sparse-Controlled Gaussian-Mesh Hybrid Representation
·2631 words·13 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ Zhejiang University
DreamMesh4D: Generating high-fidelity dynamic 3D meshes from monocular video using a novel Gaussian-mesh hybrid representation and adaptive hybrid skinning.
DOGS: Distributed-Oriented Gaussian Splatting for Large-Scale 3D Reconstruction Via Gaussian Consensus
·3216 words·16 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ National University of Singapore
DOGS: Distributed-Oriented Gaussian Splatting accelerates large-scale 3D reconstruction by distributing the training of 3D Gaussian Splatting models across multiple machines, achieving 6x faster train…
DN-4DGS: Denoised Deformable Network with Temporal-Spatial Aggregation for Dynamic Scene Rendering
·2765 words·13 mins·
loading
·
loading
Computer Vision
3D Vision
π’ University of Science and Technology of China
DN-4DGS: Real-time dynamic scene rendering is revolutionized by a denoised deformable network with temporal-spatial aggregation, achieving state-of-the-art quality.
DMesh: A Differentiable Mesh Representation
·3349 words·16 mins·
loading
·
loading
Computer Vision
3D Vision
π’ University of Maryland
DMesh: A novel differentiable mesh representation enabling efficient gradient-based optimization for diverse 3D shape applications.
Distribution Guidance Network for Weakly Supervised Point Cloud Semantic Segmentation
·2253 words·11 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Peking University
DGNet enhances weakly supervised point cloud segmentation by aligning feature embeddings to a mixture of von Mises-Fisher distributions, achieving state-of-the-art performance.
DistillNeRF: Perceiving 3D Scenes from Single-Glance Images by Distilling Neural Fields and Foundation Model Features
·2827 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ NVIDIA Research
DistillNeRF: a self-supervised learning framework enabling accurate 3D scene reconstruction from sparse, single-frame images by cleverly distilling features from offline NeRFs and 2D foundation models…
DisC-GS: Discontinuity-aware Gaussian Splatting
·2095 words·10 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Lancaster University
DisC-GS enhances Gaussian Splatting for real-time novel view synthesis by accurately rendering image discontinuities and boundaries, improving visual quality.
Director3D: Real-world Camera Trajectory and 3D Scene Generation from Text
·2781 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ Shanghai Artificial Intelligence Laboratory
Director3D generates realistic 3D scenes and camera trajectories from text descriptions using a three-stage pipeline: Cinematographer, Decorator, and Detailer.
Direct3D: Scalable Image-to-3D Generation via 3D Latent Diffusion Transformer
·2139 words·11 mins·
loading
·
loading
Computer Vision
3D Vision
π’ University of Oxford
Direct3D: Revolutionizing image-to-3D generation with a scalable, native 3D diffusion model achieving state-of-the-art quality.
DiffusionBlend: Learning 3D Image Prior through Position-aware Diffusion Score Blending for 3D Computed Tomography Reconstruction
·2570 words·13 mins·
loading
·
loading
Computer Vision
3D Vision
π’ University of Michigan
DiffusionBlend++ learns a 3D image prior via position-aware diffusion score blending, achieving state-of-the-art 3D CT reconstruction with superior efficiency.
DiffuBox: Refining 3D Object Detection with Point Diffusion
·3129 words·15 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Cornell University
DiffuBox refines 3D object detection using a novel diffusion-based approach, significantly improving accuracy across various domains by refining bounding boxes based on surrounding LiDAR point clouds.
Depth Anywhere: Enhancing 360 Monocular Depth Estimation via Perspective Distillation and Unlabeled Data Augmentation
·2664 words·13 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Stanford University
Depth Anywhere enhances 360-degree monocular depth estimation by cleverly using perspective models to label unlabeled 360-degree data, significantly improving accuracy.
Depth Anything V2
·3310 words·16 mins·
loading
·
loading
Computer Vision
3D Vision
π’ TikTok
Depth Anything V2 drastically improves monocular depth estimation by using synthetic training data, scaling up the teacher model, and employing pseudo-labeled real images. It outperforms previous met…
DEL: Discrete Element Learner for Learning 3D Particle Dynamics with Neural Rendering
·3655 words·18 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Hong Kong University of Science and Technology
DEL: Learns 3D particle dynamics from 2D images via physics-informed neural rendering, exceeding existing methods’ accuracy and robustness.
DeBaRA: Denoising-Based 3D Room Arrangement Generation
·2721 words·13 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Dassault SystΓ¨mes
DeBaRA: a novel denoising-based model generates realistic & controllable 3D room layouts, surpassing existing methods.