Skip to main content

3D Vision

MV2Cyl: Reconstructing 3D Extrusion Cylinders from Multi-View Images
·3293 words·16 mins· loading · loading
Computer Vision 3D Vision 🏢 Korea Advanced Institute of Science and Technology
MV2Cyl: A novel method reconstructs 3D extrusion cylinder CAD models directly from multi-view images, surpassing accuracy of methods using raw 3D geometry.
Multistable Shape from Shading Emerges from Patch Diffusion
·2364 words·12 mins· loading · loading
3D Vision 🏢 Harvard University
A novel diffusion model reconstructs multimodal shape distributions from shading, mirroring human multistable perception.
MultiPull: Detailing Signed Distance Functions by Pulling Multi-Level Queries at Multi-Step
·3626 words·18 mins· loading · loading
Computer Vision 3D Vision 🏢 Tsinghua University
MultiPull: a novel method reconstructing detailed 3D surfaces from raw point clouds using multi-step optimization of multi-level features, significantly improving accuracy and detail.
Multi-times Monte Carlo Rendering for Inter-reflection Reconstruction
·1845 words·9 mins· loading · loading
Computer Vision 3D Vision 🏢 Shanghai Jiao Tong University
Ref-MC2 reconstructs high-fidelity 3D objects with inter-reflections by using a novel multi-times Monte Carlo sampling strategy, achieving superior performance in accuracy and efficiency.
Multi-scale Consistency for Robust 3D Registration via Hierarchical Sinkhorn Tree
·2306 words·11 mins· loading · loading
Computer Vision 3D Vision 🏢 Tsinghua University
Hierarchical Sinkhorn Tree (HST) robustly retrieves accurate 3D point cloud correspondences using multi-scale consistency, outperforming state-of-the-art methods.
Multi-hypotheses Conditioned Point Cloud Diffusion for 3D Human Reconstruction from Occluded Images
·2520 words·12 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 KAIST
MHCDIFF: a novel pipeline using multi-hypotheses conditioned point cloud diffusion for accurate 3D human reconstruction from occluded images, outperforming state-of-the-art methods.
MotionGS: Exploring Explicit Motion Guidance for Deformable 3D Gaussian Splatting
·3095 words·15 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 University of Science and Technology of China
MotionGS enhances deformable 3D Gaussian splatting for dynamic scenes by using motion flow to guide deformation, significantly improving reconstruction accuracy and outperforming state-of-the-art meth…
Mixture of neural fields for heterogeneous reconstruction in cryo-EM
·4281 words·21 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Stanford University
Hydra: a novel cryo-EM reconstruction method resolves both conformational and compositional heterogeneity ab initio, enabling the analysis of complex, unpurified samples with state-of-the-art accuracy…
Mining and Transferring Feature-Geometry Coherence for Unsupervised Point Cloud Registration
·2514 words·12 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University, China
INTEGER: a novel unsupervised point cloud registration method leveraging feature-geometry coherence for reliable pseudo-label mining and density-invariant feature learning, achieving state-of-the-art …
MIDGArD: Modular Interpretable Diffusion over Graphs for Articulated Designs
·2491 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Intel Labs
MIDGARD: Generate high-quality, simulatable 3D articulated assets with enhanced control and interpretability using a novel diffusion-based framework.
Metric from Human: Zero-shot Monocular Metric Depth Estimation via Test-time Adaptation
·4145 words·20 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Carnegie Mellon University
Humans as landmarks: A novel zero-shot monocular metric depth estimation method leverages generative models and human mesh recovery to transfer metric scale information, achieving superior generalizat…
Meta 3D AssetGen: Text-to-Mesh Generation with High-Quality Geometry, Texture, and PBR Materials
·2436 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Meta AI
Meta 3D AssetGen: High-quality text-to-mesh generation with realistic PBR materials and lighting, exceeding prior methods in speed and accuracy.
MeshXL: Neural Coordinate Field for Generative 3D Foundation Models
·2662 words·13 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Tencent PCG
MeshXL: Autoregressively generating high-quality 3D meshes using a novel Neural Coordinate Field (NeurCF) representation and large language model approaches.
MeshFormer : High-Quality Mesh Generation with 3D-Guided Reconstruction Model
·1982 words·10 mins· loading · loading
3D Vision 🏢 University of California, San Diego
MeshFormer: High-quality 3D mesh generation from sparse views in seconds, using transformers and 3D convolutions.
Memorize What Matters: Emergent Scene Decomposition from Multitraverse
·3662 words·18 mins· loading · loading
3D Vision 🏢 NVIDIA
3D Gaussian Mapping (3DGM) achieves self-supervised camera-only 3D scene decomposition by leveraging multi-traverse driving data, memorizing permanent structures while filtering out transient objects.
ManiPose: Manifold-Constrained Multi-Hypothesis 3D Human Pose Estimation
·2484 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Valeo.ai
ManiPose: Manifold-constrained multi-hypothesis model solves 3D human pose estimation’s depth ambiguity, outperforming state-of-the-art models in pose consistency.
LuSh-NeRF: Lighting up and Sharpening NeRFs for Low-light Scenes
·2414 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 City University of Hong Kong
LuSh-NeRF: A novel model reconstructs sharp, bright NeRFs from hand-held low-light photos by sequentially modeling and removing noise and blur, outperforming existing methods.
LRM-Zero: Training Large Reconstruction Models with Synthesized Data
·3486 words·17 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Adobe Research
LRM-Zero: Training large reconstruction models solely on synthetic data, achieving quality comparable to real-data trained models.
LP-3DGS: Learning to Prune 3D Gaussian Splatting
·2308 words·11 mins· loading · loading
Computer Vision 3D Vision 🏢 Johns Hopkins University
LP-3DGS learns to optimally prune 3D Gaussian splatting, achieving significant efficiency gains without compromising rendering quality via a trainable binary mask and the Gumbel-Sigmoid method.
LoD-Loc: Aerial Visual Localization using LoD 3D Map with Neural Wireframe Alignment
·3487 words·17 mins· loading · loading
Computer Vision 3D Vision 🏢 National University of Defense Technology
LoD-Loc: A novel aerial visual localization method uses lightweight LoD 3D maps & neural wireframe alignment for accurate and efficient 6-DoF pose estimation, surpassing state-of-the-art methods.