Skip to main content

🏢 Nankai University

Zipper: Addressing Degeneracy in Algorithm-Agnostic Inference
·1827 words·9 mins· loading · loading
AI Theory Interpretability 🏢 Nankai University
Zipper: A novel statistical device resolves the degeneracy issue in algorithm-agnostic inference, enabling reliable goodness-of-fit tests with enhanced power.
Token Merging for Training-Free Semantic Binding in Text-to-Image Synthesis
·2366 words·12 mins· loading · loading
Natural Language Processing Text Generation 🏢 Nankai University
ToMe: a novel training-free method dramatically improves semantic binding in text-to-image synthesis by intelligently merging related tokens, ensuring accurate alignment between generated images and t…
To Err Like Human: Affective Bias-Inspired Measures for Visual Emotion Recognition Evaluation
·1759 words·9 mins· loading · loading
Computer Vision Image Classification 🏢 Nankai University
This paper introduces novel metrics for visual emotion recognition evaluation, considering the psychological distance between emotions to better reflect human perception, improving the assessment of m…
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation
·2393 words·12 mins· loading · loading
Image Generation 🏢 Nankai University
StoryDiffusion enhances long-range image & video generation by introducing a simple yet effective self-attention mechanism and a semantic motion predictor, achieving high content consistency without t…
SARDet-100K: Towards Open-Source Benchmark and ToolKit for Large-Scale SAR Object Detection
·2771 words·14 mins· loading · loading
Object Detection 🏢 Nankai University
SARDet-100K: A new benchmark dataset and open-source toolkit revolutionizes large-scale SAR object detection.
Real-Time Selection Under General Constraints via Predictive Inference
·2826 words·14 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Nankai University
II-COS: a novel online sample selection method effectively controls individual and interactive constraints in real-time via predictive inference, improving efficiency and addressing various practical …
OPUS: Occupancy Prediction Using a Sparse Set
·2458 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Nankai University
OPUS: a novel, real-time occupancy prediction framework using a sparse set prediction paradigm, outperforms state-of-the-art methods on Occ3D-nuScenes.
Lighting Every Darkness with 3DGS: Fast Training and Real-Time Rendering for HDR View Synthesis
·3953 words·19 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Nankai University
LE3D: Real-time HDR view synthesis from noisy RAW images is achieved using 3DGS, significantly reducing training time and improving rendering speed.
Grid4D: 4D Decomposed Hash Encoding for High-fidelity Dynamic Gaussian Splatting
·2420 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Nankai University
Grid4D: A novel 4D decomposed hash encoding boosts high-fidelity dynamic Gaussian splatting, surpassing state-of-the-art models in visual quality and rendering speed.
Faster Diffusion: Rethinking the Role of the Encoder for Diffusion Model Inference
·5119 words·25 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 Nankai University
Faster Diffusion achieves significant speedups in diffusion model inference by cleverly reusing encoder features and enabling parallel processing, eliminating the need for computationally expensive di…
Conformalized Multiple Testing after Data-dependent Selection
·2034 words·10 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Nankai University
This paper introduces Selective Conformal P-Value (SCPV), a novel method for controlling FDR in conformalized multiple testing after data-dependent selection, offering a unified theoretical framework …
AlphaPruning: Using Heavy-Tailed Self Regularization Theory for Improved Layer-wise Pruning of Large Language Models
·3772 words·18 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Nankai University
AlphaPruning leverages Heavy-Tailed Self-Regularization theory to allocate optimal layer-wise sparsity ratios in LLMs, achieving 80% sparsity in LLaMA-7B with reasonable perplexity.