Skip to main content

Spotlight Others

2024

EMR-Merging: Tuning-Free High-Performance Model Merging
·3173 words·15 mins· loading · loading
Image Classification 🏒 Fudan University
EMR-MERGING: A tuning-free model merging technique achieves high performance by electing a unified model and generating lightweight task-specific modulators, eliminating the need for additional data …
EigenVI: score-based variational inference with orthogonal function expansions
·2049 words·10 mins· loading · loading
Variational Inference 🏒 Flatiron Institute
EigenVI: a novel score-based variational inference method using orthogonal function expansions, offers closed-form solutions by solving eigenvalue problems, outperforming existing Gaussian BBVI method…
Dynamic 3D Gaussian Fields for Urban Areas
·2544 words·12 mins· loading · loading
3D Vision 🏒 ETH Zurich
4DGF, a novel neural scene representation, achieves interactive-speed novel view synthesis for large-scale dynamic urban areas by efficiently combining 3D Gaussians and neural fields.
Double-Ended Synthesis Planning with Goal-Constrained Bidirectional Search
·2386 words·12 mins· loading · loading
AI Applications Healthcare 🏒 Massachusetts Institute of Technology
Double-Ended Synthesis Planning (DESP) significantly boosts computer-aided synthesis planning by using a bidirectional search, outperforming existing methods on multiple benchmarks, especially when sp…
Doob's Lagrangian: A Sample-Efficient Variational Approach to Transition Path Sampling
·1775 words·9 mins· loading · loading
🏒 Cornell University
Researchers developed a sample-efficient variational approach for transition path sampling using Doob’s h-transform, significantly reducing computational costs while accurately capturing transition pa…
Don't Look Twice: Faster Video Transformers with Run-Length Tokenization
·1863 words·9 mins· loading · loading
Video Understanding 🏒 Carnegie Mellon University
Run-Length Tokenization (RLT) dramatically speeds up video transformer training and inference by efficiently removing redundant video tokens, matching baseline model performance with significant time …
Diversity-Driven Synthesis: Enhancing Dataset Distillation through Directed Weight Adjustment
·2093 words·10 mins· loading · loading
Image Classification 🏒 Agency for Science, Technology and Research (A*STAR)
Boosting dataset distillation, a new method, Diversity-Driven Synthesis, uses directed weight adjustment to create diverse, representative synthetic datasets, improving model performance while reducin…
Distributed-Order Fractional Graph Operating Network
·2406 words·12 mins· loading · loading
🏒 Nanyang Technological University
DRAGON: A novel GNN framework using distributed-order fractional calculus surpasses traditional methods by capturing complex graph dynamics with enhanced flexibility and performance.
Dissecting Query-Key Interaction in Vision Transformers
·3134 words·15 mins· loading · loading
Vision Transformers 🏒 University of Miami
Vision transformers’ self-attention mechanism is dissected revealing how early layers focus on similar features for perceptual grouping while later layers integrate dissimilar features for contextuali…
Disentangling the Roles of Distinct Cell Classes with Cell-Type Dynamical Systems
·2009 words·10 mins· loading · loading
🏒 Princeton University
New Cell-Type Dynamical Systems (CTDS) model disentangles neural population dynamics by incorporating distinct cell types, improving prediction accuracy and biological interpretability.
Diffusion Priors for Variational Likelihood Estimation and Image Denoising
·2184 words·11 mins· loading · loading
Image Generation 🏒 Huazhong University of Science and Technology
Adaptive likelihood estimation and MAP inference during reverse diffusion tackles real-world image noise.
Diffusion Models With Learned Adaptive Noise
·2619 words·13 mins· loading · loading
🏒 Cornell University
MuLAN, a novel learned diffusion process, achieves state-of-the-art density estimation by adaptively adding multivariate Gaussian noise at varying rates across an image, significantly reducing trainin…
Diffusion Model with Cross Attention as an Inductive Bias for Disentanglement
·3205 words·16 mins· loading · loading
Representation Learning 🏒 Microsoft Research
Diffusion models with cross-attention: a powerful inductive bias for effortless disentanglement!
DiffSF: Diffusion Models for Scene Flow Estimation
·2351 words·12 mins· loading · loading
Scene Understanding 🏒 Linkâping University
DiffSF boosts scene flow estimation accuracy and reliability by cleverly combining transformer networks with denoising diffusion models, offering state-of-the-art results and uncertainty quantificatio…
DiffLight: A Partial Rewards Conditioned Diffusion Model for Traffic Signal Control with Missing Data
·2583 words·13 mins· loading · loading
AI Applications Smart Cities 🏒 Beijing Jiaotong University
DiffLight: a novel conditional diffusion model for traffic signal control effectively addresses data-missing scenarios by unifying traffic data imputation and decision-making, demonstrating superior p…
Differentiable Task Graph Learning: Procedural Activity Representation and Online Mistake Detection from Egocentric Videos
·4229 words·20 mins· loading · loading
Video Understanding 🏒 University of Catania
This paper introduces a novel differentiable framework for learning task graphs from video demonstrations of procedural activities. By directly optimizing the weights of a task graph’s edges, the mod…
DeSparsify: Adversarial Attack Against Token Sparsification Mechanisms
·2446 words·12 mins· loading · loading
Image Classification 🏒 Ben-Gurion University of the Negev
DeSparsify: A stealthy adversarial attack exhausts vision transformer resources by exploiting token sparsification mechanisms’ dynamic nature, highlighting the need for improved resource management i…
Deep Submodular Peripteral Networks
·2349 words·12 mins· loading · loading
🏒 University of Washington
Deep Submodular Peripteral Networks (DSPNs) learn submodular functions efficiently using graded pairwise comparisons, surpassing traditional methods and demonstrating superiority in experimental desig…
Deep Learning for Computing Convergence Rates of Markov Chains
·1326 words·7 mins· loading · loading
🏒 Stanford University
Deep learning tackles Markov chain convergence rate analysis! Deep Contractive Drift Calculator (DCDC) provides sample-based bounds in Wasserstein distance, surpassing traditional methods’ limitations…
CycleNet: Enhancing Time Series Forecasting through Modeling Periodic Patterns
·2454 words·12 mins· loading · loading
🏒 South China University of Technology
CycleNet enhances long-term time series forecasting by explicitly modeling inherent periodic patterns using a novel Residual Cycle Forecasting technique, achieving state-of-the-art accuracy and effici…