Skip to main content

Posters

2024

Toward Efficient Inference for Mixture of Experts
·2411 words·12 mins· loading · loading
Natural Language Processing Machine Translation 🏢 Duke University
Unlocking the speed and efficiency of Mixture-of-Expert models, this research unveils novel optimization techniques, achieving dramatic improvements in inference throughput and resource usage.
Toward Dynamic Non-Line-of-Sight Imaging with Mamba Enforced Temporal Consistency
·2152 words·11 mins· loading · loading
Computer Vision 3D Vision 🏢 University of Science and Technology of China
Dynamic NLOS imaging gets a speed boost! New ST-Mamba method leverages temporal consistency across frames for high-resolution video reconstruction, overcoming speed limitations of traditional methods.
Toward Conditional Distribution Calibration in Survival Prediction
·4492 words·22 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 Computing Science, University of Alberta
Boost survival prediction accuracy with CSD-iPOT: a novel post-processing method achieving superior marginal & conditional calibration without sacrificing discrimination.
Toward Approaches to Scalability in 3D Human Pose Estimation
·2344 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Korea University
Boosting 3D human pose estimation: Biomechanical Pose Generator and Binary Depth Coordinates enhance accuracy and scalability.
Toward a Well-Calibrated Discrimination via Survival Outcome-Aware Contrastive Learning
·3272 words·16 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 Chung-Ang University
ConSurv: a novel contrastive learning approach for survival analysis enhances discrimination without sacrificing calibration by employing weighted sampling and aligning well with the assumption that p…
Toward a Stable, Fair, and Comprehensive Evaluation of Object Hallucination in Large Vision-Language Models
·2235 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Harbin Institute of Technology
LeHaCE: a novel framework for evaluating object hallucination in LVLMs, improving evaluation stability and fairness by accounting for instruction-induced image description length variations.
Topological obstruction to the training of shallow ReLU neural networks
·1553 words·8 mins· loading · loading
AI Theory Optimization 🏢 Politecnico Di Torino
Shallow ReLU neural networks face topological training obstructions due to gradient flow confinement on disconnected quadric hypersurfaces.
Topological Generalization Bounds for Discrete-Time Stochastic Optimization Algorithms
·8286 words·39 mins· loading · loading
AI Generated AI Theory Generalization 🏢 University of Edinburgh
New topology-based complexity measures reliably predict deep learning model generalization, outperforming existing methods and offering practical computational efficiency.
TopoLogic: An Interpretable Pipeline for Lane Topology Reasoning on Driving Scenes
·2304 words·11 mins· loading · loading
AI Generated AI Applications Autonomous Vehicles 🏢 Institute of Computing Technology, Chinese Academy of Sciences
TopoLogic uses lane geometry and query similarity to improve lane topology reasoning in autonomous driving, significantly outperforming existing methods.
TopoFR: A Closer Look at Topology Alignment on Face Recognition
·2430 words·12 mins· loading · loading
Computer Vision Face Recognition 🏢 Zhejiang University
TopoFR enhances face recognition by aligning topological structures between input and latent spaces. Using persistent homology, it preserves crucial data structure info, overcoming overfitting. A har…
Token Merging for Training-Free Semantic Binding in Text-to-Image Synthesis
·2366 words·12 mins· loading · loading
Natural Language Processing Text Generation 🏢 Nankai University
ToMe: a novel training-free method dramatically improves semantic binding in text-to-image synthesis by intelligently merging related tokens, ensuring accurate alignment between generated images and t…
To Learn or Not to Learn, That is the Question — A Feature-Task Dual Learning Model of Perceptual Learning
·1867 words·9 mins· loading · loading
Machine Learning Transfer Learning 🏢 Peking University
A new dual-learning model resolves the paradox of perceptual learning, showing how task-based and feature-based learning interact to produce both specific and transferable improvements in sensory perc…
To Err Like Human: Affective Bias-Inspired Measures for Visual Emotion Recognition Evaluation
·1759 words·9 mins· loading · loading
Computer Vision Image Classification 🏢 Nankai University
This paper introduces novel metrics for visual emotion recognition evaluation, considering the psychological distance between emotions to better reflect human perception, improving the assessment of m…
To Believe or Not to Believe Your LLM: IterativePrompting for Estimating Epistemic Uncertainty
·1940 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Google DeepMind
This paper introduces an innovative iterative prompting method for estimating epistemic uncertainty in LLMs, enabling reliable detection of hallucinations.
TinyTTA: Efficient Test-time Adaptation via Early-exit Ensembles on Edge Devices
·2263 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Cambridge
TinyTTA enables efficient test-time adaptation on memory-constrained edge devices using a novel self-ensemble and early-exit strategy, improving accuracy and reducing memory usage.
TinyLUT: Tiny Look-Up Table for Efficient Image Restoration at the Edge
·1979 words·10 mins· loading · loading
Computer Vision Image Generation 🏢 School of Integrated Circuits, Xidian University
TinyLUT achieves 10x lower memory consumption and superior accuracy in image restoration on edge devices using innovative separable mapping and dynamic discretization of LUTs.
Tiny Time Mixers (TTMs): Fast Pre-trained Models for Enhanced Zero/Few-Shot Forecasting of Multivariate Time Series
·3445 words·17 mins· loading · loading
Machine Learning Few-Shot Learning 🏢 IBM Research
Tiny Time Mixers (TTMs) achieve state-of-the-art zero/few-shot multivariate time series forecasting, outperforming existing benchmarks while drastically reducing computational requirements.
TimeXer: Empowering Transformers for Time Series Forecasting with Exogenous Variables
·2924 words·14 mins· loading · loading
Machine Learning Deep Learning 🏢 Tsinghua University
TimeXer empowers transformers for superior time series forecasting by cleverly integrating exogenous variables, achieving state-of-the-art results on diverse benchmarks.
Time-Varying LoRA: Towards Effective Cross-Domain Fine-Tuning of Diffusion Models
·3031 words·15 mins· loading · loading
Computer Vision Image Generation 🏢 Southern University of Science and Technology
Terra, a novel time-varying low-rank adapter, enables effective cross-domain fine-tuning of diffusion models by creating a continuous parameter manifold, facilitating efficient knowledge sharing and g…
Time-FFM: Towards LM-Empowered Federated Foundation Model for Time Series Forecasting
·2669 words·13 mins· loading · loading
Machine Learning Federated Learning 🏢 Hong Kong University of Science and Technology
TIME-FFM: a Federated Foundation Model empowers time series forecasting using pre-trained Language Models, tackling data scarcity and privacy concerns for superior few-shot and zero-shot predictions.