Skip to main content

Posters

2024

LSH-MoE: Communication-efficient MoE Training via Locality-Sensitive Hashing
·2125 words·10 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Peking University
LSH-MoE accelerates Mixture-of-Experts training by 1.28x-2.2x via Locality-Sensitive Hashing, significantly reducing communication costs.
LRM-Zero: Training Large Reconstruction Models with Synthesized Data
·3486 words·17 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Adobe Research
LRM-Zero: Training large reconstruction models solely on synthetic data, achieving quality comparable to real-data trained models.
LP-3DGS: Learning to Prune 3D Gaussian Splatting
·2308 words·11 mins· loading · loading
Computer Vision 3D Vision 🏢 Johns Hopkins University
LP-3DGS learns to optimally prune 3D Gaussian splatting, achieving significant efficiency gains without compromising rendering quality via a trainable binary mask and the Gumbel-Sigmoid method.
Lower Bounds of Uniform Stability in Gradient-Based Bilevel Algorithms for Hyperparameter Optimization
·1822 words·9 mins· loading · loading
Machine Learning Optimization 🏢 Gaoling School of Artificial Intelligence, Renmin University of China
This paper establishes tight lower bounds for the uniform stability of gradient-based bilevel programming algorithms used for hyperparameter optimization, resolving a key open problem regarding the ti…
Lower Bounds and Optimal Algorithms for Non-Smooth Convex Decentralized Optimization over Time-Varying Networks
·1231 words·6 mins· loading · loading
Machine Learning Optimization 🏢 Yandex Research
First optimal algorithms matching lower bounds for non-smooth convex decentralized optimization over time-varying networks are presented, substantially improving theoretical performance.
Low-Rank Optimal Transport through Factor Relaxation with Latent Coupling
·2606 words·13 mins· loading · loading
Machine Learning Optimization 🏢 Princeton University
FRLC: a novel algorithm for low-rank optimal transport using latent coupling, enabling faster computation and better interpretability for diverse applications.
Low Precision Local Training is Enough for Federated Learning
·2011 words·10 mins· loading · loading
Machine Learning Federated Learning 🏢 Fudan University
Low-precision local training, surprisingly, is sufficient for accurate federated learning, significantly reducing communication and computation costs.
Low Degree Hardness for Broadcasting on Trees
·1545 words·8 mins· loading · loading
AI Theory Optimization 🏢 University of Missouri
Low-degree polynomials fail to efficiently infer roots in broadcasting tree problems below the Kesten-Stigum bound.
LOVA3: Learning to Visual Question Answering, Asking and Assessment
·3398 words·16 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Show Lab, National University of Singapore
LOVA³ enhances MLLMs by teaching them to ask and assess image-based questions, improving their multimodal understanding and performance on various benchmarks.
LoTLIP: Improving Language-Image Pre-training for Long Text Understanding
·3183 words·15 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 University of Science and Technology of China
LoTLIP boosts language-image pre-training for superior long text understanding by cleverly integrating corner tokens and utilizing a massive dataset of 100M long-caption images.
Loss Landscape Characterization of Neural Networks without Over-Parametrization
·2958 words·14 mins· loading · loading
AI Theory Optimization 🏢 University of Basel
Deep learning optimization is revolutionized by a new function class, enabling convergence guarantees without over-parameterization and accommodating saddle points.
Lorentz-Equivariant Geometric Algebra Transformers for High-Energy Physics
·1502 words·8 mins· loading · loading
Machine Learning Deep Learning 🏢 Heidelberg University
Lorentz Geometric Algebra Transformer (L-GATr): A novel, scalable architecture for high-energy physics, achieving high-precision, data-efficient learning and outperforming existing methods on regressi…
LoRA-GA: Low-Rank Adaptation with Gradient Approximation
·2382 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Tsinghua University
LoRA-GA: A novel initialization method dramatically speeds up low-rank adaptation (LoRA) for LLMs, achieving convergence rates comparable to full fine-tuning while improving performance.
LoQT: Low-Rank Adapters for Quantized Pretraining
·2483 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Copenhagen
LoQT enables efficient large language model training on consumer hardware via quantized weights and low-rank weight updates, overcoming memory limitations.
Looks Too Good To Be True: An Information-Theoretic Analysis of Hallucinations in Generative Restoration Models
·2111 words·10 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Verily AI (Google Life Sciences)
Generative image restoration models face a critical trade-off: higher perceptual quality often leads to increased hallucinations (unreliable predictions).
LookHere: Vision Transformers with Directed Attention Generalize and Extrapolate
·4927 words·24 mins· loading · loading
Computer Vision Image Classification 🏢 Carleton University
LookHere: Vision Transformers excel at high-resolution image classification by using 2D attention masks to direct attention heads, improving generalization and extrapolation.
Lookback Prophet Inequalities
·467 words·3 mins· loading · loading
AI Theory Optimization 🏢 ENSAE, Ecole Polytechnique
This paper enhances prophet inequalities by allowing lookback, improving competitive ratios and providing algorithms for diverse observation orders, thereby bridging theory and real-world online selec…
Look, Listen, and Answer: Overcoming Biases for Audio-Visual Question Answering
·2344 words·12 mins· loading · loading
Natural Language Processing Question Answering 🏢 Xi'an Jiaotong University
New dataset MUSIC-AVQA-R and a multi-faceted cycle collaborative debiasing strategy significantly improve audio-visual question answering robustness.
Long-Tailed Out-of-Distribution Detection via Normalized Outlier Distribution Adaptation
·2335 words·11 mins· loading · loading
Computer Vision Object Detection 🏢 Beihang University
AdaptOD: a novel approach for robust OOD detection in long-tailed recognition, dynamically adapting outlier distributions to true OOD distributions using a dual-normalized energy loss for improved acc…
Long-tailed Object Detection Pretraining: Dynamic Rebalancing Contrastive Learning with Dual Reconstruction
·2225 words·11 mins· loading · loading
Computer Vision Object Detection 🏢 Nanjing University of Science and Technology
Dynamic Rebalancing Contrastive Learning with Dual Reconstruction (2DRCL) pre-training significantly boosts object detection accuracy, especially for underrepresented classes.