Skip to main content

Spotlight Others

2024

Are Language Models Actually Useful for Time Series Forecasting?
·3629 words·18 mins· loading · loading
AI Applications Finance 🏢 University of Virginia
Popular large language model (LLM)-based time series forecasting methods perform no better than simpler alternatives, often worse, and require vastly more compute.
Any2Graph: Deep End-To-End Supervised Graph Prediction With An Optimal Transport Loss
·3100 words·15 mins· loading · loading
🏢 Télécom Paris, IP Paris
Any2Graph: a novel deep learning framework using an Optimal Transport loss for accurate and efficient supervised graph prediction.
Analysing Multi-Task Regression via Random Matrix Theory with Application to Time Series Forecasting
·1646 words·8 mins· loading · loading
🏢 Huawei Noah's Ark Lab
This paper presents a novel theoretical framework for multi-task regression using random matrix theory, offering precise performance estimations and a closed-form solution for optimal hyperparameter t…
Aligner-Encoders: Self-Attention Transformers Can Be Self-Transducers
·2029 words·10 mins· loading · loading
Speech Recognition 🏢 Google
Transformers can now perform self-alignment, enabling simpler, faster speech recognition models.
Algebraic Positional Encodings
·1392 words·7 mins· loading · loading
Machine Translation 🏢 Aalto University
Revolutionizing Transformers, Algebraic Positional Encodings (APE) offers a theory-first approach to positional encoding, outperforming state-of-the-art methods without hyperparameter tuning across va…
Advancing Spiking Neural Networks for Sequential Modeling with Central Pattern Generators
·1882 words·9 mins· loading · loading
🏢 Microsoft Research
Bio-inspired CPG-PE enhances spiking neural networks’ sequential modeling by efficiently encoding position information, outperforming conventional methods across various tasks.
Adaptive Randomized Smoothing: Certified Adversarial Robustness for Multi-Step Defences
·3521 words·17 mins· loading · loading
Image Classification 🏢 University of British Columbia
Adaptive Randomized Smoothing certifies deep learning model predictions against adversarial attacks by cleverly combining randomized smoothing with adaptive, multi-step input masking for improved accu…
Adaptive Image Quality Assessment via Teaching Large Multimodal Model to Compare
·2147 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 City University of Hong Kong
Compare2Score: A novel IQA model teaches large multimodal models to translate comparative image quality judgments into continuous quality scores, significantly outperforming existing methods.
Active Classification with Few Queries under Misspecification
·253 words·2 mins· loading · loading
Active Learning 🏢 University of Texas at Austin
Learning halfspaces efficiently under noise is cracked! A novel query language enables a polylog query algorithm for Massart noise, overcoming previous limitations.
Acoustic Volume Rendering for Neural Impulse Response Fields
·2052 words·10 mins· loading · loading
Speech and Audio Acoustic Scene Analysis 🏢 University of Pennsylvania
Acoustic Volume Rendering (AVR) revolutionizes realistic audio synthesis by adapting volume rendering to model acoustic impulse responses, achieving state-of-the-art performance in novel pose synthesi…
ACES: Generating a Diversity of Challenging Programming Puzzles with Autotelic Generative Models
·2681 words·13 mins· loading · loading
🏢 Inria
Autotelic Code Search (ACES) generates diverse, challenging Python programming puzzles by iteratively using LLM-generated semantic descriptors and measuring puzzle difficulty via LLM solver success ra…
Accelerating Diffusion Models with Parallel Sampling: Inference at Sub-Linear Time Complexity
·418 words·2 mins· loading · loading
🏢 Stanford University
Researchers achieve sub-linear time complexity for diffusion model inference using parallel sampling with poly-logarithmic time complexity.
A-FedPD: Aligning Dual-Drift is All Federated Primal-Dual Learning Needs
·2374 words·12 mins· loading · loading
Federated Learning 🏢 University of Sydney
A-FedPD tackles federated learning’s ‘dual drift’ problem by aligning global and local dual variables, resulting in faster convergence and enhanced stability for primal-dual methods.
A Unified Debiasing Approach for Vision-Language Models across Modalities and Tasks
·2218 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Purdue University
SFID, a novel debiasing method, effectively mitigates bias in vision-language models across various tasks without retraining, improving fairness and efficiency.
A Textbook Remedy for Domain Shifts: Knowledge Priors for Medical Image Analysis
·2721 words·13 mins· loading · loading
Image Classification 🏢 University of Pennsylvania
KnoBo enhances deep learning models for medical image analysis by incorporating knowledge priors from medical textbooks, boosting out-of-domain performance by up to 32.4%.
A Pairwise Pseudo-likelihood Approach for Matrix Completion with Informative Missingness
·1982 words·10 mins· loading · loading
🏢 Texas A&M University
New method recovers low-rank matrices with informative missingness, offering robust, near-optimal performance.
A Near-optimal Algorithm for Learning Margin Halfspaces with Massart Noise
·223 words·2 mins· loading · loading
🏢 University of Washington
Near-optimal algorithm achieves computationally efficient learning of margin halfspaces with Massart noise, nearly matching theoretical lower bounds.
A Geometric View of Data Complexity: Efficient Local Intrinsic Dimension Estimation with Diffusion Models
·3928 words·19 mins· loading · loading
🏢 Layer 6 AI
Diffusion models power FLIPD, a fast, single-model LID estimator.
3DGS-Enhancer: Enhancing Unbounded 3D Gaussian Splatting with View-consistent 2D Diffusion Priors
·2090 words·10 mins· loading · loading
3D Vision 🏢 Clemson University
3DGS-Enhancer boosts unbounded 3D Gaussian splatting, generating high-fidelity novel views even with sparse input data using view-consistent 2D diffusion priors.
3D Gaussian Splatting as Markov Chain Monte Carlo
·1616 words·8 mins· loading · loading
3D Vision 🏢 University of British Columbia
Researchers rethink 3D Gaussian Splatting as MCMC sampling, improving rendering quality and Gaussian control via a novel relocation strategy.