Skip to main content

Deep Learning

Scalable Optimization in the Modular Norm
·2001 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 MIT
Deep learning optimization gets a major upgrade with Modula, a new method that uses the modular norm to normalize weight updates, enabling learning rate transfer across network widths and depths, thus…
SARAD: Spatial Association-Aware Anomaly Detection and Diagnosis for Multivariate Time Series
·3459 words·17 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Warwick
SARAD: A novel anomaly detection approach for multivariate time series leverages spatial information and association reduction patterns to achieve state-of-the-art performance.
SAND: Smooth imputation of sparse and noisy functional data with Transformer networks
·1394 words·7 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Davis
SAND, a novel transformer network variant, smoothly imputes sparse and noisy functional data by leveraging self-attention on derivatives, outperforming existing methods.
Sample Selection via Contrastive Fragmentation for Noisy Label Regression
·6755 words·32 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Seoul National University
ConFrag, a novel approach to noisy label regression, leverages contrastive fragmentation and neighborhood agreement to select clean samples, significantly outperforming state-of-the-art baselines on s…
SampDetox: Black-box Backdoor Defense via Perturbation-based Sample Detoxification
·2579 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 Singapore Management University
SampDetox uses diffusion models to purify poisoned machine learning samples by strategically adding noise to eliminate backdoors without compromising data integrity.
S2HPruner: Soft-to-Hard Distillation Bridges the Discretization Gap in Pruning
·2415 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 Fudan University
S2HPruner bridges the discretization gap in neural network pruning via a novel soft-to-hard distillation framework, achieving superior performance across various benchmarks without fine-tuning.
Rough Transformers: Lightweight Continuous-Time Sequence Modelling with Path Signatures
·4237 words·20 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 University of Oxford
Rough Transformers: A lightweight continuous-time sequence modeling approach using path signatures to significantly reduce computational costs, improving efficiency and accuracy, particularly for long…
Robust group and simultaneous inferences for high-dimensional single index model
·473 words·3 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Beijing Normal University
This paper introduces robust group inference procedures for high-dimensional single index models, offering substantial efficiency gains for heavy-tailed errors and handling group testing effectively w…
RMLR: Extending Multinomial Logistic Regression into General Geometries
·2575 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Trento
RMLR: A novel framework extends multinomial logistic regression to diverse geometries, overcoming limitations of existing methods by requiring minimal geometric properties for broad applicability.
Revive Re-weighting in Imbalanced Learning by Density Ratio Estimation
·2405 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 School of Artificial Intelligence, Shanghai Jiao Tong University
Revive Re-weighting in Imbalanced Learning by Density Ratio Estimation dynamically adjusts class weights during training using density ratio estimation, significantly improving model generalization, e…
Retrieval-Augmented Diffusion Models for Time Series Forecasting
·2360 words·12 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Peking University
Boosting time series forecasting accuracy, Retrieval-Augmented Diffusion Models (RATD) leverage relevant historical data to guide the diffusion process, overcoming limitations of existing models and d…
Retrieval & Fine-Tuning for In-Context Tabular Models
·3508 words·17 mins· loading · loading
Machine Learning Deep Learning 🏢 Layer6
LoCalPFN: boosting in-context tabular learning via retrieval & fine-tuning!
Rethinking the Membrane Dynamics and Optimization Objectives of Spiking Neural Networks
·2616 words·13 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 College of Artificial Intelligence, Southwest University
Boosting spiking neural network accuracy by 4.05% on ImageNet and achieving state-of-the-art results on CIFAR10-DVS and N-Caltech101 through learnable initial membrane potential and refined training s…
Rethinking Fourier Transform from A Basis Functions Perspective for Long-term Time Series Forecasting
·1994 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 School of Computing, Macquarie University
Revolutionizing long-term time series forecasting, a new Fourier Basis Mapping method enhances accuracy by precisely interpreting frequency coefficients and considering time-frequency relationships, a…
Rethinking Deep Thinking: Stable Learning of Algorithms using Lipschitz Constraints
·2740 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Southampton
Stable algorithm learning achieved by Deep Thinking networks with Lipschitz Constraints, ensuring convergence and better extrapolation to complex problems.
Reparameterized Multi-Resolution Convolutions for Long Sequence Modelling
·3179 words·15 mins· loading · loading
Machine Learning Deep Learning 🏢 University College London
MRConv: Reparameterized multi-resolution convolutions efficiently model long sequences, improving performance across various data modalities.
Relational Concept Bottleneck Models
·2454 words·12 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 University of Cambridge
Relational Concept Bottleneck Models (R-CBMs) merge interpretable CBMs with powerful GNNs for high-performing, explainable relational deep learning.
Rejection via Learning Density Ratios
·2534 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 Australian National University
This paper introduces a novel framework for classification with rejection by learning density ratios between data and idealized distributions, improving model robustness and accuracy.
Regularized Adaptive Momentum Dual Averaging with an Efficient Inexact Subproblem Solver for Training Structured Neural Network
·2322 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 National Taiwan University
RAMDA: a new algorithm ensures efficient training of structured neural networks by achieving optimal structure and outstanding predictive performance.
REDUCR: Robust Data Downsampling using Class Priority Reweighting
·2544 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 University College London
REDUCR, a novel data downsampling method, significantly improves worst-class test accuracy in imbalanced datasets by using class priority reweighting, surpassing state-of-the-art methods by ~15%.