Deep Learning
Are Multiple Instance Learning Algorithms Learnable for Instances?
·2369 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Graduate School of Data Science, Seoul National University of Science and Technology
Deep MIL algorithms’ instance-level learnability is theoretically proven, revealing crucial conditions for success and highlighting gaps in existing models.
Approximation Rate of the Transformer Architecture for Sequence Modeling
·1599 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 CNRS@CREATE LTD
This paper unveils the Transformer’s approximation power, deriving explicit Jackson-type rates to reveal its strengths and limitations in handling various sequential relationships.
Approximately Equivariant Neural Processes
·2389 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Cambridge
Boosting meta-learning, this paper introduces a novel, flexible approach to create approximately equivariant neural processes that outperform both non-equivariant and strictly equivariant counterparts…
ANT: Adaptive Noise Schedule for Time Series Diffusion Models
·4333 words·21 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Yonsei University
ANT: An adaptive noise schedule automatically determines optimal noise schedules for time series diffusion models, significantly boosting performance across diverse tasks.
An exactly solvable model for emergence and scaling laws in the multitask sparse parity problem
·2679 words·13 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Oxford
A novel multilinear model analytically explains the emergence and scaling laws of skills in the multitask sparse parity problem, accurately predicting skill emergence in neural networks.
Amortized Fourier Neural Operators
·2070 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Qing Yuan Research Institute, SEIEE, Shanghai Jiao Tong University
Amortized Fourier Neural Operators (AM-FNOs) dramatically improve efficiency in solving PDEs by using neural networks for kernel parameterization, achieving up to 31% better accuracy compared to exist…
Aligning Target-Aware Molecule Diffusion Models with Exact Energy Optimization
·2379 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Stanford University
ALIDIFF aligns target-aware molecule diffusion models with exact energy optimization, generating molecules with state-of-the-art binding energies and improved properties.
Alias-Free Mamba Neural Operator
·2498 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Zhejiang University of Technology
MambaNO: a novel neural operator achieving linear complexity and state-of-the-art accuracy in solving PDEs by cleverly balancing global and local information using an alias-free architecture.
Advancing Training Efficiency of Deep Spiking Neural Networks through Rate-based Backpropagation
·2050 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Zhejiang University
Rate-based backpropagation boosts deep spiking neural network training efficiency by leveraging rate coding, achieving comparable performance to BPTT with reduced complexity.
ADOPT: Modified Adam Can Converge with Any $eta_2$ with the Optimal Rate
·1889 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Tokyo
ADOPT, a novel adaptive gradient method, achieves optimal convergence rates without restrictive assumptions, unlike Adam, significantly improving deep learning optimization.
Addressing Spectral Bias of Deep Neural Networks by Multi-Grade Deep Learning
·2240 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Department of Mathematics and Statistics, Old Dominion University
Multi-Grade Deep Learning (MGDL) conquers spectral bias in deep neural networks by incrementally learning low-frequency components, ultimately capturing high-frequency features through composition.
Adaptive Passive-Aggressive Framework for Online Regression with Side Information
·2153 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Hong Kong University of Science and Technology
Adaptive Passive-Aggressive framework with Side information (APAS) significantly boosts online regression accuracy by dynamically adjusting thresholds and integrating side information, leading to supe…
Adaptive Depth Networks with Skippable Sub-Paths
·3982 words·19 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Incheon National University
Adaptive Depth Networks with Skippable Sub-Paths: Train once, deploy efficiently! This paper proposes a novel training method to create adaptive-depth networks, enabling on-demand model depth selectio…
Adapting to Unknown Low-Dimensional Structures in Score-Based Diffusion Models
·393 words·2 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Chinese University of Hong Kong
Score-based diffusion models are improved by a novel coefficient design, enabling efficient adaptation to unknown low-dimensional data structures and achieving a convergence rate of O(k²/√T).
Ada-MSHyper: Adaptive Multi-Scale Hypergraph Transformer for Time Series Forecasting
·3027 words·15 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Zhejiang University
Ada-MSHyper: A novel adaptive multi-scale hypergraph transformer significantly boosts time series forecasting accuracy by modeling group-wise interactions and handling complex temporal variations.
Activation Map Compression through Tensor Decomposition for Deep Learning
·2035 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Telecom Paris
Slash deep learning’s memory footprint! This paper introduces a novel activation map compression technique via tensor decomposition, significantly boosting on-device training efficiency for edge AI.
Accelerating Relative Entropy Coding with Space Partitioning
·1881 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Cambridge
Space partitioning dramatically speeds up relative entropy coding (REC) for neural compression, achieving 5-15% better bitrates than previous methods.
Absorb & Escape: Overcoming Single Model Limitations in Generating Heterogeneous Genomic Sequences
·3759 words·18 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Imperial College London
Absorb & Escape: a novel post-training sampling method that overcomes single model limitations by combining Autoregressive (AR) and Diffusion Models (DMs), generating high-quality heterogeneous genomi…
A versatile informative diffusion model for single-cell ATAC-seq data generation and analysis
·1769 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 City University of Hong Kong
ATAC-Diff: A versatile diffusion model for high-quality single-cell ATAC-seq data generation and analysis, surpassing state-of-the-art.
A two-scale Complexity Measure for Deep Learning Models
·1709 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 IBM Research
New 2sED measure effectively bounds deep learning model complexity, correlating well with training error and offering efficient computation, particularly for deep models via a layerwise approach.