Machine Learning
Segment, Shuffle, and Stitch: A Simple Layer for Improving Time-Series Representations
·3043 words·15 mins·
loading
·
loading
Machine Learning
Representation Learning
π’ Queen's University
Boost time-series model accuracy with Segment, Shuffle, and Stitch (S3)! This simple layer shuffles data segments to enhance representation learning, improving classification, forecasting, and anomaly…
Seek Commonality but Preserve Differences: Dissected Dynamics Modeling for Multi-modal Visual RL
·2815 words·14 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Peking University
Dissected Dynamics Modeling (DDM) excels at multi-modal visual reinforcement learning by cleverly separating and integrating common and unique features across different sensory inputs for more accurat…
Searching for Efficient Linear Layers over a Continuous Space of Structured Matrices
·2763 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ New York University
Revolutionizing large neural networks, this paper introduces a continuous parameterization of structured matrices, discovering that full-rank structures without parameter sharing achieve optimal scali…
Score-Optimal Diffusion Schedules
·2200 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ University of Oxford
Researchers developed a novel algorithm to automatically find optimal schedules for denoising diffusion models (DDMs), significantly improving sample quality and efficiency without manual parameter tu…
Score-based 3D molecule generation with neural fields
·4106 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Prescient Design
FuncMol: A new neural field model generates 3D molecules efficiently, outperforming existing methods by achieving an order of magnitude faster sampling speed.
Scanning Trojaned Models Using Out-of-Distribution Samples
·2406 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Sharif University of Technology
TRODO: a novel trojan detection method using out-of-distribution samples, effectively identifies trojaned classifiers even against adversarial attacks and with limited data.
Scaling transformer neural networks for skillful and reliable medium-range weather forecasting
·2398 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ UC Los Angeles
Stormer, a simple transformer model, achieves state-of-the-art medium-range weather forecasting accuracy by using weather-specific embedding, randomized dynamics forecasting, and a pressure-weighted l…
Scaling laws for learning with real and surrogate data
·4942 words·24 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Granica Computing Inc.
Boost machine learning with surrogate data! A novel weighted ERM method effectively integrates surrogate data, significantly reducing test errors even with unrelated data, guided by a predictable sca…
Scaling Law for Time Series Forecasting
·2211 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Tsinghua University
Unlocking the potential of deep learning for time series forecasting: this study reveals a scaling law influenced by dataset size, model complexity, and the crucial look-back horizon, leading to impro…
Scale-invariant Optimal Sampling for Rare-events Data and Sparse Models
·2539 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ University of Connecticut
Scale-invariant optimal subsampling tackles computational challenges in analyzing massive rare-events data with sparse models, enhancing parameter estimation and variable selection without being affec…
Scalable Optimization in the Modular Norm
·2001 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ MIT
Deep learning optimization gets a major upgrade with Modula, a new method that uses the modular norm to normalize weight updates, enabling learning rate transfer across network widths and depths, thus…
Scalable Kernel Inverse Optimization
·1850 words·9 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ Delft Center for Systems and Control
Scalable Kernel Inverse Optimization (KIO) efficiently learns unknown objective functions from data using kernel methods and a novel Sequential Selection Optimization (SSO) algorithm, enabling applica…
Scalable DBSCAN with Random Projections
·4595 words·22 mins·
loading
·
loading
AI Generated
Machine Learning
Clustering
π’ University of Auckland
sDBSCAN: Blazing-fast density-based clustering for massive datasets using random projections!
Scalable Constrained Policy Optimization for Safe Multi-agent Reinforcement Learning
·1664 words·8 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ Peking University
Scalable MAPPO-L: Decentralized training with local interactions ensures safe, high-reward multi-agent systems, even with limited communication.
Scalable Bayesian Optimization via Focalized Sparse Gaussian Processes
·2563 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
π’ Tsinghua University
FOCALBO, a hierarchical Bayesian optimization algorithm using focalized sparse Gaussian processes, efficiently tackles high-dimensional problems with massive datasets, achieving state-of-the-art perfo…
SCAFFLSA: Taming Heterogeneity in Federated Linear Stochastic Approximation and TD Learning
·1913 words·9 mins·
loading
·
loading
Machine Learning
Federated Learning
π’ CMAP, UMR 7641, Γcole Polytechnique
SCAFFLSA tames heterogeneity in federated learning, achieving logarithmic communication complexity and linear sample complexity.
SARAD: Spatial Association-Aware Anomaly Detection and Diagnosis for Multivariate Time Series
·3459 words·17 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ University of Warwick
SARAD: A novel anomaly detection approach for multivariate time series leverages spatial information and association reduction patterns to achieve state-of-the-art performance.
SAND: Smooth imputation of sparse and noisy functional data with Transformer networks
·1394 words·7 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ UC Davis
SAND, a novel transformer network variant, smoothly imputes sparse and noisy functional data by leveraging self-attention on derivatives, outperforming existing methods.
Sample-Efficient Constrained Reinforcement Learning with General Parameterization
·263 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Indian Institute of Technology Kanpur
Accelerated Primal-Dual Natural Policy Gradient (PD-ANPG) algorithm achieves a theoretical lower bound sample complexity for solving general parameterized CMDPs, improving state-of-the-art by a factor…
Sample-Efficient Agnostic Boosting
·1303 words·7 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Amazon
Agnostic boosting gets a major efficiency upgrade! A new algorithm leverages sample reuse to drastically reduce the data needed for accurate learning, closing the gap with computationally expensive al…