Skip to main content

Machine Learning

Towards a Scalable Reference-Free Evaluation of Generative Models
·3926 words·19 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Chinese University of Hong Kong
FKEA: a novel, scalable method for reference-free evaluation of generative models’ diversity using random Fourier features, overcoming computational limitations of existing entropy-based scores.
Towards a 'Universal Translator' for Neural Dynamics at Single-Cell, Single-Spike Resolution
·2778 words·14 mins· loading · loading
Machine Learning Self-Supervised Learning 🏢 Columbia University
A new self-supervised learning approach, Multi-task Masking (MtM), significantly improves the prediction accuracy of neural population activity by capturing neural dynamics at multiple spatial scales,…
Toward Global Convergence of Gradient EM for Over-Paramterized Gaussian Mixture Models
·345 words·2 mins· loading · loading
Machine Learning Optimization 🏢 University of Washington
Gradient EM for over-parameterized Gaussian Mixture Models globally converges with a sublinear rate, solving a longstanding open problem in machine learning.
To Learn or Not to Learn, That is the Question — A Feature-Task Dual Learning Model of Perceptual Learning
·1867 words·9 mins· loading · loading
Machine Learning Transfer Learning 🏢 Peking University
A new dual-learning model resolves the paradox of perceptual learning, showing how task-based and feature-based learning interact to produce both specific and transferable improvements in sensory perc…
TinyTTA: Efficient Test-time Adaptation via Early-exit Ensembles on Edge Devices
·2263 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Cambridge
TinyTTA enables efficient test-time adaptation on memory-constrained edge devices using a novel self-ensemble and early-exit strategy, improving accuracy and reducing memory usage.
Tiny Time Mixers (TTMs): Fast Pre-trained Models for Enhanced Zero/Few-Shot Forecasting of Multivariate Time Series
·3445 words·17 mins· loading · loading
Machine Learning Few-Shot Learning 🏢 IBM Research
Tiny Time Mixers (TTMs) achieve state-of-the-art zero/few-shot multivariate time series forecasting, outperforming existing benchmarks while drastically reducing computational requirements.
TimeXer: Empowering Transformers for Time Series Forecasting with Exogenous Variables
·2924 words·14 mins· loading · loading
Machine Learning Deep Learning 🏢 Tsinghua University
TimeXer empowers transformers for superior time series forecasting by cleverly integrating exogenous variables, achieving state-of-the-art results on diverse benchmarks.
Time-FFM: Towards LM-Empowered Federated Foundation Model for Time Series Forecasting
·2669 words·13 mins· loading · loading
Machine Learning Federated Learning 🏢 Hong Kong University of Science and Technology
TIME-FFM: a Federated Foundation Model empowers time series forecasting using pre-trained Language Models, tackling data scarcity and privacy concerns for superior few-shot and zero-shot predictions.
Time-Constrained Robust MDPs
·10005 words·47 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 IRT Saint-Exupéry
Time-Constrained Robust MDPs (TC-RMDPs) improve reinforcement learning by addressing limitations of traditional methods, offering a novel framework for handling real-world uncertainties and yielding m…
Time Makes Space: Emergence of Place Fields in Networks Encoding Temporally Continuous Sensory Experiences
·3838 words·19 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 University of Pennsylvania
Networks trained on continuous sensory data spontaneously develop place cell-like responses, demonstrating that time-encoded experience can create spatial maps in the brain.
This Too Shall Pass: Removing Stale Observations in Dynamic Bayesian Optimization
·4337 words·21 mins· loading · loading
AI Generated Machine Learning Optimization 🏢 IC, EPFL
W-DBO efficiently tackles stale data in dynamic Bayesian Optimization by leveraging a novel Wasserstein distance-based criterion to remove irrelevant observations, maintaining high sampling frequency …
Theoretical Investigations and Practical Enhancements on Tail Task Risk Minimization in Meta Learning
·3609 words·17 mins· loading · loading
AI Generated Machine Learning Meta Learning 🏢 College of Science, National University of Defense Technology
This research enhances meta-learning robustness by theoretically grounding and practically improving tail-risk minimization, achieving improved fast adaptation in the task space.
The tree autoencoder model, with application to hierarchical data visualization
·2243 words·11 mins· loading · loading
Machine Learning Unsupervised Learning 🏢 Dept. of Computer Science and Engineering, University of California, Merced
PCA tree: a novel hierarchical dimensionality reduction model visualized using oblique trees and local PCAs, offering speed and interpretability.
The Surprising Ineffectiveness of Pre-Trained Visual Representations for Model-Based Reinforcement Learning
·2250 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Bosch Center for Artificial Intelligence
Contrary to expectations, pre-trained visual representations surprisingly don’t improve model-based reinforcement learning’s sample efficiency or generalization; data diversity and network architectu…
The surprising efficiency of temporal difference learning for rare event prediction
·1614 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Courant Institute of Mathematical Sciences, New York University
TD learning surprisingly outperforms Monte Carlo methods for rare event prediction in Markov chains, achieving relative accuracy with polynomially, instead of exponentially, many observed transitions.
The Star Geometry of Critic-Based Regularizer Learning
·1709 words·9 mins· loading · loading
Machine Learning Unsupervised Learning 🏢 University of California, Los Angeles
Star geometry reveals optimal data-driven regularizers!
The Selective $G$-Bispectrum and its Inversion: Applications to $G$-Invariant Networks
·2369 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 UCLouvain
This paper introduces a selective G-Bispectrum algorithm, slashing the computational complexity from O(|G|^2) to O(|G|), making G-invariant deep learning faster and more scalable.
The Prevalence of Neural Collapse in Neural Multivariate Regression
·2414 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 New York University Abu Dhabi
Neural networks exhibit ‘Neural Regression Collapse’ (NRC) during training, where feature vectors collapse to subspaces spanned by principal components of features and weights, and the weight vector G…
The Power of Extrapolation in Federated Learning
·2710 words·13 mins· loading · loading
AI Generated Machine Learning Federated Learning 🏢 GenAI Center of Excellence
Federated learning gets a speed boost: New extrapolation strategies significantly improve FedProx’s convergence, offering both theoretical backing and practical enhancements.
The Poisson Midpoint Method for Langevin Dynamics: Provably Efficient Discretization for Diffusion Models
·2297 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 Cornell University
Poisson Midpoint Method quadratically accelerates Langevin Monte Carlo for diffusion models, achieving high-quality image generation with significantly fewer computations.