AI Theory
Smoothed Online Classification can be Harder than Batch Classification
·302 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ University of Michigan
Smoothed online classification can be harder than batch classification when label spaces are unbounded, challenging existing assumptions in machine learning.
Smoke and Mirrors in Causal Downstream Tasks
·2586 words·13 mins·
loading
·
loading
AI Theory
Causality
π’ Institute of Science and Technology Austria
AI for science faces hidden biases in causal inference; this paper reveals these flaws using ant behavior data, introducing ISTAnt benchmark, and provides guidelines for more accurate causal AI.
Slack-Free Spiking Neural Network Formulation for Hypergraph Minimum Vertex Cover
·1466 words·7 mins·
loading
·
loading
AI Theory
Optimization
π’ Intel Labs
A novel slack-free spiking neural network efficiently solves the Hypergraph Minimum Vertex Cover problem on neuromorphic hardware, outperforming CPU-based methods in both speed and energy consumption.
SkipPredict: When to Invest in Predictions for Scheduling
·2285 words·11 mins·
loading
·
loading
AI Theory
Optimization
π’ Harvard University
SkipPredict optimizes scheduling by prioritizing cheap predictions and using expensive ones only when necessary, achieving cost-effective performance.
Sketchy Moment Matching: Toward Fast and Provable Data Selection for Finetuning
·1518 words·8 mins·
loading
·
loading
AI Theory
Generalization
π’ Courant Institute
Sketchy Moment Matching (SkMM) is a fast and theoretically sound data selection method for deep learning finetuning. By controlling variance-bias tradeoffs in high dimensions, SkMM drastically reduces…
Shuffling Gradient-Based Methods for Nonconvex-Concave Minimax Optimization
·337 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ IBM Research
New shuffling gradient methods achieve state-of-the-art oracle complexity for nonconvex-concave minimax optimization problems, offering improved performance and efficiency.
Sharpness-Aware Minimization Activates the Interactive Teaching's Understanding and Optimization
·1829 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ School of Artificial Intelligence, Jilin University
Sharpness Reduction Interactive Teaching (SRIT) boosts interactive teaching’s performance by integrating SAM’s generalization capabilities, leading to improved model accuracy and generalization.
Shaping the distribution of neural responses with interneurons in a recurrent circuit model
·1538 words·8 mins·
loading
·
loading
AI Theory
Optimization
π’ Center for Computational Neuroscience, Flatiron Institute
Researchers developed a recurrent neural circuit model that efficiently transforms sensory signals into neural representations by dynamically adjusting interneuron connectivity and activation function…
SGD vs GD: Rank Deficiency in Linear Networks
·381 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ EPFL
SGD surprisingly diminishes network rank, unlike GD, due to a repulsive force between eigenvalues, offering insights into deep learning generalization.
Sequential Signal Mixing Aggregation for Message Passing Graph Neural Networks
·2468 words·12 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Technion
Sequential Signal Mixing Aggregation (SSMA) boosts message-passing graph neural network performance by effectively mixing neighbor features, achieving state-of-the-art results across various benchmark…
Sequential Probability Assignment with Contexts: Minimax Regret, Contextual Shtarkov Sums, and Contextual Normalized Maximum Likelihood
·217 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Toronto
This paper introduces contextual Shtarkov sums, a new complexity measure characterizing minimax regret in sequential probability assignment with contexts, and derives the minimax optimal algorithm, co…
Separation and Bias of Deep Equilibrium Models on Expressivity and Learning Dynamics
·2192 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Peking University
Deep Equilibrium Models (DEQs) outperform standard neural networks, but lack theoretical understanding. This paper provides general separation results showing DEQ’s superior expressivity and character…
Semidefinite Relaxations of the Gromov-Wasserstein Distance
·2209 words·11 mins·
loading
·
loading
AI Theory
Optimization
π’ National University of Singapore
This paper introduces a novel, tractable semidefinite program (SDP) relaxation for the Gromov-Wasserstein distance, enabling the computation of globally optimal transportation plans.
Semi-Random Matrix Completion via Flow-Based Adaptive Reweighting
·349 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ MIT
New nearly-linear time algorithm achieves high-accuracy semi-random matrix completion, overcoming previous limitations on accuracy and noise tolerance.
SEEV: Synthesis with Efficient Exact Verification for ReLU Neural Barrier Functions
·1687 words·8 mins·
loading
·
loading
AI Theory
Safety
π’ Washington University in St. Louis
SEEV framework efficiently verifies ReLU neural barrier functions by reducing activation regions and using tight over-approximations, significantly improving verification efficiency without sacrificin…
Secret Collusion among AI Agents: Multi-Agent Deception via Steganography
·5189 words·25 mins·
loading
·
loading
AI Generated
AI Theory
Safety
π’ UC Berkeley
AI agents can secretly collude using steganography, hiding their interactions from oversight. This research formalizes this threat, analyzes LLMs’ capabilities, and proposes mitigation strategies.
Score-based generative models are provably robust: an uncertainty quantification perspective
·293 words·2 mins·
loading
·
loading
AI Theory
Robustness
π’ UniversitΓ© CΓ΄te D'Azur
Score-based generative models are provably robust to multiple error sources, as shown via a novel Wasserstein uncertainty propagation theorem.
Schur Nets: exploiting local structure for equivariance in higher order graph neural networks
·1825 words·9 mins·
loading
·
loading
AI Theory
Representation Learning
π’ University of Chicago
Schur Nets boost higher-order GNNs by efficiently exploiting local graph structure for automorphism equivariance, achieving improved performance without the computational burden of traditional methods…
Scaling Laws in Linear Regression: Compute, Parameters, and Data
·1352 words·7 mins·
loading
·
loading
AI Theory
Optimization
π’ UC Berkeley
Deep learning’s neural scaling laws defy conventional wisdom; this paper uses infinite-dimensional linear regression to theoretically explain this phenomenon, showing that implicit regularization of S…
Scalable Neural Network Verification with Branch-and-bound Inferred Cutting Planes
·2551 words·12 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ University of Illinois Urbana-Champaign
BICCOS: Scalable neural network verification via branch-and-bound inferred cutting planes.