AI Theory
MILP-StuDio: MILP Instance Generation via Block Structure Decomposition
·3596 words·17 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Science and Technology of China
MILP-StuDio generates high-quality mixed-integer linear programming instances by preserving crucial block structures, significantly improving learning-based solver performance.
MG-Net: Learn to Customize QAOA with Circuit Depth Awareness
·2515 words·12 mins·
loading
·
loading
AI Theory
Optimization
π’ School of Computer Science, Faculty of Engineering, University of Sydney
MG-Net dynamically designs optimal mixer Hamiltonians for QAOA, overcoming the limitation of fixed-depth quantum circuits and significantly improving approximation ratios.
Metric Transforms and Low Rank Representations of Kernels for Fast Attention
·275 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ University of California, Berkeley
Researchers unveil novel linear-algebraic tools revealing the limits of fast attention, classifying positive definite kernels for Manhattan distance, and fully characterizing metric transforms for Man…
Metric Space Magnitude for Evaluating the Diversity of Latent Representations
·6876 words·33 mins·
loading
·
loading
AI Generated
AI Theory
Representation Learning
π’ University of Edinburgh
Novel metric space magnitude measures rigorously quantify the diversity of latent representations across multiple scales, showing superior performance in detecting mode collapse and characterizing emb…
Mechanism design augmented with output advice
·374 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Aristotle University of Thessaloniki
Mechanism design enhanced with output advice improves approximation guarantees by using imperfect predictions of the output, not agent types, offering robust, practical solutions.
Measuring Goal-Directedness
·1615 words·8 mins·
loading
·
loading
AI Theory
Ethics
π’ Imperial College London
New metric, Maximum Entropy Goal-Directedness (MEG), quantifies AI goal-directedness, crucial for assessing AI safety and agency.
Mean-Field Langevin Dynamics for Signed Measures via a Bilevel Approach
·350 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Γcole Polytechnique FΓ©dΓ©rale De Lausanne
This paper presents a novel bilevel approach to extend mean-field Langevin dynamics to solve convex optimization problems over signed measures, achieving stronger guarantees and faster convergence rat…
Mean-Field Analysis for Learning Subspace-Sparse Polynomials with Gaussian Input
·272 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ MIT
Researchers establish basis-free conditions for SGD learnability in two-layer neural networks learning subspace-sparse polynomials with Gaussian input, offering insights into training dynamics.
Maximizing utility in multi-agent environments by anticipating the behavior of other learners
·1732 words·9 mins·
loading
·
loading
AI Theory
Optimization
π’ MIT
Optimizing against learning agents: New algorithms and computational limits revealed!
MatrixNet: Learning over symmetry groups using learned group representations
·1841 words·9 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Northeastern University
MatrixNet learns efficient group representations for improved deep learning on symmetry groups, achieving higher sample efficiency and generalization than existing methods.
Matrix Denoising with Doubly Heteroscedastic Noise: Fundamental Limits and Optimal Spectral Methods
·1782 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Institute of Science and Technology Austria
Optimal matrix denoising with doubly heteroscedastic noise achieved!
Matching the Statistical Query Lower Bound for $k$-Sparse Parity Problems with Sign Stochastic Gradient Descent
·2323 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ UC Los Angeles
Sign Stochastic Gradient Descent (SGD) achieves optimal sample complexity for solving k-sparse parity problems, matching Statistical Query lower bounds.
Masked Hard-Attention Transformers Recognize Exactly the Star-Free Languages
·1697 words·8 mins·
loading
·
loading
AI Generated
Natural Language Processing
AI Theory
π’ University of Notre Dame
Masked hard-attention transformers, with strict masking, precisely capture star-free languages, matching the expressive power of linear temporal logic.
Marrying Causal Representation Learning with Dynamical Systems for Science
·3100 words·15 mins·
loading
·
loading
AI Generated
AI Theory
Representation Learning
π’ Institute of Science and Technology Austria
This study marries causal representation learning with dynamical systems to enable parameter identification in real-world scientific data, unlocking downstream causal analysis for various applications…
Markov Equivalence and Consistency in Differentiable Structure Learning
·2350 words·12 mins·
loading
·
loading
AI Theory
Causality
π’ Carnegie Mellon University
Researchers developed a new, differentiable score function for learning causal relationships from data that reliably recovers the simplest causal model, even with complex data.
Marginal Causal Flows for Validation and Inference
·1827 words·9 mins·
loading
·
loading
AI Theory
Causality
π’ University of Oxford
Frugal Flows: Generate realistic causal benchmarks with exact marginal causal effects, enabling robust causal method validation.
MambaLRP: Explaining Selective State Space Sequence Models
·3148 words·15 mins·
loading
·
loading
AI Theory
Interpretability
π’ Google DeepMind
MambaLRP enhances explainability of Mamba sequence models by ensuring faithful relevance propagation, achieving state-of-the-art explanation performance, and uncovering model biases.
MALT Powers Up Adversarial Attacks
·1855 words·9 mins·
loading
·
loading
AI Theory
Robustness
π’ Weizmann Institute of Science
MALT: a novel adversarial attack, is 5x faster than AutoAttack, achieving higher success rates on CIFAR-100 and ImageNet by exploiting mesoscopic almost linearity in neural networks.
MAC Advice for facility location mechanism design
·1881 words·9 mins·
loading
·
loading
AI Theory
Robustness
π’ Tel Aviv University
Improved facility location mechanisms are designed using ‘Mostly Approximately Correct’ predictions, exceeding prior bounds despite large prediction errors.
Low Degree Hardness for Broadcasting on Trees
·1545 words·8 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Missouri
Low-degree polynomials fail to efficiently infer roots in broadcasting tree problems below the Kesten-Stigum bound.