AI Theory
Expectation Alignment: Handling Reward Misspecification in the Presence of Expectation Mismatch
·527 words·3 mins·
loading
·
loading
AI Generated
AI Theory
Safety
🏢 Colorado State University
This paper introduces Expectation Alignment (EAL), a novel framework and interactive algorithm to address reward misspecification in AI, aligning AI behavior with user expectations.
Exogenous Matching: Learning Good Proposals for Tractable Counterfactual Estimation
·2755 words·13 mins·
loading
·
loading
AI Theory
Causality
🏢 Shanghai Key Laboratory of Trustworthy Computing, East China Normal University
Exogenous Matching learns optimal proposals for efficient counterfactual estimation by transforming variance minimization into conditional distribution learning, outperforming existing methods.
Exactly Minimax-Optimal Locally Differentially Private Sampling
·1615 words·8 mins·
loading
·
loading
AI Theory
Privacy
🏢 KAIST
This paper provides the first exact minimax-optimal mechanisms for locally differentially private sampling, applicable across all f-divergences.
Exact, Tractable Gauss-Newton Optimization in Deep Reversible Architectures Reveal Poor Generalization
·2136 words·11 mins·
loading
·
loading
AI Theory
Optimization
🏢 MediaTek Research
Exact Gauss-Newton optimization in deep reversible networks surprisingly reveals poor generalization, despite faster training, challenging existing deep learning optimization theories.
Exact Gradients for Stochastic Spiking Neural Networks Driven by Rough Signals
·528 words·3 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 University of Copenhagen
New framework uses rough path theory to enable gradient-based training of SSNNs driven by rough signals, allowing for noise in spike timing and network dynamics.
Evidence of Learned Look-Ahead in a Chess-Playing Neural Network
·4142 words·20 mins·
loading
·
loading
AI Generated
AI Theory
Interpretability
🏢 UC Berkeley
Chess AI Leela Zero surprisingly uses learned look-ahead, internally representing future optimal moves, significantly improving its strategic decision-making.
Evaluating alignment between humans and neural network representations in image-based learning tasks
·3856 words·19 mins·
loading
·
loading
AI Generated
AI Theory
Representation Learning
🏢 Helmholtz Computational Health Center
Pretrained neural networks surprisingly capture fundamental aspects of human cognition, enabling generalization in image-based learning tasks, as demonstrated by aligning neural network representation…
Estimating Heterogeneous Treatment Effects by Combining Weak Instruments and Observational Data
·1848 words·9 mins·
loading
·
loading
AI Theory
Causality
🏢 Cornell University
This study develops a novel two-stage framework for accurately predicting conditional average treatment effects using both observational data and weak instrumental variables, overcoming limitations of…
Estimating Generalization Performance Along the Trajectory of Proximal SGD in Robust Regression
·1899 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Rutgers University
New consistent estimators precisely track generalization error during robust regression’s iterative model training, enabling optimal stopping iteration for minimized error.
Entrywise error bounds for low-rank approximations of kernel matrices
·1461 words·7 mins·
loading
·
loading
AI Theory
Optimization
🏢 Imperial College London
This paper provides novel entrywise error bounds for low-rank kernel matrix approximations, showing how many data points are needed to get statistically consistent results for low-rank approximations.
Entropy testing and its application to testing Bayesian networks
·328 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Sydney
This paper presents near-optimal algorithms for entropy identity testing, significantly improving Bayesian network testing efficiency.
Enriching Disentanglement: From Logical Definitions to Quantitative Metrics
·3435 words·17 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 University of Tokyo
This paper presents a novel approach to deriving theoretically grounded disentanglement metrics by linking logical definitions to quantitative measures, offering strong theoretical guarantees and easi…
Enhancing Robustness of Last Layer Two-Stage Fair Model Corrections
·2233 words·11 mins·
loading
·
loading
AI Theory
Fairness
🏢 Arizona State University
Boosting fair machine learning’s robustness against noisy labels, this work introduces a novel label-spreading method, achieving state-of-the-art worst-group accuracy.
Enhancing Robustness of Graph Neural Networks on Social Media with Explainable Inverse Reinforcement Learning
·1977 words·10 mins·
loading
·
loading
AI Theory
Robustness
🏢 Hong Kong University of Science and Technology
MoE-BiEntIRL: A novel explainable inverse reinforcement learning method enhances GNN robustness against diverse social media attacks by reconstructing attacker policies and generating more robust trai…
Energy-based Epistemic Uncertainty for Graph Neural Networks
·4139 words·20 mins·
loading
·
loading
AI Theory
Robustness
🏢 Technical University of Munich
GEBM: a novel graph-based energy model for robust GNN uncertainty estimation.
Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space
·3166 words·15 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Harvard University
Generative models learn hidden capabilities suddenly during training, which can be explained and predicted using a novel ‘concept space’ framework that analyzes learning dynamics and concept signal.
Emergence of heavy tails in homogenized stochastic gradient descent
·1472 words·7 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Northwestern Polytechnical University
Homogenized SGD reveals heavy-tailed neural network parameters, offering quantifiable bounds on tail-index and showcasing the interplay between optimization hyperparameters and model generalization.
Elliptical Attention
·3508 words·17 mins·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 FPT Software AI Center
Elliptical Attention enhances transformers by using a Mahalanobis distance metric, stretching the feature space to focus on contextually relevant information, thus improving robustness and reducing re…
Efficiently Learning Significant Fourier Feature Pairs for Statistical Independence Testing
·1689 words·8 mins·
loading
·
loading
AI Theory
Causality
🏢 Shanghai Key Lab of Intelligent Information Processing, and School of Computer Science, Fudan University
This research introduces LFHSIC, a novel, linear-time independence test that significantly outperforms existing methods, especially for high-dimensional data, by learning optimal Fourier feature pairs…
Efficient Streaming Algorithms for Graphlet Sampling
·1741 words·9 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Saarland University
STREAM-UGS: a novel semi-streaming algorithm for efficient graphlet sampling, enabling fast analysis of massive graphs with limited memory.