Skip to main content

Posters

2024

On the Necessity of Collaboration for Online Model Selection with Decentralized Data
·1427 words·7 mins· loading · loading
Machine Learning Federated Learning 🏒 School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen
Federated online model selection needs collaboration only when clients have limited computing power; otherwise, independent learning suffices.
On the Minimax Regret for Contextual Linear Bandits and Multi-Armed Bandits with Expert Advice
·360 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Tokyo
This paper provides novel algorithms and matching lower bounds for multi-armed bandits with expert advice and contextual linear bandits, resolving open questions and advancing theoretical understandin…
On the Limitations of Fractal Dimension as a Measure of Generalization
·1917 words·9 mins· loading · loading
Machine Learning Deep Learning 🏒 University of Oxford
Fractal dimension, while showing promise, fails to consistently predict neural network generalization due to hyperparameter influence and adversarial initializations; prompting further research.
On the Inductive Bias of Stacking Towards Improving Reasoning
·2018 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Google Research
MIDAS: A novel training method improves language model reasoning by efficiently stacking middle layers, surprisingly boosting downstream task performance without increasing pretraining perplexity.
On the Impacts of the Random Initialization in the Neural Tangent Kernel Theory
·1555 words·8 mins· loading · loading
AI Theory Generalization 🏒 Tsinghua University
Standard initialization in neural networks negatively impacts generalization ability under Neural Tangent Kernel theory, contradicting real-world performance, urging the development of improved theore…
On the Impact of Feature Heterophily on Link Prediction with Graph Neural Networks
·2063 words·10 mins· loading · loading
AI Theory Representation Learning 🏒 University of Michigan
Graph Neural Networks (GNNs) struggle with heterophilic link prediction; this paper introduces formal definitions, theoretical analysis, improved designs, and real-world benchmarks to address this cha…
On the Identifiability of Hybrid Deep Generative Models: Meta-Learning as a Solution
·1881 words·9 mins· loading · loading
Machine Learning Meta Learning 🏒 Rochester Institute of Technology
Meta-learning solves hybrid deep generative model unidentifiability!
On the Expressivity and Sample Complexity of Node-Individualized Graph Neural Networks
·2191 words·11 mins· loading · loading
AI Generated AI Theory Generalization 🏒 Max Planck Institute of Biochemistry
Boosting GNN expressivity and generalization: Novel node individualization schemes lower sample complexity, improving substructure identification.
On the Expressive Power of Tree-Structured Probabilistic Circuits
·1425 words·7 mins· loading · loading
AI Theory Optimization 🏒 University of Illinois Urbana-Champaign
Tree-structured probabilistic circuits are surprisingly efficient: this paper proves a quasi-polynomial upper bound on their size, showing they’re almost as expressive as more complex DAG structures.
On the Efficiency of ERM in Feature Learning
·316 words·2 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Toronto
ERM’s efficiency in feature learning surprisingly remains high even with massive feature maps; its excess risk asymptotically matches an oracle procedure’s, implying potential for streamlined feature-…
On the Curses of Future and History in Future-dependent Value Functions for Off-policy Evaluation
·299 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Illinois Urbana-Champaign
This paper tackles the ‘curse of horizon’ in off-policy evaluation for partially observable Markov decision processes (POMDPs) by proposing novel coverage assumptions, enabling polynomial estimation e…
On the Convergence of Loss and Uncertainty-based Active Learning Algorithms
·2336 words·11 mins· loading · loading
AI Generated Machine Learning Active Learning 🏒 Meta
New active learning algorithm, Adaptive-Weight Sampling (AWS), achieves faster convergence with theoretical guarantees, improving data efficiency for machine learning.
On the Computational Landscape of Replicable Learning
·348 words·2 mins· loading · loading
AI Theory Optimization 🏒 Yale University
This paper reveals surprising computational connections between algorithmic replicability and other learning paradigms, offering novel algorithms and demonstrating separations between replicability an…
On the Computational Complexity of Private High-dimensional Model Selection
·1501 words·8 mins· loading · loading
AI Theory Privacy 🏒 University of Michigan
This paper proposes a computationally efficient, differentially private best subset selection method for high-dimensional sparse linear regression, achieving both strong statistical utility and provab…
On the Complexity of Teaching a Family of Linear Behavior Cloning Learners
·1819 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Washington
A novel algorithm, TIE, optimally teaches a family of linear behavior cloning learners, achieving instance-optimal teaching dimension while providing efficient approximation for larger action spaces.
On the Complexity of Learning Sparse Functions with Statistical and Gradient Queries
·484 words·3 mins· loading · loading
Machine Learning Deep Learning 🏒 Toyota Technological Institute at Chicago
Learning sparse functions efficiently with gradient methods is challenging; this paper introduces Differentiable Learning Queries (DLQ) to precisely characterize gradient query complexity, revealing s…
On the Complexity of Identification in Linear Structural Causal Models
·1403 words·7 mins· loading · loading
AI Theory Causality 🏒 Saarland University
New polynomial-space algorithm for causal parameter identification in linear models vastly improves upon existing methods, showing that this crucial task is computationally hard.
On the Comparison between Multi-modal and Single-modal Contrastive Learning
·455 words·3 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏒 RIKEN AIP
Multi-modal contrastive learning surpasses single-modal by leveraging inter-modal correlations to improve feature learning and downstream task performance, as demonstrated through a novel theoretical …
On the cohesion and separability of average-link for hierarchical agglomerative clustering
·1663 words·8 mins· loading · loading
AI Theory Optimization 🏒 Departmento De InformÑtica, PUC-RIO
Average-link hierarchical clustering gets a comprehensive evaluation using new criteria, showing it outperforms other methods when both cohesion and separability matter.
On the Benefits of Public Representations for Private Transfer Learning under Distribution Shift
·1927 words·10 mins· loading · loading
AI Generated AI Theory Privacy 🏒 Carnegie Mellon University
Public data boosts private AI accuracy even with extreme distribution shifts, improving private model training by up to 67% in three tasks. This is due to shared low-dimensional representations betwe…