Skip to main content

🏢 Courant Institute

Sketchy Moment Matching: Toward Fast and Provable Data Selection for Finetuning
·1518 words·8 mins· loading · loading
AI Theory Generalization 🏢 Courant Institute
Sketchy Moment Matching (SkMM) is a fast and theoretically sound data selection method for deep learning finetuning. By controlling variance-bias tradeoffs in high dimensions, SkMM drastically reduces…
Realizable $H$-Consistent and Bayes-Consistent Loss Functions for Learning to Defer
·1495 words·8 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Courant Institute
New surrogate loss functions for learning-to-defer achieve Bayes-consistency, realizable H-consistency, and H-consistency bounds simultaneously, resolving open questions and improving L2D performance.
Multi-Label Learning with Stronger Consistency Guarantees
·239 words·2 mins· loading · loading
Machine Learning Optimization 🏢 Courant Institute
Novel surrogate losses with label-independent H-consistency bounds enable stronger guarantees for multi-label learning.
Mixed Dynamics In Linear Networks: Unifying the Lazy and Active Regimes
·521 words·3 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Courant Institute
A new formula unifies lazy and active neural network training regimes, revealing a mixed regime that combines their strengths for faster convergence and low-rank bias.
A Universal Growth Rate for Learning with Smooth Surrogate Losses
·1364 words·7 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Courant Institute
This paper reveals a universal square-root growth rate for H-consistency bounds of smooth surrogate losses in classification, significantly advancing our understanding of loss function selection.