🏢 Courant Institute
Sketchy Moment Matching: Toward Fast and Provable Data Selection for Finetuning
·1518 words·8 mins·
loading
·
loading
AI Theory
Generalization
🏢 Courant Institute
Sketchy Moment Matching (SkMM) is a fast and theoretically sound data selection method for deep learning finetuning. By controlling variance-bias tradeoffs in high dimensions, SkMM drastically reduces…
Realizable $H$-Consistent and Bayes-Consistent Loss Functions for Learning to Defer
·1495 words·8 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Courant Institute
New surrogate loss functions for learning-to-defer achieve Bayes-consistency, realizable H-consistency, and H-consistency bounds simultaneously, resolving open questions and improving L2D performance.
Multi-Label Learning with Stronger Consistency Guarantees
·239 words·2 mins·
loading
·
loading
Machine Learning
Optimization
🏢 Courant Institute
Novel surrogate losses with label-independent H-consistency bounds enable stronger guarantees for multi-label learning.
Mixed Dynamics In Linear Networks: Unifying the Lazy and Active Regimes
·521 words·3 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Courant Institute
A new formula unifies lazy and active neural network training regimes, revealing a mixed regime that combines their strengths for faster convergence and low-rank bias.
A Universal Growth Rate for Learning with Smooth Surrogate Losses
·1364 words·7 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Courant Institute
This paper reveals a universal square-root growth rate for H-consistency bounds of smooth surrogate losses in classification, significantly advancing our understanding of loss function selection.