Skip to main content

🏢 Northwestern University

Soft-Label Integration for Robust Toxicity Classification
·2918 words·14 mins· loading · loading
AI Generated Natural Language Processing Text Classification 🏢 Northwestern University
Boosting toxicity classification robustness, this paper introduces a novel bi-level optimization framework integrating crowdsourced soft-labels and GroupDRO to enhance resistance against out-of-distri…
Semantic Feature Learning for Universal Unsupervised Cross-Domain Retrieval
·1809 words·9 mins· loading · loading
Computer Vision Cross-Modal Retrieval 🏢 Northwestern University
Universal Unsupervised Cross-Domain Retrieval (U2CDR) framework learns semantic features to enable accurate retrieval even when category spaces differ across domains.
Provably Optimal Memory Capacity for Modern Hopfield Models: Transformer-Compatible Dense Associative Memories as Spherical Codes
·1812 words·9 mins· loading · loading
AI Theory Representation Learning 🏢 Northwestern University
Researchers achieve provably optimal memory capacity in transformer-compatible Hopfield models by framing the problem as an optimal spherical code arrangement, resulting in a novel sublinear time algo…
Provably Mitigating Overoptimization in RLHF: Your SFT Loss is Implicitly an Adversarial Regularizer
·1760 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Northwestern University
RLHF’s overoptimization problem is mitigated by RPO, a novel algorithm that uses SFT loss as an implicit adversarial regularizer, ensuring efficient and effective LLM alignment.
On Statistical Rates and Provably Efficient Criteria of Latent Diffusion Transformers (DiTs)
·402 words·2 mins· loading · loading
AI Theory Generalization 🏢 Northwestern University
Latent Diffusion Transformers (DiTs) achieve almost-linear time training and inference through low-rank gradient approximations and efficient criteria, overcoming high dimensionality challenges.
Improving self-training under distribution shifts via anchored confidence with theoretical guarantees
·2507 words·12 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏢 Northwestern University
Anchored Confidence (AnCon) significantly improves self-training under distribution shifts by using a temporal ensemble to smooth noisy pseudo-labels, achieving 8-16% performance gains without computa…
Harnessing Multiple Correlated Networks for Exact Community Recovery
·2191 words·11 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Northwestern University
Unlocking latent community structures from multiple correlated networks is now possible with greater precision, as this research pinpoints the information-theoretic threshold for exact recovery, even …
Embedding Dimension of Contrastive Learning and $k$-Nearest Neighbors
·2134 words·11 mins· loading · loading
AI Generated Machine Learning Representation Learning 🏢 Northwestern University
Discover optimal embedding dimensions for contrastive learning & k-NN using graph arboricity; achieve efficient model design & performance.
Efficient Graph Matching for Correlated Stochastic Block Models
·2079 words·10 mins· loading · loading
AI Theory Optimization 🏢 Northwestern University
Efficient algorithm achieves near-perfect graph matching in correlated stochastic block models, resolving a key open problem and enabling improved community detection.