🏢 Apple
Progressive Entropic Optimal Transport Solvers
·4169 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
🏢 Apple
Progressive Entropic Optimal Transport (PROGOT) solvers efficiently and robustly compute optimal transport plans and maps, even at large scales, by progressively scheduling parameters.
Private Stochastic Convex Optimization with Heavy Tails: Near-Optimality from Simple Reductions
·397 words·2 mins·
loading
·
loading
AI Theory
Privacy
🏢 Apple
Achieving near-optimal rates for differentially private stochastic convex optimization with heavy-tailed gradients is possible using simple reduction-based techniques.
Private Online Learning via Lazy Algorithms
·475 words·3 mins·
loading
·
loading
AI Generated
AI Theory
Privacy
🏢 Apple
New transformation boosts privacy in online learning!
Learning Elastic Costs to Shape Monge Displacements
·1828 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Apple
Learn optimal transport maps with structured displacements using elastic costs and a novel bilevel loss function!
Kaleido Diffusion: Improving Conditional Diffusion Models with Autoregressive Latent Modeling
·2911 words·14 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Apple
Kaleido Diffusion boosts the diversity of images generated by diffusion models without sacrificing quality, using autoregressive latent modeling to add more control and interpretability to the image g…
Instance-Optimal Private Density Estimation in the Wasserstein Distance
·338 words·2 mins·
loading
·
loading
AI Theory
Privacy
🏢 Apple
Instance-optimal private density estimation algorithms, adapting to data characteristics for improved accuracy in the Wasserstein distance, are introduced.
How Far Can Transformers Reason? The Globality Barrier and Inductive Scratchpad
·3573 words·17 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Apple
Transformers struggle with complex reasoning tasks. This paper introduces ‘globality degree’ to measure task difficulty and shows that high globality hinders efficient learning. However, using ‘induc…
Grounding Multimodal Large Language Models in Actions
·3629 words·18 mins·
loading
·
loading
AI Generated
Multimodal Learning
Embodied AI
🏢 Apple
Researchers unveil unified architecture for grounding multimodal large language models in actions, showing superior performance with learned tokenization for continuous actions and semantic alignment …
Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum
·3234 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Apple
This paper introduces dataset decomposition (DD), a novel approach to accelerate LLM training while enhancing performance. DD significantly reduces training time by decomposing datasets into buckets …
Bridging semantics and pragmatics in information-theoretic emergent communication
·1593 words·8 mins·
loading
·
loading
Natural Language Processing
Dialogue Systems
🏢 Apple
AI agents learn human-like communication, combining semantic categorization and pragmatic context-sensitive reasoning, through a novel information-theoretic framework.
Aggregate-and-Adapt Natural Language Prompts for Downstream Generalization of CLIP
·2881 words·14 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Apple
Aggregate-and-Adapt Prompt Embedding (AAPE) boosts CLIP’s downstream generalization by distilling textual knowledge from natural language prompts, achieving competitive performance across various visi…