🏢 Meta AI
Temporal Graph Neural Tangent Kernel with Graphon-Guaranteed
·1777 words·9 mins·
loading
·
loading
AI Theory
Representation Learning
🏢 Meta AI
Temp-G³NTK: a novel temporal graph neural tangent kernel guarantees convergence to graphon NTK, offering superior performance in temporal graph classification and node-level tasks.
Stochastic Optimal Control Matching
·1801 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Meta AI
Stochastic Optimal Control Matching (SOCM) significantly reduces errors in stochastic optimal control by learning a matching vector field using a novel iterative diffusion optimization technique.
Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
·4652 words·22 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Meta AI
Rainbow Teaming: a novel black-box approach generates diverse adversarial prompts to enhance LLM robustness and safety, achieving over 90% attack success rate across various models.
No Free Delivery Service: Epistemic limits of passive data collection in complex social systems
·2178 words·11 mins·
loading
·
loading
AI Theory
Generalization
🏢 Meta AI
Passive data collection in complex social systems invalidates standard AI model validation; new methods are needed.
Meta 3D AssetGen: Text-to-Mesh Generation with High-Quality Geometry, Texture, and PBR Materials
·2436 words·12 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Meta AI
Meta 3D AssetGen: High-quality text-to-mesh generation with realistic PBR materials and lighting, exceeding prior methods in speed and accuracy.
Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Length
·1897 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Meta AI
MEGALODON: A new neural architecture for LLMs, enabling unlimited context length with improved efficiency and accuracy.
Learning to grok: Emergence of in-context learning and skill composition in modular arithmetic tasks
·3569 words·17 mins·
loading
·
loading
Large Language Models
🏢 Meta AI
Large language models surprisingly solve unseen arithmetic tasks; this work reveals how they learn to compose simple skills into complex ones through in-context learning, showing a transition from mem…
Iteration Head: A Mechanistic Study of Chain-of-Thought
·2483 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Meta AI
Researchers reveal how Chain-of-Thought reasoning emerges in transformers via specialized ‘iteration heads’, improving LLM performance and offering insights into mechanistic interpretability.
Global Lyapunov functions: a long-standing open problem in mathematics, with symbolic transformers
·2454 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Meta AI
AI-powered sequence-to-sequence transformers surpass human and algorithmic abilities in discovering Lyapunov functions for dynamical systems, solving a long-standing open problem in mathematics.
FlowLLM: Flow Matching for Material Generation with Large Language Models as Base Distributions
·2004 words·10 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Meta AI
FlowLLM revolutionizes material design by cleverly merging large language models and Riemannian flow matching, yielding a 300% boost in stable material generation!
CALVIN: Improved Contextual Video Captioning via Instruction Tuning
·2746 words·13 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 Meta AI
CALVIN: Instruction tuning boosts contextual video captioning, achieving state-of-the-art results!
A Polar coordinate system represents syntax in large language models
·1633 words·8 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Meta AI
LLMs spontaneously encode syntax using a polar coordinate system, representing syntactic relations via relative direction and distance of word embeddings.