Natural Language Processing
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
·3884 words·19 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Tsinghua University
Smaller LLMs can outperform larger ones by strategically increasing computation during inference, defying conventional LLM scaling.
Training Language Models for Social Deduction with Multi-Agent Reinforcement Learning
·507 words·3 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Stanford University
Language models learn effective social deduction strategies in a virtual game by using their goal to predict useful information as a dense reward signal, doubling win rates compared to standard RL.
The Curse of Depth in Large Language Models
·2429 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Medical Artificial Intelligence Laboratory, Westlake University
Deep layers in LLMs underperform due to Pre-Layer Normalization; LayerNorm Scaling resolves this by controlling output variance, significantly improving training efficiency.
LM2: Large Memory Models
·2722 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Convergence Labs Ltd
LM2: Large Memory Models enhance Transformers by adding an auxiliary memory module, significantly improving multi-step reasoning and long-context information synthesis.
APE: Faster and Longer Context-Augmented Generation via Adaptive Parallel Encoding
·6090 words·29 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
APE: a novel method significantly speeds up context-augmented generation (CAG). By using adaptive parallel encoding, APE achieves a 4.5x speedup and maintains high accuracy even with 128K length cont…
Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach
·5939 words·28 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of Maryland
Boost LLM reasoning power at test time by recursively processing latent information, enabling dramatic performance gains with fewer parameters.
QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
·3320 words·16 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 ISTA
QuEST enables stable, accurate LLM training using only 1-bit weights and activations, achieving Pareto-optimal performance compared to higher-precision models.
Generating Symbolic World Models via Test-time Scaling of Large Language Models
·2722 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Hong Kong University of Science and Technology
LLMs excel at complex reasoning but struggle with planning; this paper introduces a test-time scaling approach that enhances LLMs’ PDDL reasoning, enabling high-quality PDDL domain generation, outperf…
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails
·2622 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of California, Los Angeles
DuoGuard: a novel two-player RL framework generates high-quality synthetic data, improving multilingual LLM safety by outperforming state-of-the-art models with a significantly smaller model size and …
ARR: Question Answering with Large Language Models via Analyzing, Retrieving, and Reasoning
·8117 words·39 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Question Answering
🏢 University of British Columbia
ARR: A novel zero-shot prompting method significantly boosts LLM performance on diverse question-answering tasks by explicitly incorporating question analysis, information retrieval, and step-by-step …
Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions
·6016 words·29 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Brown University
Simple interactions can easily elicit harmful outputs from LLMs, which are often overlooked. The SPEAK EASY framework and HARMSCORE metric expose this vulnerability and provide tools for better safet…
PILAF: Optimal Human Preference Sampling for Reward Modeling
·2374 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 NYU
PILAF optimizes human feedback in reward modeling for better LLM alignment by using a novel response sampling strategy that aligns reward modeling with value optimization.
MAGA: MAssive Genre-Audience Reformulation to Pretraining Corpus Expansion
·2840 words·14 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 ByteDance
MAGA reformulates existing corpora to massively expand LLM pretraining data, boosting performance across various model sizes while maintaining quality.
Llasa: Scaling Train-Time and Inference-Time Compute for Llama-based Speech Synthesis
·3315 words·16 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Text Generation
🏢 Hong Kong University of Science and Technology
Llasa, a novel single-Transformer TTS model, achieves state-of-the-art performance by scaling both training and inference compute, improving naturalness, prosody and emotional expressiveness.
Linear Correlation in LM's Compositional Generalization and Hallucination
·8299 words·39 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 UC San Diego
Language models surprisingly exhibit linear relationships when composing knowledge; this linearity, resilient to fine-tuning, predicts compositional generalization and hallucination.
CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference
·1697 words·8 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Chinese University of Hong Kong
CMOE efficiently transforms dense LLMs into sparse MoE architectures via expert carving, enabling fast inference without extensive retraining.
BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
·2217 words·11 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Salesforce AI Research
BOLT bootstraps Long Chain-of-Thought reasoning in LLMs without distillation, achieving impressive results across various benchmarks.
Beyond Prompt Content: Enhancing LLM Performance via Content-Format Integrated Prompt Optimization
·2451 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Microsoft Research
Researchers jointly optimize prompt content and format to significantly boost Large Language Model (LLM) performance.
Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning
·3144 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Meta AI
Boosting language model reasoning: A novel hybrid approach using latent tokens drastically shortens reasoning traces, improving model performance and efficiency.
Teaching Language Models to Critique via Reinforcement Learning
·4328 words·21 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of Hong Kong
LLMs learn to critique and refine their output via reinforcement learning, significantly improving code generation.