Large Language Models
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails
·2622 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of California, Los Angeles
DuoGuard: a novel two-player RL framework generates high-quality synthetic data, improving multilingual LLM safety by outperforming state-of-the-art models with a significantly smaller model size and …
Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions
·6016 words·29 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Brown University
Simple interactions can easily elicit harmful outputs from LLMs, which are often overlooked. The SPEAK EASY framework and HARMSCORE metric expose this vulnerability and provide tools for better safet…
PILAF: Optimal Human Preference Sampling for Reward Modeling
·2374 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 NYU
PILAF optimizes human feedback in reward modeling for better LLM alignment by using a novel response sampling strategy that aligns reward modeling with value optimization.
MAGA: MAssive Genre-Audience Reformulation to Pretraining Corpus Expansion
·2840 words·14 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 ByteDance
MAGA reformulates existing corpora to massively expand LLM pretraining data, boosting performance across various model sizes while maintaining quality.
Linear Correlation in LM's Compositional Generalization and Hallucination
·8299 words·39 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 UC San Diego
Language models surprisingly exhibit linear relationships when composing knowledge; this linearity, resilient to fine-tuning, predicts compositional generalization and hallucination.
CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference
·1697 words·8 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Chinese University of Hong Kong
CMOE efficiently transforms dense LLMs into sparse MoE architectures via expert carving, enabling fast inference without extensive retraining.
BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
·2217 words·11 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Salesforce AI Research
BOLT bootstraps Long Chain-of-Thought reasoning in LLMs without distillation, achieving impressive results across various benchmarks.
Beyond Prompt Content: Enhancing LLM Performance via Content-Format Integrated Prompt Optimization
·2451 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Microsoft Research
Researchers jointly optimize prompt content and format to significantly boost Large Language Model (LLM) performance.
Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning
·3144 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Meta AI
Boosting language model reasoning: A novel hybrid approach using latent tokens drastically shortens reasoning traces, improving model performance and efficiency.
Teaching Language Models to Critique via Reinforcement Learning
·4328 words·21 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of Hong Kong
LLMs learn to critique and refine their output via reinforcement learning, significantly improving code generation.
LIMO: Less is More for Reasoning
·2691 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Generative Al Research
LIMO: Few examples unlock complex reasoning in LLMs, challenging assumptions about data-hungry models and achieving state-of-the-art results with minimal training.
Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2
·4637 words·22 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Google DeepMind
AlphaGeometry2 surpasses average IMO gold medalists in solving geometry problems!
Analyze Feature Flow to Enhance Interpretation and Steering in Language Models
·5882 words·28 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 T-Tech
Researchers unveil a data-free method to visualize and control feature flow in LLMs, enhancing interpretability and enabling targeted model steering.
Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search
·3854 words·19 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 MIT
Satori: A novel 7B LLM achieves state-of-the-art mathematical reasoning via autoregressive search.
QLASS: Boosting Language Agent Inference via Q-Guided Stepwise Search
·2983 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 UC Los Angeles
QLASS boosts language agent inference by using Q-values to guide a stepwise search, improving efficiency and performance even with limited data.
On Teacher Hacking in Language Model Distillation
·2783 words·14 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Google DeepMind
Language model distillation suffers from ’teacher hacking’, where student models over-optimize flawed teacher models, degrading true performance. This paper identifies this issue and offers effective…
ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning
·2452 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 University of Washington
LLMs struggle with complex logical reasoning; ZebraLogic benchmark reveals a ‘curse of complexity’, highlighting inherent limitations and guiding future research.
The Differences Between Direct Alignment Algorithms are a Blur
·3273 words·16 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 T-Tech
Direct alignment algorithms are a blur, but this paper shows how a simple SFT phase and a scaling parameter significantly improve alignment quality, regardless of the specific reward function used.
Process Reinforcement through Implicit Rewards
·3889 words·19 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Tsinghua University
PRIME (Process Reinforcement through IMplicit rEwards) revolutionizes LLM training by efficiently using implicit process rewards from online policy rollouts and outcome labels, significantly boosting …
PlotGen: Multi-Agent LLM-based Scientific Data Visualization via Multimodal Feedback
·523 words·3 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 IGDTUW, Delhi
PlotGen: A novel multi-agent LLM framework automates accurate scientific data visualization via multimodal feedback, boosting novice productivity and improving visualization accuracy.