Skip to main content

Natural Language Processing

LIMO: Less is More for Reasoning
·2691 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Generative Al Research
LIMO: Few examples unlock complex reasoning in LLMs, challenging assumptions about data-hungry models and achieving state-of-the-art results with minimal training.
Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2
·4637 words·22 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Google DeepMind
AlphaGeometry2 surpasses average IMO gold medalists in solving geometry problems!
DreamDPO: Aligning Text-to-3D Generation with Human Preferences via Direct Preference Optimization
·2709 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Text Generation 🏢 Zhejiang University
DreamDPO: Revolutionizing text-to-3D generation by directly aligning outputs with human preferences via innovative preference optimization.
Analyze Feature Flow to Enhance Interpretation and Steering in Language Models
·5882 words·28 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 T-Tech
Researchers unveil a data-free method to visualize and control feature flow in LLMs, enhancing interpretability and enabling targeted model steering.
Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search
·3854 words·19 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 MIT
Satori: A novel 7B LLM achieves state-of-the-art mathematical reasoning via autoregressive search.
QLASS: Boosting Language Agent Inference via Q-Guided Stepwise Search
·2983 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 UC Los Angeles
QLASS boosts language agent inference by using Q-values to guide a stepwise search, improving efficiency and performance even with limited data.
On Teacher Hacking in Language Model Distillation
·2783 words·14 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Google DeepMind
Language model distillation suffers from ’teacher hacking’, where student models over-optimize flawed teacher models, degrading true performance. This paper identifies this issue and offers effective…
ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning
·2452 words·12 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 University of Washington
LLMs struggle with complex logical reasoning; ZebraLogic benchmark reveals a ‘curse of complexity’, highlighting inherent limitations and guiding future research.
The Differences Between Direct Alignment Algorithms are a Blur
·3273 words·16 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 T-Tech
Direct alignment algorithms are a blur, but this paper shows how a simple SFT phase and a scaling parameter significantly improve alignment quality, regardless of the specific reward function used.
Process Reinforcement through Implicit Rewards
·3889 words·19 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Tsinghua University
PRIME (Process Reinforcement through IMplicit rEwards) revolutionizes LLM training by efficiently using implicit process rewards from online policy rollouts and outcome labels, significantly boosting …
PlotGen: Multi-Agent LLM-based Scientific Data Visualization via Multimodal Feedback
·523 words·3 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 IGDTUW, Delhi
PlotGen: A novel multi-agent LLM framework automates accurate scientific data visualization via multimodal feedback, boosting novice productivity and improving visualization accuracy.
PhD Knowledge Not Required: A Reasoning Challenge for Large Language Models
·1257 words·6 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Wellesley College
New benchmark challenges LLMs with general knowledge puzzles, revealing reasoning gaps and suggesting improvements for future models.
Lifelong Sequential Knowledge Editing without Model Degradation
·13067 words·62 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 UC Berkeley
ENCORE enables lifelong sequential knowledge editing in LLMs without performance loss, achieving 10,000 edits while maintaining downstream accuracy.
Jailbreaking with Universal Multi-Prompts
·5963 words·28 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 National Taiwan University
JUMP outperforms existing methods by optimizing universal multi-prompts for jailbreaking LLMs, offering a more efficient and generalizable approach to LLM adversarial attacks.
FastKV: KV Cache Compression for Fast Long-Context Processing with Token-Selective Propagation
·4575 words·22 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Department of Electrical and Computer Engineering, Seoul National University
FastKV: A novel KV cache compression method speeds up long-context LLM processing 2x by selectively propagating tokens and using GQA-aware compression, maintaining accuracy.
DeepRAG: Thinking to Retrieval Step by Step for Large Language Models
·2973 words·14 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Question Answering 🏢 Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences
DeepRAG enhances LLM reasoning by strategically integrating retrieval, modeled as an MDP, improving accuracy by 21.99% and retrieval efficiency.
ChartCitor: Multi-Agent Framework for Fine-Grained Chart Visual Attribution
·228 words·2 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Question Answering 🏢 Adobe Research
ChartCitor: A multi-agent LLM framework combats LLM hallucination in ChartQA by providing fine-grained visual citations, enhancing user trust and productivity.
Almost Surely Safe Alignment of Large Language Models at Inference-Time
·2605 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Peking University
InferenceGuard ensures almost-sure safe LLM responses at inference time by framing safe generation as a constrained Markov Decision Process in the LLM’s latent space, achieving high safety rates witho…
A Probabilistic Inference Approach to Inference-Time Scaling of LLMs using Particle-Based Monte Carlo Methods
·2572 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 MIT
Boosting Large Language Model (LLM) inference speed using probabilistic inference via particle-based Monte Carlo methods achieves 4-16x better scaling than deterministic search approaches.
WILDCHAT-50M: A Deep Dive Into the Role of Synthetic Data in Post-Training
·3471 words·17 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 NYU
WILDCHAT-50M: Largest public chat dataset refines LLM post-training, showing superior SFT performance with fewer samples.