Skip to main content

Natural Language Processing

Typed-RAG: Type-aware Multi-Aspect Decomposition for Non-Factoid Question Answering
·1842 words·9 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Question Answering 🏒 Pohang University of Science and Technology
Typed-RAG enhances non-factoid QA by type-aware decomposition, refining retrieval and generation for nuanced, user-aligned answers.
Survey on Evaluation of LLM-based Agents
·396 words·2 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Hebrew University of Jerusalem
A comprehensive survey on evaluation methodologies for LLM-based agents, analyzing benchmarks and frameworks across key dimensions like capabilities, applications, and generalist performance.
Stop Overthinking: A Survey on Efficient Reasoning for Large Language Models
·3774 words·18 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Rice University
LLMs survey: Model, output, and prompt-based strategies for efficient reasoning, mitigating ‘overthinking’ for faster, cheaper, and real-world applications.
MathFusion: Enhancing Mathematic Problem-solving of LLM through Instruction Fusion
·2769 words·13 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Renmin University of China
MathFusion: Instruction Fusion enhances LLM’s math problem-solving!
Deceptive Humor: A Synthetic Multilingual Benchmark Dataset for Bridging Fabricated Claims with Humorous Content
·3312 words·16 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Text Classification 🏒 IIIT Dharwad
New dataset bridges fabricated claims with humor for spotting online deception!
CaKE: Circuit-aware Editing Enables Generalizable Knowledge Learners
·3734 words·18 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 University of California, Los Angeles
CaKE: Editing LLMs to Enhance Knowledge Generalization Across Reasoning Tasks.
BigO(Bench) -- Can LLMs Generate Code with Controlled Time and Space Complexity?
·3082 words·15 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 FAIR at Meta
BIGO(Bench) can help LLMs generate code with controlled time/space complexity, addressing the gap in current evaluations and encouraging further exploration.
Temporal Consistency for LLM Reasoning Process Error Identification
·3234 words·16 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Princeton University
A new test-time method, Temporal Consistency, is introduced to improve LLM reasoning by leveraging iterative self-reflection.
Pensez: Less Data, Better Reasoning -- Rethinking French LLM
·3508 words·17 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 UniversitΓ© Grenoble Alpes
Pensez: Strategic fine-tuning beats massive data for superior reasoning in French LLMs, challenging conventional wisdom.
$Ο†$-Decoding: Adaptive Foresight Sampling for Balanced Inference-Time Exploration and Exploitation
·3341 words·16 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Shanghai AI Lab
Ξ¦-Decoding: Adaptive foresight sampling balances inference-time exploration and exploitation for better LLM reasoning.
Investigating Human-Aligned Large Language Model Uncertainty
·1326 words·7 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Vanderbilt University
This research explores how well LLM uncertainty measures align with human uncertainty, finding Bayesian and top-k entropy measures show promise.
GKG-LLM: A Unified Framework for Generalized Knowledge Graph Construction
·2910 words·14 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Information Extraction 🏒 Xi'an Jiaotong University
GKG-LLM: Unifying Knowledge Graph Construction with a novel 3-stage framework, empowering domain adaptation & resource efficiency.
New Trends for Modern Machine Translation with Large Reasoning Models
·518 words·3 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Machine Translation 🏒 University of Edinburgh
LRMs transform MT with reasoning, handling context, culture, and nuance for better translations.
Light-R1: Curriculum SFT, DPO and RL for Long COT from Scratch and Beyond
·1730 words·9 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Qiyuan Tech
Light-R1: Trains long COT models from scratch using curriculum SFT, DPO, and RL, achieving SOTA performance and strong generalization with limited resources.
Perplexity Trap: PLM-Based Retrievers Overrate Low Perplexity Documents
·3678 words·18 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Information Extraction 🏒 Renmin University of China
PLM retrievers overrate low-perplexity docs, causing source bias. This paper reveals the causal effect & offers a fix!
BiasEdit: Debiasing Stereotyped Language Models via Model Editing
·2942 words·14 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 University of California, San Diego
BIASEDIT: Efficiently debiasing language models via lightweight network edits!
SEAP: Training-free Sparse Expert Activation Pruning Unlock the Brainpower of Large Language Models
·3962 words·19 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 Renmin University of China
SEAP: Unlock LLM brainpower w/ training-free sparse expert activation pruning! Boost efficiency, keep accuracy. Optimize LLMs now!
WildIFEval: Instruction Following in the Wild
·2601 words·13 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Text Generation 🏒 Hebrew University of Jerusalem
WILDIFEVAL: Instruction Following in the Wild.
Beyond Decoder-only: Large Language Models Can be Good Encoders for Machine Translation
·6125 words·29 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Machine Translation 🏒 NLP Lab, Northeastern University, Shenyang, China
LLMs as MT encoders enhance efficiency & generalization!
WritingBench: A Comprehensive Benchmark for Generative Writing
·4038 words·19 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Text Generation 🏒 Alibaba Group
WritingBench: A new benchmark for generative writing evaluation, enhancing LLMs across diverse domains.