Skip to main content

Natural Language Processing

The Relationship Between Reasoning and Performance in Large Language Models -- o3 (mini) Thinks Harder, Not Longer
·2673 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Vrije Universiteit Brussel
LLMs: 03-mini achieves superior accuracy without longer reasoning chains, suggesting ’thinking harder’ matters more than ’thinking longer'.
LightThinker: Thinking Step-by-Step Compression
·1662 words·8 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph
LightThinker: LLMs dynamically compress intermediate steps, reducing memory & boosting reasoning efficiency without sacrificing accuracy.
Unstructured Evidence Attribution for Long Context Query Focused Summarization
·3830 words·18 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Text Summarization 🏢 University of Copenhagen
LLMs struggle with positional bias and lack transparency when summarizing long contexts. This paper introduces SUnsET dataset and fine-tuning methods to improve unstructured evidence citation and summ…
SurveyX: Academic Survey Automation via Large Language Models
·2720 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Renmin University of China
SURVEYX automates academic survey generation, enhancing content and citation quality.
StructFlowBench: A Structured Flow Benchmark for Multi-turn Instruction Following
·3134 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Dialogue Systems 🏢 School of Artificial Intelligence, Jilin University
Current LLM evaluation benchmarks often overlook the structural dependencies in multi-turn dialogues, treating them as simple concatenations of single-turn interactions. This approach neglects user in…
How Much Knowledge Can You Pack into a LoRA Adapter without Harming LLM?
·2645 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 AIRI
Packing new knowledge into LoRA adapters can harm LLMs! A delicate balance is needed to prevent performance decline.
FR-Spec: Accelerating Large-Vocabulary Language Models via Frequency-Ranked Speculative Sampling
·3445 words·17 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Text Generation 🏢 Tsinghua University
FR-Spec: Frequency-Ranked Speculative Sampling accelerates LLMs by optimizing vocabulary space compression, achieving 1.12x speedup over EAGLE-2.
Does Time Have Its Place? Temporal Heads: Where Language Models Recall Time-specific Information
·4876 words·23 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Korea University
LLMs have ‘Temporal Heads’ that process time-specific facts!
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
·3075 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Zhejiang University
LORAM: Train small, infer large LLMs by memory-efficient LoRA training. Enables 70B parameter model training on a 20G HBM GPU, replacing A100-80G. Reduces parameter storage cost by 15.81x.
SIFT: Grounding LLM Reasoning in Contexts via Stickers
·3144 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Shanghai Jiao Tong University
SIFT: Grounds LLM reasoning with ‘Stickers’ to highlight context and improve accuracy without extra training.
REFIND: Retrieval-Augmented Factuality Hallucination Detection in Large Language Models
·582 words·3 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Pohang University of Science and Technology
REFIND: Detects LLM hallucinations by directly leveraging retrieved documents, using a novel Context Sensitivity Ratio.
MoM: Linear Sequence Modeling with Mixture-of-Memories
·2764 words·13 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Shanghai AI Laboratory
MoM: Enhancing linear sequence modeling via mixture-of-memories for improved recall and reduced memory interference.
LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization
·2370 words·12 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 National University of Singapore
LongPO: Self-evolve LLMs to excel in long contexts via short-to-long preference optimization, boosting performance without sacrificing short-context skills.
Is That Your Final Answer? Test-Time Scaling Improves Selective Question Answering
·2478 words·12 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Question Answering 🏢 Johns Hopkins University
Test-time scaling + confidence = better QA!
Craw4LLM: Efficient Web Crawling for LLM Pretraining
·3024 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Tsinghua University
CRAW4LLM: Efficiently crawls web pages for LLM pretraining by prioritizing influence scores, boosting data quality & cutting crawling waste.
Autellix: An Efficient Serving Engine for LLM Agents as General Programs
·4705 words·23 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 UC Berkeley
Autellix: Efficient LLM Serving for Agents
Think Inside the JSON: Reinforcement Strategy for Strict LLM Schema Adherence
·1174 words·6 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 MasterControl AI Research
ThinkJSON presents a reinforcement learning strategy to enforce strict schema adherence in LLM generation.
SafeRoute: Adaptive Model Selection for Efficient and Accurate Safety Guardrails in Large Language Models
·2481 words·12 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 KAIST
SafeRoute efficiently enhances LLM safety by adaptively using smaller and larger safety guard models, maximizing accuracy while minimizing costs.
Rethinking Diverse Human Preference Learning through Principal Component Analysis
·2799 words·14 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Rice University
Decomposed Reward Models (DRMs) extract diverse human preferences from binary comparisons using PCA, enabling flexible and interpretable LLM alignment.
Perovskite-LLM: Knowledge-Enhanced Large Language Models for Perovskite Solar Cell Research
·3084 words·15 mins· loading · loading
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 Hong Kong University of Science and Technology
Perovskite-LLM: a new knowledge-enhanced system boosts perovskite solar cell research by integrating a domain-specific knowledge graph, high-quality datasets, and specialized LLMs for superior knowled…