Large Language Models
Pensez: Less Data, Better Reasoning -- Rethinking French LLM
·3508 words·17 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ UniversitΓ© Grenoble Alpes
Pensez: Strategic fine-tuning beats massive data for superior reasoning in French LLMs, challenging conventional wisdom.
$Ο$-Decoding: Adaptive Foresight Sampling for Balanced Inference-Time Exploration and Exploitation
·3341 words·16 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Shanghai AI Lab
Ξ¦-Decoding: Adaptive foresight sampling balances inference-time exploration and exploitation for better LLM reasoning.
Investigating Human-Aligned Large Language Model Uncertainty
·1326 words·7 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Vanderbilt University
This research explores how well LLM uncertainty measures align with human uncertainty, finding Bayesian and top-k entropy measures show promise.
Light-R1: Curriculum SFT, DPO and RL for Long COT from Scratch and Beyond
·1730 words·9 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Qiyuan Tech
Light-R1: Trains long COT models from scratch using curriculum SFT, DPO, and RL, achieving SOTA performance and strong generalization with limited resources.
BiasEdit: Debiasing Stereotyped Language Models via Model Editing
·2942 words·14 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ University of California, San Diego
BIASEDIT: Efficiently debiasing language models via lightweight network edits!
SEAP: Training-free Sparse Expert Activation Pruning Unlock the Brainpower of Large Language Models
·3962 words·19 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Renmin University of China
SEAP: Unlock LLM brainpower w/ training-free sparse expert activation pruning! Boost efficiency, keep accuracy. Optimize LLMs now!
Sketch-of-Thought: Efficient LLM Reasoning with Adaptive Cognitive-Inspired Sketching
·1708 words·9 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ KAIST
Sketch-of-Thought(SoT) reduces LLM token usage by up to 76% while maintaining (or improving) accuracy via cognitive-inspired sketching.
Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts
·3804 words·18 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Shanghai AI Laboratory
Linear-MoE: Integrates Linear Sequence Modeling with Mixture-of-Experts, achieving efficiency gains and competitive performance in large language models.
TinyR1-32B-Preview: Boosting Accuracy with Branch-Merge Distillation
·570 words·3 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Peking University
TinyR1-32B-Preview: A novel branch-merge distillation approach that significantly enhances model accuracy and reduces computational costs for LLMs.
FuseChat-3.0: Preference Optimization Meets Heterogeneous Model Fusion
·449 words·3 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ School of Computer Science and Engineering, Sun Yat-Sen University, China
FuseChat-3.0: Heterogeneous model fusion boosts LLM performance via preference optimization, creating efficient and powerful language models.
Beyond RAG: Task-Aware KV Cache Compression for Comprehensive Knowledge Reasoning
·2528 words·12 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ SAP Labs
Task-aware KV cache compression enables efficient knowledge reasoning in LLMs.
An Empirical Study on Eliciting and Improving R1-like Reasoning Models
·3690 words·18 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Renmin University of China
This paper explores and improves R1-like reasoning models through RL and tool manipulation, achieving significant accuracy gains.
Process-based Self-Rewarding Language Models
·3066 words·15 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Nanjing University
Process-based Self-Rewarding advances LLMs, surpassing human reasoning in math by step-wise self-evaluation.
Wikipedia in the Era of LLMs: Evolution and Risks
·3967 words·19 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Huazhong University of Science and Technology
LLMs modestly affect Wikipedia, subtly altering content and potentially skewing NLP benchmarks.
Mask-DPO: Generalizable Fine-grained Factuality Alignment of LLMs
·2943 words·14 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Shanghai Jiao Tong University
Mask-DPO: Fine-grained Factuality Alignment improves LLMs’ factuality by masking sentence-level errors during DPO training for enhanced knowledge alignment.
LINGOLY-TOO: Disentangling Memorisation from Reasoning with Linguistic Templatisation and Orthographic Obfuscation
·4618 words·22 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ University of Oxford
LINGOLY-TOO: A new benchmark to disentangle memorization from reasoning in LLMs using linguistic templatization and orthographic obfuscation.
Word Form Matters: LLMs' Semantic Reconstruction under Typoglycemia
·2734 words·13 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ MBZUAI
LLMs primarily rely on word form, unlike humans, when reconstructing semantics, indicating a need for context-aware mechanisms to enhance LLMs’ adaptability.
When an LLM is apprehensive about its answers -- and when its uncertainty is justified
·3209 words·16 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Skolkovo Institute of Science and Technology (Skoltech)
This paper investigates when LLMs are apprehensive and when their uncertainty is justified.
SampleMix: A Sample-wise Pre-training Data Mixing Strategey by Coordinating Data Quality and Diversity
·2929 words·14 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Meituan Group
SampleMix: Sample-wise Pre-training Data Mixing by Coordinating Data Quality and Diversity
Liger: Linearizing Large Language Models to Gated Recurrent Structures
·4096 words·20 mins·
loading
·
loading
AI Generated
π€ Daily Papers
Natural Language Processing
Large Language Models
π’ Shanghai AI Laboratory
Liger: LLMs linearized to gated recurrent models, enabling efficient deployment via key matrix repurposing and LoRA fine-tuning.