Skip to main content

Natural Language Processing

Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies
·2496 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Stanford University
Boosting LLM performance: This research shows how larger language models need bigger vocabularies for optimal efficiency and performance.
Scaling Laws for Reward Model Overoptimization in Direct Alignment Algorithms
·2949 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Stanford University
Direct Alignment Algorithms (DAAs) for LLM alignment suffer from over-optimization, even without explicit reward models; this paper empirically demonstrates this and proposes scaling laws to understan…
SaulLM-54B & SaulLM-141B: Scaling Up Domain Adaptation for the Legal Domain
·2626 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 CINES
SaulLM-54B & SaulLM-141B achieve state-of-the-art performance on legal tasks by scaling up model size, employing a specialized instruction-following protocol, and aligning model outputs with human pre…
SafeWorld: Geo-Diverse Safety Alignment
·3977 words·19 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Los Angeles
SAFEWORLD: a new benchmark reveals and fixes LLMs’ struggle with diverse safety standards.
S$^{2}$FT: Efficient, Scalable and Generalizable LLM Fine-tuning by Structured Sparsity
·1908 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
S2FT: Structured Sparse Fine-Tuning achieves state-of-the-art LLM fine-tuning performance, training efficiency, and inference scalability by selecting sparsely and computing densely.
S-STE: Continuous Pruning Function for Efficient 2:4 Sparse Pre-training
·2718 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Tsinghua University
S-STE achieves efficient 2:4 sparse pre-training by introducing a novel continuous pruning function, overcoming the limitations of previous methods and leading to improved accuracy and speed.
RSA: Resolving Scale Ambiguities in Monocular Depth Estimators through Language Descriptions
·2064 words·10 mins· loading · loading
Natural Language Processing Vision-Language Models 🏢 Yale University
RSA: Language unlocks metric depth from single images!
RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
·3132 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Hong Kong University of Science and Technology
RouterDC: A query-based router trained via dual contrastive learning assembles multiple LLMs, significantly outperforming individual LLMs and existing routing methods on both in- and out-of-distributi…
RL on Incorrect Synthetic Data Scales the Efficiency of LLM Math Reasoning by Eight-Fold
·2612 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
Leveraging model-generated synthetic data for LLM finetuning significantly improves efficiency when using both positive and strategically constructed negative examples, resulting in an eight-fold incr…
Risk-Averse Fine-tuning of Large Language Models
·3716 words·18 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Amazon
Risk-Averse RLHF fine-tunes LLMs to minimize toxic outputs while maintaining performance.
Richelieu: Self-Evolving LLM-Based Agents for AI Diplomacy
·2475 words·12 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Peking University
Richelieu: a self-evolving LLM-based AI agent masters Diplomacy, a complex game requiring strategic planning and negotiation, without human data, by integrating self-play for continuous improvement.
Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference
·2947 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Santa Barbara
Reverse the forget-retain objectives for efficient LLM unlearning!
Rethinking Memory and Communication Costs for Efficient Data Parallel Training of Large Language Models
·2992 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Ant Group
PaRO boosts LLM training speed by up to 266% through refined model state partitioning and optimized communication.
Rethinking LLM Memorization through the Lens of Adversarial Compression
·2014 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
Researchers propose Adversarial Compression Ratio (ACR) to assess LLM memorization, offering an adversarial, flexible, and computationally efficient method for monitoring data misuse and compliance.
ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search
·2799 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Tsinghua University
ReST-MCTS*: A novel LLM self-training method using process reward guided tree search, outperforming existing methods by generating higher-quality reasoning traces for improved model accuracy.
Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
·5026 words·24 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Cambridge
This research unveils a compute-optimal recipe for fine-tuning language models into high-quality text embedding models, offering practical guidance and scaling laws for resource-constrained settings.
Representation Noising: A Defence Mechanism Against Harmful Finetuning
·3502 words·17 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Dalhousie University
RepNoise: a novel defense against harmful fine-tuning of LLMs by removing information about harmful representations, generalizing across different harmful tasks, and maintaining LLM capabilities.
ReMoDetect: Reward Models Recognize Aligned LLM's Generations
·4799 words·23 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Korea Advanced Institute of Science and Technology
ReMoDetect leverages reward models to identify and classify LLM-generated text. By using continual preference fine-tuning and incorporating human/LLM mixed text, ReMoDetect achieves state-of-the-art p…
Reinforcing LLM Agents via Policy Optimization with Action Decomposition
·2925 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Shanghai Jiao Tong University
POAD enhances LLM agents by decomposing language agent optimization to the token level, achieving finer-grained credit assignment and improved learning efficiency and generalization.
Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs
·3598 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Illinois Urbana-Champaign
Regularizing hidden states improves reward model generalization in RLHF for LLMs, boosting accuracy and mitigating over-optimization.