Skip to main content

Natural Language Processing

Interpreting Learned Feedback Patterns in Large Language Models
·2900 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Oxford
Researchers developed methods to measure and interpret the divergence between learned feedback patterns (LFPs) in LLMs and human preferences, helping minimize discrepancies between LLM behavior and tr…
Instruction Tuning With Loss Over Instructions
·4022 words·19 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University College London
Boost LLM performance with INSTRUCTION MODELLING: a simple yet effective instruction tuning method that improves model outputs by over 100% in some cases by applying loss to both instructions and outp…
Instance-adaptive Zero-shot Chain-of-Thought Prompting
·2222 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 College of Computer Science and Technology, Jilin University
Instance-adaptive prompting significantly improves zero-shot Chain-of-Thought reasoning in LLMs by dynamically selecting prompts tailored to each instance, leading to consistent performance gains acro…
Initialization is Critical to Whether Transformers Fit Composite Functions by Reasoning or Memorizing
·2409 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Shanghai Jiao Tong University
Transformer model initialization dramatically affects whether it reasons or memorizes, impacting performance on compositional tasks.
Information Re-Organization Improves Reasoning in Large Language Models
·2018 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Zhejiang University
InfoRE: A novel method improving large language models’ reasoning by reorganizing information to highlight logical relationships, resulting in a 4% average accuracy boost across various tasks.
InfoRM: Mitigating Reward Hacking in RLHF via Information-Theoretic Reward Modeling
·5629 words·27 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Wuhan University
InfoRM tackles reward hacking in RLHF using an information-theoretic approach, enhancing generalizability and enabling overoptimization detection.
InfLLM: Training-Free Long-Context Extrapolation for LLMs with an Efficient Context Memory
·2045 words·10 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Tsinghua University
InfLLM: Training-free long-context extrapolation for LLMs via efficient context memory.
Inevitable Trade-off between Watermark Strength and Speculative Sampling Efficiency for Language Models
·2218 words·11 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Maryland
Injecting watermarks into LLM outputs while speeding up generation is impossible; this paper proves this trade-off and offers methods prioritizing either watermark strength or speed.
INDICT: Code Generation with Internal Dialogues of Critiques for Both Security and Helpfulness
·3011 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Salesforce Research
INDICT, a novel framework, empowers LLMs with internal dialogues of critiques to enhance code generation, prioritizing both safety and helpfulness, resulting in +10% absolute improvement across variou…
Incentivizing Quality Text Generation via Statistical Contracts
·1392 words·7 mins· loading · loading
Natural Language Processing Text Generation 🏢 Technion - Israel Institute of Technology
Cost-robust contracts, inspired by statistical hypothesis tests, incentivize quality in LLM text generation, overcoming the moral hazard of pay-per-token models.
In-Context Learning with Representations: Contextual Generalization of Trained Transformers
·1880 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
Transformers learn contextual information for generalization to unseen examples and tasks, even with limited training data, converging linearly to a global minimum.
In-Context Learning State Vector with Inner and Momentum Optimization
·2182 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Harbin Institute of Technology (Shenzhen)
This paper introduces inner and momentum optimization to enhance the state vector for in-context learning, improving performance and scalability in LLMs.
In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization
·436 words·3 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC Berkeley
Linear Transformer Blocks (LTBs) achieve near-optimal in-context learning (ICL) for linear regression by effectively implementing one-step gradient descent with learnable initialization, a significant…
Improving Sparse Decomposition of Language Model Activations with Gated Sparse Autoencoders
·4021 words·19 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Google DeepMind
Gated Sparse Autoencoders (GSAEs) achieve Pareto improvement over baseline SAEs for unsupervised feature discovery in language models, resolving the shrinkage bias of L1 penalty by separating feature …
Improving Gloss-free Sign Language Translation by Reducing Representation Density
·3386 words·16 mins· loading · loading
AI Generated Natural Language Processing Machine Translation 🏢 Tencent AI Lab
SignCL, a novel contrastive learning strategy, significantly boosts gloss-free sign language translation by mitigating representation density, achieving substantial performance gains.
Improving Context-Aware Preference Modeling for Language Models
·1939 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Microsoft Research
Context-aware preference modeling improves language model alignment by resolving ambiguity through a two-step process: context selection followed by context-specific preference evaluation. The approa…
Improved Generation of Adversarial Examples Against Safety-aligned LLMs
·2198 words·11 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 UC Davis
Researchers developed novel methods to improve the generation of adversarial examples against safety-aligned LLMs, achieving significantly higher attack success rates compared to existing techniques.
Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
·2125 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Sea AI Lab
Improved few-shot jailbreaking techniques efficiently circumvent aligned language models and their defenses, achieving high success rates even against advanced protection methods.
Implicit Regularization of Sharpness-Aware Minimization for Scale-Invariant Problems
·2239 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 ETH Zurich
Boosting deep learning generalization, this work unveils SAM’s implicit regularization using ‘balancedness’, a new metric. A resource-efficient variant, BAR, achieves 95% computational savings with i…
Implicit Optimization Bias of Next-token Prediction in Linear Models
·1645 words·8 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of British Columbia
Researchers reveal implicit optimization biases in next-token prediction for language models, showing how gradient descent selects solutions based on data sparsity and a novel margin concept, impactin…