Large Language Models
Instruction Tuning With Loss Over Instructions
·4022 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University College London
Boost LLM performance with INSTRUCTION MODELLING: a simple yet effective instruction tuning method that improves model outputs by over 100% in some cases by applying loss to both instructions and outp…
Instance-adaptive Zero-shot Chain-of-Thought Prompting
·2222 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 College of Computer Science and Technology, Jilin University
Instance-adaptive prompting significantly improves zero-shot Chain-of-Thought reasoning in LLMs by dynamically selecting prompts tailored to each instance, leading to consistent performance gains acro…
Initialization is Critical to Whether Transformers Fit Composite Functions by Reasoning or Memorizing
·2409 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Shanghai Jiao Tong University
Transformer model initialization dramatically affects whether it reasons or memorizes, impacting performance on compositional tasks.
Information Re-Organization Improves Reasoning in Large Language Models
·2018 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Zhejiang University
InfoRE: A novel method improving large language models’ reasoning by reorganizing information to highlight logical relationships, resulting in a 4% average accuracy boost across various tasks.
InfoRM: Mitigating Reward Hacking in RLHF via Information-Theoretic Reward Modeling
·5629 words·27 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Wuhan University
InfoRM tackles reward hacking in RLHF using an information-theoretic approach, enhancing generalizability and enabling overoptimization detection.
InfLLM: Training-Free Long-Context Extrapolation for LLMs with an Efficient Context Memory
·2045 words·10 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Tsinghua University
InfLLM: Training-free long-context extrapolation for LLMs via efficient context memory.
Inevitable Trade-off between Watermark Strength and Speculative Sampling Efficiency for Language Models
·2218 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Maryland
Injecting watermarks into LLM outputs while speeding up generation is impossible; this paper proves this trade-off and offers methods prioritizing either watermark strength or speed.
Induced Model Matching: Restricted Models Help Train Full-Featured Models
·2402 words·12 mins·
loading
·
loading
Large Language Models
🏢 University of Illinois Chicago
Restricted models often outperform full-featured models when training data is limited. This paper introduces Induced Model Matching (IMM), a novel technique that uses a restricted model as a guide to…
INDICT: Code Generation with Internal Dialogues of Critiques for Both Security and Helpfulness
·3011 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Salesforce Research
INDICT, a novel framework, empowers LLMs with internal dialogues of critiques to enhance code generation, prioritizing both safety and helpfulness, resulting in +10% absolute improvement across variou…
In-Context Learning with Representations: Contextual Generalization of Trained Transformers
·1880 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
Transformers learn contextual information for generalization to unseen examples and tasks, even with limited training data, converging linearly to a global minimum.
In-Context Learning State Vector with Inner and Momentum Optimization
·2182 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Harbin Institute of Technology (Shenzhen)
This paper introduces inner and momentum optimization to enhance the state vector for in-context learning, improving performance and scalability in LLMs.
In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization
·436 words·3 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 UC Berkeley
Linear Transformer Blocks (LTBs) achieve near-optimal in-context learning (ICL) for linear regression by effectively implementing one-step gradient descent with learnable initialization, a significant…
Improving Sparse Decomposition of Language Model Activations with Gated Sparse Autoencoders
·4021 words·19 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Google DeepMind
Gated Sparse Autoencoders (GSAEs) achieve Pareto improvement over baseline SAEs for unsupervised feature discovery in language models, resolving the shrinkage bias of L1 penalty by separating feature …
Improving Context-Aware Preference Modeling for Language Models
·1939 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
Context-aware preference modeling improves language model alignment by resolving ambiguity through a two-step process: context selection followed by context-specific preference evaluation. The approa…
Improved Generation of Adversarial Examples Against Safety-aligned LLMs
·2198 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 UC Davis
Researchers developed novel methods to improve the generation of adversarial examples against safety-aligned LLMs, achieving significantly higher attack success rates compared to existing techniques.
Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
·2125 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Sea AI Lab
Improved few-shot jailbreaking techniques efficiently circumvent aligned language models and their defenses, achieving high success rates even against advanced protection methods.
Implicit Regularization of Sharpness-Aware Minimization for Scale-Invariant Problems
·2239 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 ETH Zurich
Boosting deep learning generalization, this work unveils SAM’s implicit regularization using ‘balancedness’, a new metric. A resource-efficient variant, BAR, achieves 95% computational savings with i…
Implicit Optimization Bias of Next-token Prediction in Linear Models
·1645 words·8 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of British Columbia
Researchers reveal implicit optimization biases in next-token prediction for language models, showing how gradient descent selects solutions based on data sparsity and a novel margin concept, impactin…
Imitating Language via Scalable Inverse Reinforcement Learning
·3278 words·16 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Google DeepMind
This study presents a novel Inverse Reinforcement Learning (IRL) approach for fine-tuning large language models, offering improved performance and generation diversity compared to standard methods.
IDGen: Item Discrimination Induced Prompt Generation for LLM Evaluation
·2083 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Tencent AI Lab
IDGen synthesizes LLM evaluation prompts using Item Discrimination theory, creating a more challenging and discriminative dataset than previous methods.