Natural Language Processing
Reflective Multi-Agent Collaboration based on Large Language Models
·2567 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Gaoling School of Artificial Intelligence, Renmin University of China
COPPER enhances LLM-based multi-agent collaboration via a self-reflection mechanism and counterfactual PPO. It improves reflection quality, alleviates credit assignment issues, and shows strong perfo…
Reference Trustable Decoding: A Training-Free Augmentation Paradigm for Large Language Models
·2065 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Wuhan University
Reference Trustable Decoding (RTD) revolutionizes large language model adaptation by offering a training-free method, enabling efficient and cost-effective task adaptation without parameter adjustment…
Reducing Transformer Key-Value Cache Size with Cross-Layer Attention
·2727 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 MIT CSAIL
Cross-Layer Attention (CLA) shrinks Transformer Key-Value cache 2x, improving LLMs’ memory efficiency without accuracy loss.
Recursive Introspection: Teaching Language Model Agents How to Self-Improve
·2681 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
RISE: Recursive Introspection teaches LLMs to iteratively improve their responses, enabling self-correction and enhanced performance on challenging reasoning tasks.
REBORN: Reinforcement-Learned Boundary Segmentation with Iterative Training for Unsupervised ASR
·2781 words·14 mins·
loading
·
loading
AI Generated
Natural Language Processing
Speech Recognition
🏢 National Taiwan University
REBORN: An iterative training framework significantly improves unsupervised ASR by learning optimal speech segment boundaries using reinforcement learning, outperforming existing methods.
Reawakening knowledge: Anticipatory recovery from catastrophic interference via structured training
·2387 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 New York University
Overparameterized neural networks surprisingly recover from catastrophic interference when trained cyclically on repeated data sequences, exhibiting anticipatory knowledge reactivation.
Reasons and Solutions for the Decline in Model Performance after Editing
·2167 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
Boosting large language model performance after knowledge editing: A new method (D4S) minimizes model damage by regulating the explosive growth of parameter layers, enabling multiple effective edits.
Realizable $H$-Consistent and Bayes-Consistent Loss Functions for Learning to Defer
·1495 words·8 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Courant Institute
New surrogate loss functions for learning-to-defer achieve Bayes-consistency, realizable H-consistency, and H-consistency bounds simultaneously, resolving open questions and improving L2D performance.
RankRAG: Unifying Context Ranking with Retrieval-Augmented Generation in LLMs
·2153 words·11 mins·
loading
·
loading
Natural Language Processing
Question Answering
🏢 Georgia Tech
RankRAG: One LLM, dual-purpose instruction-tuning for superior RAG!
Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
·4652 words·22 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Meta AI
Rainbow Teaming: a novel black-box approach generates diverse adversarial prompts to enhance LLM robustness and safety, achieving over 90% attack success rate across various models.
QUEST: Quality-Aware Metropolis-Hastings Sampling for Machine Translation
·2191 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Machine Translation
🏢 University of Washington
QUEST, a novel Metropolis-Hastings sampling method, generates high-quality & diverse machine translations by using quality metrics as energy functions, overcoming limitations of likelihood-based and r…
Query-Based Adversarial Prompt Generation
·1773 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Washington
Researchers developed a query-based attack that generates adversarial prompts, fooling language models into producing harmful outputs with significantly higher success rates than previous methods, eff…
QuaRot: Outlier-Free 4-Bit Inference in Rotated LLMs
·3782 words·18 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 ETH Zurich
QuaRot: Revolutionizing 4-bit LLM inference with lossless quantization via rotation!
Quantifying the Gain in Weak-to-Strong Generalization
·2368 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Stanford University
Weakly supervised strong models outperform weak models; this gain is precisely quantified by the strong model’s misfit error on weak labels.
Quantifying and Optimizing Global Faithfulness in Persona-driven Role-playing
·2570 words·13 mins·
loading
·
loading
Natural Language Processing
Dialogue Systems
🏢 UC San Diego
New APC metric precisely quantifies & optimizes global faithfulness in persona-driven role-playing, offering a fine-grained, explainable evaluation and improving AI character consistency.
QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptation
·3333 words·16 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 MIT
QuanTA: Quantum-inspired Tensor Adaptation efficiently fine-tunes LLMs with high-rank updates, surpassing low-rank methods like LoRA for complex tasks while minimizing additional parameters.
QBB: Quantization with Binary Bases for LLMs
·1816 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Samsung AI Cambridge
QBB: A novel post-training quantization method for LLMs dramatically improves efficiency by replacing multiplications with summations, achieving state-of-the-art results with minimal accuracy loss.
Provably Transformers Harness Multi-Concept Word Semantics for Efficient In-Context Learning
·1305 words·7 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Department of Computer Science, City University of Hong Kong
Transformers excel at in-context learning (ICL), solving new tasks with just prompts. This paper provides a mathematical explanation, showing how transformers use multi-concept word semantics to achie…
Provably Mitigating Overoptimization in RLHF: Your SFT Loss is Implicitly an Adversarial Regularizer
·1760 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Northwestern University
RLHF’s overoptimization problem is mitigated by RPO, a novel algorithm that uses SFT loss as an implicit adversarial regularizer, ensuring efficient and effective LLM alignment.
ProtGO: Function-Guided Protein Modeling for Unified Representation Learning
·1875 words·9 mins·
loading
·
loading
AI Generated
Natural Language Processing
Representation Learning
🏢 Westlake University
ProtGO: A novel unified framework integrating protein sequence, structure & function for superior representation learning, significantly outperforming current methods.