Skip to main content

Natural Language Processing

Enhancing LLM’s Cognition via Structurization
·3694 words·18 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Zhejiang University
LLMs struggle with complex, long-form text. This paper introduces ‘context structurization,’ transforming unstructured text into a structured format to enhance LLM comprehension. Experiments across …
Enhancing Large Vision Language Models with Self-Training on Image Comprehension
·3514 words·17 mins· loading · loading
AI Generated Natural Language Processing Vision-Language Models 🏢 UC Los Angeles
Self-Training on Image Comprehension (STIC) significantly boosts Large Vision Language Model (LVLM) performance using unlabeled image data. STIC generates a preference dataset for image descriptions …
Enhancing Large Language Models through Adaptive Tokenizers
·1963 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Huawei Noah's Ark Lab
Adaptive tokenizers enhance LLMs by dynamically optimizing vocabulary during training, improving accuracy without increasing vocabulary size.
Enhancing In-Context Learning Performance with just SVD-Based Weight Pruning: A Theoretical Perspective
·2209 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Renmin University of China
SVD-based weight pruning surprisingly boosts in-context learning in large language models, especially when applied to deeper layers, offering a novel approach to model compression and efficiency.
End-to-End Ontology Learning with Large Language Models
·6489 words·31 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Cambridge
OLLM: An end-to-end LLM method builds ontologies from scratch, outperforming subtask approaches and improving semantic accuracy with novel evaluation metrics.
Embedding-Aligned Language Models
·2605 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Google Research
EAGLE: Guiding LLMs using latent embeddings for controlled text generation.
Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning
·3571 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Shanghai Jiao Tong University
Novel trajectory volatility score (TV Score) significantly improves out-of-distribution detection in mathematical reasoning by leveraging dynamic embedding trajectories, outperforming existing GLM met…
Elo Uncovered: Robustness and Best Practices in Language Model Evaluation
·2182 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cohere
Elo rating’s reliability for LLM evaluation is challenged, revealing inconsistencies and suggesting new, more robust methods are needed for accurate model ranking.
EffiLearner: Enhancing Efficiency of Generated Code via Self-Optimization
·2876 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Hong Kong
EFFI-LEARNER: A novel self-optimization framework dramatically improves the efficiency of LLM-generated code by iteratively refining code based on execution profiles.
Efficient Sketches for Training Data Attribution and Studying the Loss Landscape
·3015 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Google DeepMind
Novel sketching algorithms enable scalable gradient and Hessian analysis for large language models, revealing insights into their intrinsic dimensionality and challenging existing assumptions.
Efficient Prompt Optimization Through the Lens of Best Arm Identification
·4323 words·21 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Virginia
TRIPLE: Efficient prompt optimization using fixed-budget best-arm identification.
Efficient Multi-task LLM Quantization and Serving for Multiple LoRA Adapters
·2138 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Peking University
LoRA-Inlaid: a novel multi-task LLM serving system boosts throughput by 1.58x, latency by 1.76x, and job completion time by 2x, while improving SLO attainment by 10x, all while maintaining model quali…
Efficient multi-prompt evaluation of LLMs
·2504 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Michigan
PromptEval efficiently estimates LLM performance across many prompts, providing robust performance metrics and enabling reliable LLM comparisons.
Efficient Minimum Bayes Risk Decoding using Low-Rank Matrix Completion Algorithms
·3719 words·18 mins· loading · loading
AI Generated Natural Language Processing Machine Translation 🏢 Google
Fast approximation of Minimum Bayes Risk (MBR) decoding achieved using low-rank matrix completion algorithms, drastically reducing computational cost without sacrificing translation quality.
Efficient LLM Scheduling by Learning to Rank
·2254 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC San Diego
Learning to rank request outputs improves LLM scheduling, resulting in 2.8x lower chatbot latency and 6.5x higher synthetic data generation throughput.
Efficient LLM Jailbreak via Adaptive Dense-to-sparse Constrained Optimization
·1755 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
Adaptive Dense-to-sparse Constrained Optimization (ADC) efficiently jailbreaks LLMs by transforming discrete token optimization into a continuous process, achieving higher success rates than existing …
Efficient Large Multi-modal Models via Visual Context Compression
·2910 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Johns Hopkins University
LLaVolta significantly boosts multi-modal LLMs by using visual context compression, achieving substantial training cost reduction and enhanced inference efficiency without performance loss.
Efficient Contextual LLM Cascades through Budget-Constrained Policy Learning
·3825 words·18 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Michigan
TREACLE: a reinforcement learning policy efficiently selects LLMs and prompts, achieving up to 85% cost savings while maintaining high accuracy in answering reasoning questions.
Edit Distance Robust Watermarks via Indexing Pseudorandom Codes
·245 words·2 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 MIT
This paper presents a novel watermarking scheme for language models that is both undetectable and robust to a constant fraction of adversarial edits (insertions, deletions, substitutions).
Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision
·4056 words·20 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Carnegie Mellon University
AI alignment beyond human supervision is achieved via easy-to-hard generalization: training reward models on easy tasks to effectively evaluate and improve generators on harder tasks, achieving superh…