Skip to main content

Large Language Models

Enhancing In-Context Learning Performance with just SVD-Based Weight Pruning: A Theoretical Perspective
·2209 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Renmin University of China
SVD-based weight pruning surprisingly boosts in-context learning in large language models, especially when applied to deeper layers, offering a novel approach to model compression and efficiency.
End-to-End Ontology Learning with Large Language Models
·6489 words·31 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 University of Cambridge
OLLM: An end-to-end LLM method builds ontologies from scratch, outperforming subtask approaches and improving semantic accuracy with novel evaluation metrics.
Embedding-Aligned Language Models
·2605 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Google Research
EAGLE: Guiding LLMs using latent embeddings for controlled text generation.
Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning
·3571 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 Shanghai Jiao Tong University
Novel trajectory volatility score (TV Score) significantly improves out-of-distribution detection in mathematical reasoning by leveraging dynamic embedding trajectories, outperforming existing GLM met…
Elo Uncovered: Robustness and Best Practices in Language Model Evaluation
·2182 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Cohere
Elo rating’s reliability for LLM evaluation is challenged, revealing inconsistencies and suggesting new, more robust methods are needed for accurate model ranking.
EffiLearner: Enhancing Efficiency of Generated Code via Self-Optimization
·2876 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Hong Kong
EFFI-LEARNER: A novel self-optimization framework dramatically improves the efficiency of LLM-generated code by iteratively refining code based on execution profiles.
Efficient Sketches for Training Data Attribution and Studying the Loss Landscape
·3015 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 Google DeepMind
Novel sketching algorithms enable scalable gradient and Hessian analysis for large language models, revealing insights into their intrinsic dimensionality and challenging existing assumptions.
Efficient Prompt Optimization Through the Lens of Best Arm Identification
·4323 words·21 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 University of Virginia
TRIPLE: Efficient prompt optimization using fixed-budget best-arm identification.
Efficient Multi-task LLM Quantization and Serving for Multiple LoRA Adapters
·2138 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Peking University
LoRA-Inlaid: a novel multi-task LLM serving system boosts throughput by 1.58x, latency by 1.76x, and job completion time by 2x, while improving SLO attainment by 10x, all while maintaining model quali…
Efficient multi-prompt evaluation of LLMs
·2504 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Michigan
PromptEval efficiently estimates LLM performance across many prompts, providing robust performance metrics and enabling reliable LLM comparisons.
Efficient LLM Scheduling by Learning to Rank
·2254 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏒 UC San Diego
Learning to rank request outputs improves LLM scheduling, resulting in 2.8x lower chatbot latency and 6.5x higher synthetic data generation throughput.
Efficient LLM Jailbreak via Adaptive Dense-to-sparse Constrained Optimization
·1755 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Carnegie Mellon University
Adaptive Dense-to-sparse Constrained Optimization (ADC) efficiently jailbreaks LLMs by transforming discrete token optimization into a continuous process, achieving higher success rates than existing …
Efficient Large Multi-modal Models via Visual Context Compression
·2910 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 Johns Hopkins University
LLaVolta significantly boosts multi-modal LLMs by using visual context compression, achieving substantial training cost reduction and enhanced inference efficiency without performance loss.
Efficient Contextual LLM Cascades through Budget-Constrained Policy Learning
·3825 words·18 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Michigan
TREACLE: a reinforcement learning policy efficiently selects LLMs and prompts, achieving up to 85% cost savings while maintaining high accuracy in answering reasoning questions.
Efficient Adversarial Training in LLMs with Continuous Attacks
·2099 words·10 mins· loading · loading
Large Language Models 🏒 Mila, Université De Montréal
Boosting LLM robustness against attacks efficiently: Continuous adversarial training in embedding space outperforms discrete methods, achieving improved robustness with less computation.
Edit Distance Robust Watermarks via Indexing Pseudorandom Codes
·245 words·2 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 MIT
This paper presents a novel watermarking scheme for language models that is both undetectable and robust to a constant fraction of adversarial edits (insertions, deletions, substitutions).
Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision
·4056 words·20 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Carnegie Mellon University
AI alignment beyond human supervision is achieved via easy-to-hard generalization: training reward models on easy tasks to effectively evaluate and improve generators on harder tasks, achieving superh…
EAI: Emotional Decision-Making of LLMs in Strategic Games and Ethical Dilemmas
·4154 words·20 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 AIRI
LLMs’ emotional decision-making is assessed using a novel framework, EAI, showing that emotions significantly alter ethical and strategic choices in games. This reveals crucial biases, necessitati…
DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs
·4529 words·22 mins· loading · loading
Large Language Models 🏒 Tsinghua University
DuQuant: Dual transformations distribute outliers for stronger quantized LLMs.
DropBP: Accelerating Fine-Tuning of Large Language Models by Dropping Backward Propagation
·2987 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Seoul National University
DropBP: Accelerate LLM fine-tuning by 44% while preserving accuracy!