Skip to main content

🏢 Sea AI Lab

Pipeline Parallelism with Controllable Memory
·3116 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Sea AI Lab
New pipeline parallelism framework achieves up to 55% higher throughput and 50% less memory usage in large language model training by systematically controlling activation memory.
Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
·2125 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Sea AI Lab
Improved few-shot jailbreaking techniques efficiently circumvent aligned language models and their defenses, achieving high success rates even against advanced protection methods.
Amortized Eigendecomposition for Neural Networks
·2211 words·11 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Sea AI Lab
Accelerate neural network training using ‘amortized eigendecomposition’ – a novel method replacing expensive eigendecomposition with faster QR decomposition while preserving accuracy.