🏢 Sea AI Lab
Pipeline Parallelism with Controllable Memory
·3116 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Sea AI Lab
New pipeline parallelism framework achieves up to 55% higher throughput and 50% less memory usage in large language model training by systematically controlling activation memory.
Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
·2125 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Sea AI Lab
Improved few-shot jailbreaking techniques efficiently circumvent aligned language models and their defenses, achieving high success rates even against advanced protection methods.
Amortized Eigendecomposition for Neural Networks
·2211 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Sea AI Lab
Accelerate neural network training using ‘amortized eigendecomposition’ – a novel method replacing expensive eigendecomposition with faster QR decomposition while preserving accuracy.