Skip to main content

🏢 University of Alberta

NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks
·2943 words·14 mins
AI Generated 🤗 Daily Papers Natural Language Processing Large Language Models 🏢 University of Alberta
NeuZip dynamically compresses neural network weights, achieving memory-efficient training and inference without performance loss, significantly reducing the memory footprint of large language models.