DeepSpeed-FastGen: High-throughput Text Generation for LLMs via MII and DeepSpeed-Inference Permalink
title: "ZeRO-Inference: 20X faster inference through weight quantization and KV cache offloading" excerpt: "" link: https://github.com/deepspeedai/DeepSpeedE...
Partition-aware ZeRO with up to 2x reduction in communication time!
DeepSpeed was used to train the world’s largest language model.