From 99257d9da70ca4f92dd3616aa33fea7178077335 Mon Sep 17 00:00:00 2001 From: youkaichao Date: Thu, 11 Jul 2024 11:38:40 -0700 Subject: [PATCH] [doc] update pipeline parallel in readme (#6347) Signed-off-by: Alvant --- README.md | 2 +- docs/source/index.rst | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index cced85f17e257..dac4b513cd2a2 100644 --- a/README.md +++ b/README.md @@ -56,7 +56,7 @@ vLLM is flexible and easy to use with: - Seamless integration with popular Hugging Face models - High-throughput serving with various decoding algorithms, including *parallel sampling*, *beam search*, and more -- Tensor parallelism support for distributed inference +- Tensor parallelism and pipieline parallelism support for distributed inference - Streaming outputs - OpenAI-compatible API server - Support NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs and GPUs, PowerPC CPUs diff --git a/docs/source/index.rst b/docs/source/index.rst index 67c039f25e98d..174d91b8d6a01 100644 --- a/docs/source/index.rst +++ b/docs/source/index.rst @@ -38,7 +38,7 @@ vLLM is flexible and easy to use with: * Seamless integration with popular HuggingFace models * High-throughput serving with various decoding algorithms, including *parallel sampling*, *beam search*, and more -* Tensor parallelism support for distributed inference +* Tensor parallelism and pipieline parallelism support for distributed inference * Streaming outputs * OpenAI-compatible API server * Support NVIDIA GPUs and AMD GPUs