mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2026-04-23 00:17:25 +08:00
[Docs] fix PaddleOCR-VL docs bug (#4702)
This commit is contained in:
@@ -24,7 +24,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
|
||||
--max-model-len 16384 \
|
||||
--max-num-batched-tokens 16384 \
|
||||
--gpu-memory-utilization 0.8 \
|
||||
--max-num-seqs 128 \
|
||||
--max-num-seqs 128
|
||||
```
|
||||
**Example 2:** Deploying a 16K Context Service on a Single A100 GPU
|
||||
```shell
|
||||
@@ -36,7 +36,7 @@ python -m fastdeploy.entrypoints.openai.api_server \
|
||||
--max-model-len 16384 \
|
||||
--max-num-batched-tokens 16384 \
|
||||
--gpu-memory-utilization 0.8 \
|
||||
--max-num-seqs 256 \
|
||||
--max-num-seqs 256
|
||||
```
|
||||
|
||||
An example is a set of configurations that can run stably while also delivering relatively good performance. If you have further requirements for precision or performance, please continue reading the content below.
|
||||
|
||||
Reference in New Issue
Block a user