Skip to content
This repository was archived by the owner on Aug 7, 2025. It is now read-only.

Commit 168a2c7

Browse files
resolved the PR comments.
1 parent c08e3db commit 168a2c7

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

docs/batch_inference_with_ts.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -45,9 +45,9 @@ Started from Torchserve 0.4.1, there are two methods to configure TorchServe to
4545
1. provide the batch configuration information through [**POST /models** API](management_api.md).
4646
2. provide the batch configuration information through configuration file, config.properties.
4747

48-
The configuration that we are interested in is the following:
48+
The configuration properties that we are interested in are the following:
4949

50-
1. `batch_size`: This is the maximum batch size that a model is expected to handle.
50+
1. `batch_size`: This is the maximum batch size in `ms` that a model is expected to handle.
5151
2. `max_batch_delay`: This is the maximum batch delay time TorchServe waits to receive `batch_size` number of requests. If TorchServe doesn't receive `batch_size` number of
5252
requests before this timer time's out, it sends what ever requests that were received to the model `handler`.
5353

0 commit comments

Comments
 (0)