You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Aug 7, 2025. It is now read-only.
Copy file name to clipboardExpand all lines: CONTRIBUTING.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,11 +21,14 @@ If you are interested in contributing to TorchServe, your contributions will fal
21
21
```
22
22
> Supported cuda versions as cu111, cu102, cu101, cu92
23
23
24
-
- Execute sanity suite
24
+
- Run sanity suite
25
25
```bash
26
-
python ./torchserve_sanity.py
26
+
python torchserve_sanity.py
27
27
```
28
+
- Run Regression test`python test/regression_tests.py`
28
29
- For running individual test suites refer [code_coverage](docs/code_coverage.md) documentation
30
+
- If you are updating an existing model make sure that performance hasn't degraded by running [benchmarks](https://github.com/pytorch/serve/tree/master/benchmarks) on the master branch and your branch and verify there is no performance regression
31
+
- For large changes make sure to run the [automated benchmark suite](https://github.com/pytorch/serve/tree/master/test/benchmark) which will run the apache bench tests on several configurations of CUDA and EC2 instances
29
32
- If you need more context on a particular issue, please create raise a ticket on [`TorchServe` GH repo](https://github.com/pytorch/serve/issues/new/choose) or connect to [PyTorch's slack channel](https://pytorch.slack.com/)
30
33
31
34
Once you finish implementing a feature or bug-fix, please send a Pull Request to https://github.com/pytorch/serve. Use this [template](pull_request_template.md) when creating a Pull Request.
After you execute the `torchserve`command above, TorchServe runs on your host, listening for inference requests.
138
142
139
-
**Note**: If you specify model(s) when you run TorchServe, it automatically scales backend workers to the number equal to available vCPUs (if you run on a CPU instance) or to the number of available GPUs (if you run on a GPU instance). In case of powerful hosts with a lot of compute resoures (vCPUs or GPUs), this start up and autoscaling process might take considerable time. If you want to minimize TorchServe start up time you should avoid registering and scaling the model during start up time and move that to a later point by using corresponding [Management API](docs/management_api.md#register-a-model), which allows finer grain control of the resources that are allocated for any particular model).
143
+
**Note**: If you specify model(s) when you run TorchServe, it automatically scales backend workers to the number equal to available vCPUs (if you run on a CPU instance) or to the number of available GPUs (if you run on a GPU instance). In case of powerful hosts with a lot of compute resources (vCPUs or GPUs), this start up and autoscaling process might take considerable time. If you want to minimize TorchServe start up time you should avoid registering and scaling the model during start up time and move that to a later point by using corresponding [Management API](docs/management_api.md#register-a-model), which allows finer grain control of the resources that are allocated for any particular model).
140
144
141
145
### Get predictions from a model
142
146
@@ -212,6 +216,11 @@ To stop the currently running TorchServe instance, run:
212
216
torchserve --stop
213
217
```
214
218
219
+
### Inspect the logs
220
+
All the logs you've seen as output to stdout related to model registration, management, inference are recorded in the `/logs` folder.
221
+
222
+
High level performance data like Throughput or Percentile Precision can be generated with [Benchmark](benchmark/README.md) and visualized in a report.
223
+
215
224
### Concurrency And Number of Workers
216
225
TorchServe exposes configurations that allow the user to configure the number of worker threads on CPU and GPUs. There is an important config property that can speed up the server depending on the workload.
217
226
*Note: the following property has bigger impact under heavy workloads.*
@@ -239,9 +248,12 @@ Feel free to skim the full list of [available examples](examples/README.md)
239
248
## Learn More
240
249
241
250
* [Full documentation on TorchServe](docs/README.md)
242
-
* [Manage models API](docs/management_api.md)
251
+
* [Model Management API](docs/management_api.md)
243
252
* [Inference API](docs/inference_api.md)
253
+
* [Metrics API](docs/metrics.md)
244
254
* [Package models for use with TorchServe](model-archiver/README.md)
255
+
* [Deploying TorchServe with Kubernetes](kubernetes/README.md)
* [TorchServe model zoo for pre-trained and pre-packaged models-archives](docs/model_zoo.md)
246
258
247
259
## Contributing
@@ -255,4 +267,4 @@ To file a bug or request a feature, please file a GitHub issue. For filing pull
255
267
## Disclaimer
256
268
This repository is jointly operated and maintained by Amazon, Facebook and a number of individual contributors listed in the [CONTRIBUTORS](https://github.com/pytorch/serve/graphs/contributors) file. For questions directed at Facebook, please send an email to [email protected]. For questions directed at Amazon, please send an email to [email protected]. For all other questions, please open up an issue in this repository [here](https://github.com/pytorch/serve/issues).
257
269
258
-
*TorchServe acknowledges the [Multi Model Server (MMS)](https://github.com/awslabs/multi-model-server) project from which it was derived*
270
+
*TorchServe acknowledges the [Multi Model Server (MMS)](https://github.com/awslabs/multi-model-server) project from which it was derived*
@@ -315,6 +316,9 @@ The reports are generated at location "/tmp/benchmark/"
315
316
### Sample latency graph
316
317

317
318
319
+
# Benchmarking Apache Bench AWS
320
+
If you're making a large change to TorchServe it's best to run an [automated benchmarking suite on AWS](https://github.com/pytorch/serve/tree/master/test/benchmark) so that you can test multiple CUDA versions and EC2 hardware configurations easily.
0 commit comments