Skip to content

Commit d593c53

Browse files
committed
minor changes
1 parent 7ded23a commit d593c53

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

docs/performance.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ This page is intended to provide clarity on how to obtain the benchmark numbers
1818
---------------------------------------
1919

2020
## Methodology and Infrastructure checklist
21-
As stated previously we encourage the community to run and the benchmarks on their own infrastructure and specific use case. As part of a reproducible and stable metodology we recommend that for each tested version/variation:
21+
As stated previously we encourage the community to run the benchmarks on their own infrastructure and specific use case. As part of a reproducible and stable metodology we recommend that for each tested version/variation:
2222

2323
- Monitoring should be used to assert that the machines running the the benchmark client do not become the performance bottleneck.
2424

@@ -42,10 +42,10 @@ The following example will:
4242
- Enable multi-threaded client and use 2 benchmark client threads, meaning the benchmark client maximum CPU usage will be 200%.
4343
- Be executed without pipelining (pipeline value of 1).
4444
- If server replies with errors, show them on stdout.
45-
- Benchmark [AI.MODELRUN](https://oss.redislabs.com/redisai/commands/#aimodelrun) command a model stored as a key's value using its specified backend and device.
45+
- Benchmark [AI.MODELEXECUTE](https://oss.redislabs.com/redisai/commands/#aimodelexecute) command over a model stored as a key's value using its specified backend and device.
4646

4747
```
48-
redis-benchmark -e --threads 2 -c 50 -P 1 -n 100000 AI.MODELRUN <key> INPUTS <input> [input ...] OUTPUTS <output> [output ...]
48+
redis-benchmark --threads 2 -c 50 -P 1 -n 100000 AI.MODELEXECUTE <key> INPUTS <input_count> <input> [input ...] OUTPUTS <output_count> <output> [output ...]
4949
```
5050

5151
## Using memtier-benchmark
@@ -59,22 +59,22 @@ The following example will:
5959
- Store the benchmark result in `result.json`.
6060
- Be executed without pipelining (pipeline value of 1).
6161
- If server replies with errors, show them on stdout.
62-
- Benchmark both [AI.MODELRUN](https://oss.redislabs.com/redisai/commands/#aimodelrun) command a model stored as a key's value using its specified backend and device, and [AI.SCRIPTRUN](https://oss.redislabs.com/redisai/commands/#aiscriptrun) stored as a key's value on its specified device.
62+
- Benchmark both [AI.MODELEXECUTE](https://oss.redislabs.com/redisai/commands/#aimodelexecute) command a model stored as a key's value using its specified backend and device, and [AI.SCRIPTEXECUTE](https://oss.redislabs.com/redisai/commands/#aiscriptexecute) stored as a key's value on its specified device.
6363

6464
```
65-
memtier_benchmark --clients 50 --threads 4 --requests 10000 --pipeline 1 --json-out-file results.json --command "AI.MODELRUN model_key INPUTS input1 OUTPUTS output1" --command "AI.SCRIPTRUN script_key INPUTS input2 OUTPUTS output2"
65+
memtier_benchmark --clients 50 --threads 4 --requests 10000 --pipeline 1 --json-out-file results.json --command "AI.MODELEXECUTE model_key INPUTS input_count input1 ... OUTPUTS output_count output1 ..." --command "AI.SCRIPTEXECUTE script_key entry_point INPUTS input_count input1 ... OUTPUTS output_count output1 ..."
6666
```
6767

6868
## Using aibench
6969

7070
_AIBench_ is a collection of Go programs that are used to generate datasets and then benchmark the inference performance of various Model Servers. The intent is to make the AIBench extensible so that a variety of use cases and Model Servers can be included and benchmarked.
7171

7272

73-
We recommend that you follow the detailed install steps [here](https://github.com/RedisAI/aibench#installation) and refer to the per-use [documentation](https://github.com/RedisAI/aibench#current-use-cases).
73+
We recommend that you follow the detailed installation steps [here](https://github.com/RedisAI/aibench#installation) and refer to the per-use [documentation](https://github.com/RedisAI/aibench#current-use-cases).
7474

7575
### Current DL solutions supported:
7676

77-
- [RedisAI](https://redisai.io): an AI serving engine for real-time applications built by Redis Labs and Tensorwerk, seamlessly plugged into ​Redis.
77+
- [RedisAI](https://redisai.io): an AI serving engine for real-time applications built by Redis and Tensorwerk, seamlessly plugged into ​Redis.
7878
- [Nvidia Triton Inference Server](https://docs.nvidia.com/deeplearning/triton-inference-server): An open source inference serving software that lets teams deploy trained AI models from any framework (TensorFlow, TensorRT, PyTorch, ONNX Runtime, or a custom framework), from local storage or Google Cloud Platform or AWS S3 on any GPU- or CPU-based infrastructure.
7979
- [TorchServe](https://pytorch.org/serve/): built and maintained by Amazon Web Services (AWS) in collaboration with Facebook, TorchServe is available as part of the PyTorch open-source project.
8080
- [Tensorflow Serving](https://www.tensorflow.org/tfx/guide/serving): a high-performance serving system, wrapping TensorFlow and maintained by Google.

0 commit comments

Comments
 (0)