You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/performance.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ This page is intended to provide clarity on how to obtain the benchmark numbers
18
18
---------------------------------------
19
19
20
20
## Methodology and Infrastructure checklist
21
-
As stated previously we encourage the community to run and the benchmarks on their own infrastructure and specific use case. As part of a reproducible and stable metodology we recommend that for each tested version/variation:
21
+
As stated previously we encourage the community to run the benchmarks on their own infrastructure and specific use case. As part of a reproducible and stable metodology we recommend that for each tested version/variation:
22
22
23
23
- Monitoring should be used to assert that the machines running the the benchmark client do not become the performance bottleneck.
24
24
@@ -42,10 +42,10 @@ The following example will:
42
42
- Enable multi-threaded client and use 2 benchmark client threads, meaning the benchmark client maximum CPU usage will be 200%.
43
43
- Be executed without pipelining (pipeline value of 1).
44
44
- If server replies with errors, show them on stdout.
45
-
- Benchmark [AI.MODELRUN](https://oss.redislabs.com/redisai/commands/#aimodelrun) command a model stored as a key's value using its specified backend and device.
45
+
- Benchmark [AI.MODELEXECUTE](https://oss.redislabs.com/redisai/commands/#aimodelexecute) command over a model stored as a key's value using its specified backend and device.
- Be executed without pipelining (pipeline value of 1).
61
61
- If server replies with errors, show them on stdout.
62
-
- Benchmark both [AI.MODELRUN](https://oss.redislabs.com/redisai/commands/#aimodelrun) command a model stored as a key's value using its specified backend and device, and [AI.SCRIPTRUN](https://oss.redislabs.com/redisai/commands/#aiscriptrun) stored as a key's value on its specified device.
62
+
- Benchmark both [AI.MODELEXECUTE](https://oss.redislabs.com/redisai/commands/#aimodelexecute) command a model stored as a key's value using its specified backend and device, and [AI.SCRIPTEXECUTE](https://oss.redislabs.com/redisai/commands/#aiscriptexecute) stored as a key's value on its specified device.
_AIBench_ is a collection of Go programs that are used to generate datasets and then benchmark the inference performance of various Model Servers. The intent is to make the AIBench extensible so that a variety of use cases and Model Servers can be included and benchmarked.
71
71
72
72
73
-
We recommend that you follow the detailed install steps [here](https://github.com/RedisAI/aibench#installation) and refer to the per-use [documentation](https://github.com/RedisAI/aibench#current-use-cases).
73
+
We recommend that you follow the detailed installation steps [here](https://github.com/RedisAI/aibench#installation) and refer to the per-use [documentation](https://github.com/RedisAI/aibench#current-use-cases).
74
74
75
75
### Current DL solutions supported:
76
76
77
-
-[RedisAI](https://redisai.io): an AI serving engine for real-time applications built by Redis Labs and Tensorwerk, seamlessly plugged into Redis.
77
+
-[RedisAI](https://redisai.io): an AI serving engine for real-time applications built by Redis and Tensorwerk, seamlessly plugged into Redis.
78
78
-[Nvidia Triton Inference Server](https://docs.nvidia.com/deeplearning/triton-inference-server): An open source inference serving software that lets teams deploy trained AI models from any framework (TensorFlow, TensorRT, PyTorch, ONNX Runtime, or a custom framework), from local storage or Google Cloud Platform or AWS S3 on any GPU- or CPU-based infrastructure.
79
79
-[TorchServe](https://pytorch.org/serve/): built and maintained by Amazon Web Services (AWS) in collaboration with Facebook, TorchServe is available as part of the PyTorch open-source project.
80
80
-[Tensorflow Serving](https://www.tensorflow.org/tfx/guide/serving): a high-performance serving system, wrapping TensorFlow and maintained by Google.
0 commit comments