You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: examples/cpp/aot_inductor/bert/README.md
+2
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,8 @@ This example uses AOTInductor to compile the [google-bert/bert-base-uncased](htt
2
2
3
3
Then, this example loads model and runs prediction using libtorch. The handler C++ source code for this examples can be found [here](src).
4
4
5
+
**Note**: Please note that due to an issue in Pytorch 2.2.1 the AOTInductor model can not be placed on a specific GPU through the API. This issue is resolved in the PT 2.3 nightlies. Please install the nightlies if you want to run multiple model worker on different GPUs.
6
+
5
7
### Setup
6
8
1. Follow the instructions in [README.md](../../../../cpp/README.md) to build the TorchServe C++ backend.
Copy file name to clipboardexpand all lines: examples/cpp/aot_inductor/resnet/README.md
+2
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,8 @@
1
1
This example uses AOTInductor to compile the Resnet50 into an so file which is then executed using libtorch.
2
2
The handler C++ source code for this examples can be found [here](src).
3
3
4
+
**Note**: Please note that due to an issue in Pytorch 2.2.1 the AOTInductor model can not be placed on a specific GPU through the API. This issue is resolved in the PT 2.3 nightlies. Please install the nightlies if you want to run multiple model worker on different GPUs.
5
+
4
6
### Setup
5
7
1. Follow the instructions in [README.md](../../../../cpp/README.md) to build the TorchServe C++ backend.
0 commit comments