cancel
Showing results for 
Search instead for 
Did you mean: 
TonyD1
Cadet
Cadet
  • 37 Views

API for inference (OpenVINNO)

In the Openshift AI lab I tried to deploy my own model using a saved onnx model and OpenVINNO in serving model. In S3 bucket I uploaded the .onnx file

It was successful and in status ready. When I try to use POST Rest API I get an error

"The model with requested version is not found"

Name:mymodel namespace: personalmodels

Model is using 4 numerical inputs and one numerical output

As a host, I used both option https://mymodel-predictor-personalmodels.apps.ocp4.example.com/v2/models/mymodel/infer and https://mymodel-personalmodels.apps.ocp4.example.com/v2/models/mymodel/infer

Then I passed the input data

curl -sk -X 'POST' <host>
-H 'accept: application/json' -H 'Content-Type: application/json' \
-d '{
"inputs": [
{
"data": [0.2,4,2,5],
"datatype": "FP32", "name": "inputs", "shape": [1,4]
}
],
"outputs": [{"name": "output0"}]
}' 

Which API calls should work for OpenVINNO/onnx? Why does it complain for version? 

 

0 Replies
Join the discussion
You must log in to join this conversation.