The Intel OpenVINO Inference Engine is a set of libraries for executing convolutional neural networks.
GoCV support for the Intel OpenVINO Inference Engine will be able to be found here in the "gocv.io/x/gocv/openvino/ie" package.
Support in GoCV for the Intel OpenVINO Inference Engine requires version 2019 R3+ in order to work.
This code loads a Caffe model, and then uses OpenVINO inference engine to prepare it for execution on the GPU:
net := gocv.ReadNet("/path/to/your/model.caffemodel", "/path/to/your/config.proto")
if net.Empty() {
fmt.Println("Error reading network model")
return
}
// GPU usage
net.SetPreferableBackend(gocv.NetBackendType(gocv.NetBackendOpenVINO))
net.SetPreferableTarget(gocv.NetTargetType(gocv.NetTargetFP16))
// Intel Neural Compute Stick 2 usage
net.SetPreferableBackend(gocv.NetBackendType(gocv.NetBackendOpenVINO))
net.SetPreferableTarget(gocv.NetTargetType(gocv.NetTargetVPU))