forked from openvinotoolkit/openvino
-
Notifications
You must be signed in to change notification settings - Fork 0
/
vcpkg.json
143 lines (143 loc) · 4.57 KB
/
vcpkg.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
{
"$schema": "https://raw.githubusercontent.com/microsoft/vcpkg-tool/main/docs/vcpkg.schema.json",
"name": "openvino",
"version": "2024.5.0",
"maintainers": "OpenVINO Developers <openvino@intel.com>",
"summary": "This is a port for Open Visual Inference And Optimization toolkit for AI inference",
"description": [
"Intel® Distribution of OpenVINO™ toolkit is an open-source toolkit for optimizing ",
"and deploying AI inference. It can be used to develop applications and solutions based ",
"on deep learning tasks, such as: emulation of human vision, automatic speech recognition, ",
"natural language processing, recommendation systems, etc. It provides high-performance ",
"and rich deployment options, from edge to cloud"
],
"homepage": "https://github.com/openvinotoolkit/openvino",
"documentation": "https://docs.openvino.ai/latest/index.html",
"license": "Apache-2.0",
"builtin-baseline": "88a0bf87b5efd6270502dfe4dde75dd155bd992b",
"dependencies": [
{
"name": "pkgconf",
"host": true
},
"pugixml",
{
"name": "tbb",
"version>=": "2021.10.0#2"
},
"rapidjson",
{
"name": "xbyak",
"platform": "!(arm | uwp)",
"version>=": "6.69"
}
],
"default-features": [
"auto",
"auto-batch",
{
"name": "cpu",
"platform": "!(windows & arm)"
},
{
"name": "gpu",
"platform": "(x64 | arm64) & !(arm64 & windows) & !(osx | uwp)"
},
"hetero",
"ir",
"onnx",
"paddle",
"pytorch",
"tensorflow",
"tensorflow-lite"
],
"features": {
"auto": {
"description": "Enables Auto plugin for inference"
},
"auto-batch": {
"description": "Enables Auto Batch plugin for inference, useful for throughput mode"
},
"cpu": {
"description": "Enables CPU plugin for inference",
"supports": "!(windows & arm)"
},
"gpu": {
"description": "Enables GPU plugin for inference",
"supports": "(x64 | arm64) & !(arm64 & windows) & !(osx | uwp)",
"dependencies": [
"opencl"
]
},
"hetero": {
"description": "Enables Hetero plugin for inference"
},
"ir": {
"description": "Enables IR frontend for reading models in OpenVINO IR format"
},
"onnx": {
"description": "Enables ONNX frontend for reading models in ONNX format",
"dependencies": [
{
"name": "onnx",
"version>=": "1.16.2"
},
{
"name": "protobuf",
"version>=": "3.21.2"
},
{
"name": "protobuf",
"host": true,
"version>=": "3.21.2"
}
]
},
"paddle": {
"description": "Enables PaddlePaddle frontend for reading models in PaddlePaddle format",
"dependencies": [
{
"name": "protobuf",
"version>=": "3.21.2"
},
{
"name": "protobuf",
"host": true,
"version>=": "3.21.2"
}
]
},
"pytorch": {
"description": "Enables PyTorch frontend to convert models in PyTorch format"
},
"tensorflow": {
"description": "Enables TensorFlow frontend for reading models in TensorFlow format",
"dependencies": [
{
"name": "protobuf",
"version>=": "3.21.2"
},
{
"name": "protobuf",
"host": true,
"version>=": "3.21.2"
},
"snappy"
]
},
"tensorflow-lite": {
"description": "Enables TensorFlow Lite frontend for reading models in TensorFlow Lite format",
"dependencies": [
{
"name": "flatbuffers",
"version>=": "2.0.6"
},
{
"name": "flatbuffers",
"host": true,
"version>=": "2.0.6"
}
]
}
}
}