Skip to content

WIP: add OpenVINO support#320

Draft
zhol0777 wants to merge 3 commits intoNeptuneHub:mainfrom
zhol0777:openvino_support
Draft

WIP: add OpenVINO support#320
zhol0777 wants to merge 3 commits intoNeptuneHub:mainfrom
zhol0777:openvino_support

Conversation

@zhol0777
Copy link

@zhol0777 zhol0777 commented Feb 7, 2026

Allows for accelerated inferencing on Intel hardware. To allow for this, a few small changes are made:

  • addition of requirements/openvino.txt, which just lets pip install the OpenVINO runtime
  • util.provider, to prevent code-reuse with
    • get_available_providers, which filters out supported ONNXRuntime providers
  • tasks.analysis.get_provider_options, which prevents code-reuse
  • modifications to Dockerfile, to allow for that optional package installation described earlier

Allows for accelerated inferencing on Intel hardware. To allow for
this, a few small changes are made:

* addition of requirements/openvino.txt, which just lets pip install
  the OpenVINO runtime
* util.provider, to prevent code-reuse with
  * get_available_providers, which filters out supported ONNXRuntime
    providers
* tasks.analysis.get_provider_options, which prevents code-reuse
* modifications to Dockerfile, to allow for that optional package
  installation described earlier

# --- OpenVINO Acceleration ---
RENDER_GID= # render group ID (use `stat -c "%g" /dev/dri/renderD128` on host to verify)
OPENVINO_CONFIG_JSON_PATH= # path to have openvino load config https://onnxruntime.ai/docs/execution-providers/OpenVINO-ExecutionProvider.html#load_config No newline at end of file
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe I could add something to the docs about what normal config.json would ook like

CLAP_ENABLED=true

# --- OpenVINO Acceleration ---
RENDER_GID= # render group ID (use `stat -c "%g" /dev/dri/renderD128` on host to verify)
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

gotta add to config.py still, as well as docker-compose reference files

Comment on lines +1329 to +1330
def get_provider_options(cuda_do_copy_in_default_stream: bool = False,
cuda_conv_algo_search_mode: str = 'EXHAUSTIVE') -> list[tuple[str, dict[str, Any]]]:
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

unsure if this is the right place to put the function, or in new util.provider

available_providers = provider.get_available_providers()
if 'OpenVINOExecutionProvider' in available_providers:
vino_options = {
'device_type': 'AUTO',
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

might need to try other device_type... need to play with AUTO/HETERO/MULTI. am only testing on an N100 trying to get the IGP to do help out with tagging my library

ref: https://onnxruntime.ai/docs/execution-providers/OpenVINO-ExecutionProvider.html#device_type

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

setting to 'MULTI:GPU,CPU' does make the process appear in intel_gpu_top. kinda curious to perf if i just force GPU only and turn off CPU execution provider

@zhol0777
Copy link
Author

zhol0777 commented Feb 7, 2026

also needs to update docker-compose reference files and github actions build, just wanted to test the water first

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant