`provider_options` for running on GPU

#3
by cecheta - opened

The GPU microsoft/Phi-3.5-mini-instruct-onnx model was recently updated, however provider_options in genai_config.json is now set to []. When trying to run the model using the package onnxruntime-genai-cuda, it now seems to run on CPU, even if the device has GPU.

The provider_options have been intentionally left blank since the GPU ONNX model can run on both CUDA and DirectML. You can add back the execution provider information in your downloadedgenai_config.json file or you can set that information at runtime (note thatargs.provider can be cpu, cuda, or dml).

kvaishnavi changed discussion status to closed

Sign up or log in to comment