We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 670f63c commit 98bf067Copy full SHA for 98bf067
multimode_llm_adapter.py
@@ -32,7 +32,7 @@ def initialise(self) -> None:
32
print("*** Multi-modal LLM using CPU only: This module requires > 16Gb RAM")
33
self.inference_device = "CPU"
34
self.device = "cpu"
35
- self.inference_library = "ONNX"
+ self.inference_library = "ONNX/DML" if self.system_info.os == "Windows" else "ONNX"
36
self.model_repo = "microsoft/Phi-3-vision-128k-instruct-onnx-cpu"
37
self.model_filename = None # "Phi-3-vision-128k-instruct.gguf"
38
self.models_dir = "cpu-int4-rtn-block-32-acc-level-4"
0 commit comments