How is genAI compatible with Onnxruntime API? #824
elephantpanda
started this conversation in
General
Replies: 1 comment
-
I have a similar question regarding configuring Cuda EP. As discussed here https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#requirements, there are some aspects of CUDA EP that need tweaking for different models, but the current
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
The GenAI is very nice to get things up and running, and I installed the c# DirectML version from nuget. Unfortunately it is not a very low-level API so it's hard to do things like manipulate the input/output token vectors etc. Is it compatible at all with the Onnx managed runtime for c# or is it entirely it's own thing? I feel like I may have to rewrite most of it using the lower level Onnx Runtime c# api.
For example there is the function processor.ProcessImages which creates a namedTensors object. But there's not much I can do with that except feed it into the SetInputs function. Ideally I would like to inspect it, or use the tokenizer to decode it or things like that.
Beta Was this translation helpful? Give feedback.
All reactions