How is genAI compatible with Onnxruntime API? #824
elephantpanda
started this conversation in
General
Replies: 1 comment
-
I have a similar question regarding configuring Cuda EP. As discussed here https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#requirements, there are some aspects of CUDA EP that need tweaking for different models, but the current
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The GenAI is very nice to get things up and running, and I installed the c# DirectML version from nuget. Unfortunately it is not a very low-level API so it's hard to do things like manipulate the input/output token vectors etc. Is it compatible at all with the Onnx managed runtime for c# or is it entirely it's own thing? I feel like I may have to rewrite most of it using the lower level Onnx Runtime c# api.
For example there is the function processor.ProcessImages which creates a namedTensors object. But there's not much I can do with that except feed it into the SetInputs function. Ideally I would like to inspect it, or use the tokenizer to decode it or things like that.
Beta Was this translation helpful? Give feedback.
All reactions