Support scenarios where inputs_embeds is a model input #664
Replies: 9 comments
-
Here is the
|
Beta Was this translation helpful? Give feedback.
-
eres313 currently Could you share some details around how you plan to retrieve the input embeddings for your use case before they can be passed into onnxruntime-genai as inputs to the model? Do you intend to do this outside of onnxruntime? |
Beta Was this translation helpful? Give feedback.
-
Yes, my input embeddings are fixed, I dont want to recompute them every time, so it will be outside the onnxruntime. |
Beta Was this translation helpful? Give feedback.
-
The model is taking |
Beta Was this translation helpful? Give feedback.
-
also have the same issue while running the phi-3 vision onnx |
Beta Was this translation helpful? Give feedback.
-
henrywang0314 running phi-3 vision onnx model should not give you that error. What version of onnxruntime-genai are you using? Could you share the output of |
Beta Was this translation helpful? Give feedback.
-
OrkhanHI We can add some work in onnxruntime-genai to support scenarios where the input is inputs_embeds. For now, this is not supported. |
Beta Was this translation helpful? Give feedback.
-
Same issue with phi-3-vision when following the instructions here: https://onnxruntime.ai/docs/genai/tutorials/phi3-v.html pip list | grep onnxruntime-genai python phi3v.py -m cpu-int4-rtn-block-32-acc-level-4 |
Beta Was this translation helpful? Give feedback.
-
Please use the release candidate 0.3.0-rc2 Use pip install --pre onnxruntime-genai |
Beta Was this translation helpful? Give feedback.
-
Hi while converting the model with
extra_options = {"exclude_embeds": 1}
I am getting below error while loading the modelBeta Was this translation helpful? Give feedback.
All reactions