You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I created an application that uses the UAE-large-V1 model inside Transformers.js and was able to embed sentences in a browser without issues. The model would return a single vector for a single input:
When I hosted the model on Huggingface using their inference endpoint solution, it no longer works as expected. Instead of returning a single vector, it returns a variable length of 1024 dimension vectors.
Sample input:
{
"inputs": "Where are you"
}
This returns a list of lists of lists of numbers.
Is there a way to make hosted model return a single vector? And why does the the model act differently based on where it's hosted?
The text was updated successfully, but these errors were encountered:
I created an application that uses the UAE-large-V1 model inside Transformers.js and was able to embed sentences in a browser without issues. The model would return a single vector for a single input:
When I hosted the model on Huggingface using their inference endpoint solution, it no longer works as expected. Instead of returning a single vector, it returns a variable length of 1024 dimension vectors.
Sample input:
This returns a list of lists of lists of numbers.
Is there a way to make hosted model return a single vector? And why does the the model act differently based on where it's hosted?
The text was updated successfully, but these errors were encountered: