diff --git a/integrations/gradient/src/haystack_integrations/components/embedders/gradient/gradient_document_embedder.py b/integrations/gradient/src/haystack_integrations/components/embedders/gradient/gradient_document_embedder.py index 0ecfcbb98..34fc1f87f 100644 --- a/integrations/gradient/src/haystack_integrations/components/embedders/gradient/gradient_document_embedder.py +++ b/integrations/gradient/src/haystack_integrations/components/embedders/gradient/gradient_document_embedder.py @@ -127,6 +127,3 @@ def run(self, documents: List[Document]): doc.embedding = embedding return {"documents": documents} - - -# TEST! TO BE REMOVED diff --git a/integrations/instructor_embedders/src/haystack_integrations/components/embedders/instructor_embedders/embedding_backend/instructor_backend.py b/integrations/instructor_embedders/src/haystack_integrations/components/embedders/instructor_embedders/embedding_backend/instructor_backend.py index a22eaff32..efe35e9b7 100644 --- a/integrations/instructor_embedders/src/haystack_integrations/components/embedders/instructor_embedders/embedding_backend/instructor_backend.py +++ b/integrations/instructor_embedders/src/haystack_integrations/components/embedders/instructor_embedders/embedding_backend/instructor_backend.py @@ -42,6 +42,3 @@ def __init__( def embed(self, data: List[List[str]], **kwargs) -> List[List[float]]: embeddings = self.model.encode(data, **kwargs).tolist() return embeddings - - -# TEST! TO BE REMOVED