diff --git a/rag/llm/embedding_model.py b/rag/llm/embedding_model.py index 98f9bcbc1..9774e7dd9 100644 --- a/rag/llm/embedding_model.py +++ b/rag/llm/embedding_model.py @@ -352,8 +352,7 @@ class FastEmbed(DefaultEmbedding): # Using the internal tokenizer to encode the texts and get the total # number of tokens encoding = self._model.model.tokenizer.encode(text) - embedding = next(self._model.query_embed(text)).tolist() - + embedding = next(self._model.query_embed(text)) return np.array(embedding), len(encoding.ids)