From 85480f6292b4c68baa4af88b4e1f2f3802dab2b9 Mon Sep 17 00:00:00 2001 From: zhou <269138475@qq.com> Date: Fri, 21 Mar 2025 15:25:48 +0800 Subject: [PATCH] Fix: the error of Ollama embeddings interface returning "500 Internal Server Error" (#6350) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit ### What problem does this PR solve? Fix the error where the Ollama embeddings interface returns a “500 Internal Server Error” when using models such as xiaobu-embedding-v2 for embedding. ### Type of change - [x] Bug Fix (non-breaking change which fixes an issue) --- rag/llm/embedding_model.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/rag/llm/embedding_model.py b/rag/llm/embedding_model.py index 17bb84ef5..dba11b7db 100644 --- a/rag/llm/embedding_model.py +++ b/rag/llm/embedding_model.py @@ -260,14 +260,16 @@ class OllamaEmbed(Base): tks_num = 0 for txt in texts: res = self.client.embeddings(prompt=txt, - model=self.model_name) + model=self.model_name, + options={"use_mmap": True}) arr.append(res["embedding"]) tks_num += 128 return np.array(arr), tks_num def encode_queries(self, text): res = self.client.embeddings(prompt=text, - model=self.model_name) + model=self.model_name, + options={"use_mmap": True}) return np.array(res["embedding"]), 128 @@ -834,4 +836,4 @@ class GPUStackEmbed(OpenAIEmbed): print(key,base_url) self.client = OpenAI(api_key=key, base_url=base_url) - self.model_name = model_name \ No newline at end of file + self.model_name = model_name