diff --git a/conf/llm_factories.json b/conf/llm_factories.json index fda846621..549e4aab3 100644 --- a/conf/llm_factories.json +++ b/conf/llm_factories.json @@ -909,184 +909,7 @@ "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", - "llm": [ - { - "llm_name": "deepseek.r1-v1:0", - "tags": "LLM,CHAT,128k", - "max_tokens": 128000, - "model_type": "chat", - "is_tools": false - }, - { - "llm_name": "ai21.j2-ultra-v1", - "tags": "LLM,CHAT,8k", - "max_tokens": 8191, - "model_type": "chat" - }, - { - "llm_name": "ai21.j2-mid-v1", - "tags": "LLM,CHAT,8k", - "max_tokens": 8191, - "model_type": "chat" - }, - { - "llm_name": "cohere.command-text-v14", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "cohere.command-light-text-v14", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "cohere.command-r-v1:0", - "tags": "LLM,CHAT,128k", - "max_tokens": 131072, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "cohere.command-r-plus-v1:0", - "tags": "LLM,CHAT,128k", - "max_tokens": 128000, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "anthropic.claude-v2", - "tags": "LLM,CHAT,100k", - "max_tokens": 102400, - "model_type": "chat" - }, - { - "llm_name": "anthropic.claude-v2:1", - "tags": "LLM,CHAT,200k", - "max_tokens": 204800, - "model_type": "chat" - }, - { - "llm_name": "anthropic.claude-3-sonnet-20240229-v1:0", - "tags": "LLM,CHAT,200k", - "max_tokens": 204800, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "anthropic.claude-3-5-sonnet-20240620-v1:0", - "tags": "LLM,CHAT,200k", - "max_tokens": 204800, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "anthropic.claude-3-haiku-20240307-v1:0", - "tags": "LLM,CHAT,200k", - "max_tokens": 204800, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "anthropic.claude-3-opus-20240229-v1:0", - "tags": "LLM,CHAT,200k", - "max_tokens": 204800, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "anthropic.claude-instant-v1", - "tags": "LLM,CHAT,100k", - "max_tokens": 102400, - "model_type": "chat" - }, - { - "llm_name": "amazon.titan-text-express-v1", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "amazon.titan-text-premier-v1:0", - "tags": "LLM,CHAT,32k", - "max_tokens": 32768, - "model_type": "chat" - }, - { - "llm_name": "amazon.titan-text-lite-v1", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "meta.llama2-13b-chat-v1", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "meta.llama2-70b-chat-v1", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "meta.llama3-8b-instruct-v1:0", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "meta.llama3-70b-instruct-v1:0", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "mistral.mistral-7b-instruct-v0:2", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "mistral.mixtral-8x7b-instruct-v0:1", - "tags": "LLM,CHAT,4k", - "max_tokens": 4096, - "model_type": "chat" - }, - { - "llm_name": "mistral.mistral-large-2402-v1:0", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "mistral.mistral-small-2402-v1:0", - "tags": "LLM,CHAT,8k", - "max_tokens": 8192, - "model_type": "chat" - }, - { - "llm_name": "amazon.titan-embed-text-v2:0", - "tags": "TEXT EMBEDDING", - "max_tokens": 8192, - "model_type": "embedding" - }, - { - "llm_name": "cohere.embed-english-v3", - "tags": "TEXT EMBEDDING", - "max_tokens": 2048, - "model_type": "embedding" - }, - { - "llm_name": "cohere.embed-multilingual-v3", - "tags": "TEXT EMBEDDING", - "max_tokens": 2048, - "model_type": "embedding" - } - ] + "llm": [] }, { "name": "Gemini", diff --git a/rag/llm/chat_model.py b/rag/llm/chat_model.py index 0ed660430..f00915dd2 100644 --- a/rag/llm/chat_model.py +++ b/rag/llm/chat_model.py @@ -234,6 +234,7 @@ class AzureChat(Base): def __init__(self, key, model_name, **kwargs): api_key = json.loads(key).get("api_key", "") api_version = json.loads(key).get("api_version", "2024-02-01") + super().__init__(key, model_name, kwargs["base_url"]) self.client = AzureOpenAI(api_key=api_key, azure_endpoint=kwargs["base_url"], api_version=api_version) self.model_name = model_name @@ -974,6 +975,7 @@ class LmStudioChat(Base): raise ValueError("Local llm url cannot be None") if base_url.split("/")[-1] != "v1": base_url = os.path.join(base_url, "v1") + super().__init__(key, model_name, base_url) self.client = OpenAI(api_key="lm-studio", base_url=base_url) self.model_name = model_name