diff --git a/conf/llm_factories.json b/conf/llm_factories.json index 681480ca2..324cb0b05 100644 --- a/conf/llm_factories.json +++ b/conf/llm_factories.json @@ -2973,13 +2973,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "Pro/deepseek-ai/DeepSeek-V3-1226", - "tags": "LLM,CHAT,64k", - "max_tokens": 64000, - "model_type": "chat", - "is_tools": true - }, { "llm_name": "Pro/deepseek-ai/DeepSeek-V3.1", "tags": "LLM,CHAT,160k", @@ -3022,20 +3015,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": true - }, { "llm_name": "deepseek-ai/DeepSeek-V2.5", "tags": "LLM,CHAT,32k", @@ -3092,13 +3071,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "THUDM/chatglm3-6b", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": false - }, { "llm_name": "Pro/THUDM/glm-4-9b-chat", "tags": "LLM,CHAT,128k", @@ -3120,13 +3092,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "Qwen/QwQ-32B-Preview", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": false - }, { "llm_name": "Qwen/Qwen2.5-Coder-32B-Instruct", "tags": "LLM,CHAT,32k", @@ -3190,13 +3155,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "internlm/internlm2_5-20b-chat", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": true - }, { "llm_name": "internlm/internlm2_5-7b-chat", "tags": "LLM,CHAT,32k", @@ -3232,13 +3190,6 @@ "model_type": "chat", "is_tools": false }, - { - "llm_name": "Pro/Qwen/Qwen2-1.5B-Instruct", - "tags": "LLM,CHAT,32k", - "max_tokens": 32000, - "model_type": "chat", - "is_tools": false - }, { "llm_name": "BAAI/bge-m3", "tags": "LLM,EMBEDDING,8k",