diff --git a/conf/llm_factories.json b/conf/llm_factories.json index 90602c65f..d3b2dcc1c 100644 --- a/conf/llm_factories.json +++ b/conf/llm_factories.json @@ -283,20 +283,6 @@ "model_type": "chat", "is_tools": true }, - { - "llm_name": "glm-4.5", - "tags": "LLM,CHAT,131K", - "max_tokens": 131000, - "model_type": "chat", - "is_tools": true - }, - { - "llm_name": "deepseek-v3.1", - "tags": "LLM,CHAT,128k", - "max_tokens": 128000, - "model_type": "chat", - "is_tools": true - }, { "llm_name": "hunyuan-a13b-instruct", "tags": "LLM,CHAT,256k", @@ -338,6 +324,34 @@ "max_tokens": 262000, "model_type": "chat", "is_tools": true + }, + { + "llm_name": "deepseek-ocr", + "tags": "LLM,8k", + "max_tokens": 8000, + "model_type": "chat", + "is_tools": true + }, + { + "llm_name": "qwen3-235b-a22b-instruct-2507", + "tags": "LLM,CHAT,256k", + "max_tokens": 256000, + "model_type": "chat", + "is_tools": true + }, + { + "llm_name": "glm-4.6", + "tags": "LLM,CHAT,200k", + "max_tokens": 200000, + "model_type": "chat", + "is_tools": true + }, + { + "llm_name": "minimax-m2", + "tags": "LLM,CHAT,200k", + "max_tokens": 200000, + "model_type": "chat", + "is_tools": true } ] },