mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-08 20:42:30 +08:00
Feat: add LLM provider PPIO (#5013)
### What problem does this PR solve? Add a LLM provider: PPIO ### Type of change - [x] New Feature (non-breaking change which adds functionality) - [x] Documentation Update
This commit is contained in:
@ -2789,6 +2789,128 @@
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"name": "PPIO",
|
||||
"logo": "",
|
||||
"tags": "LLM",
|
||||
"status": "1",
|
||||
"llm": [
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1/community",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-v3/community",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-v3",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1-distill-llama-70b",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1-distill-qwen-32b",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1-distill-qwen-14b",
|
||||
"tags": "LLM,CHAT,64k",
|
||||
"max_tokens": 64000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "deepseek/deepseek-r1-distill-llama-8b",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "qwen/qwen-2.5-72b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "qwen/qwen-2-vl-72b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "meta-llama/llama-3.2-3b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "qwen/qwen2.5-32b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32000,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "baichuan/baichuan2-13b-chat",
|
||||
"tags": "LLM,CHAT,14k",
|
||||
"max_tokens": 14336,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "meta-llama/llama-3.1-70b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "meta-llama/llama-3.1-8b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "01-ai/yi-1.5-34b-chat",
|
||||
"tags": "LLM,CHAT,16k",
|
||||
"max_tokens": 16384,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "01-ai/yi-1.5-9b-chat",
|
||||
"tags": "LLM,CHAT,16k",
|
||||
"max_tokens": 16384,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "thudm/glm-4-9b-chat",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
},
|
||||
{
|
||||
"llm_name": "qwen/qwen-2-7b-instruct",
|
||||
"tags": "LLM,CHAT,32k",
|
||||
"max_tokens": 32768,
|
||||
"model_type": "chat"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"name": "01.AI",
|
||||
"logo": "",
|
||||
|
||||
Reference in New Issue
Block a user