Feat: add LLM provider PPIO (#5013)

### What problem does this PR solve?

Add a LLM provider: PPIO

### Type of change

- [x] New Feature (non-breaking change which adds functionality)
- [x] Documentation Update
This commit is contained in:
saikidev
2025-02-17 12:03:26 +08:00
committed by GitHub
parent 88daa349f9
commit d2929e432e
8 changed files with 138 additions and 1 deletions

View File

@ -2789,6 +2789,128 @@
}
]
},
{
"name": "PPIO",
"logo": "",
"tags": "LLM",
"status": "1",
"llm": [
{
"llm_name": "deepseek/deepseek-r1/community",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-v3/community",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-r1",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-v3",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-r1-distill-llama-70b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-r1-distill-qwen-32b",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-r1-distill-qwen-14b",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-r1-distill-llama-8b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-2.5-72b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-2-vl-72b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3.2-3b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen2.5-32b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "baichuan/baichuan2-13b-chat",
"tags": "LLM,CHAT,14k",
"max_tokens": 14336,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3.1-70b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3.1-8b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-1.5-34b-chat",
"tags": "LLM,CHAT,16k",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-1.5-9b-chat",
"tags": "LLM,CHAT,16k",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "thudm/glm-4-9b-chat",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-2-7b-instruct",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
}
]
},
{
"name": "01.AI",
"logo": "",