mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-20 04:39:00 +08:00
Feat: Add GPT-5.2 & pro (#11929)
### What problem does this PR solve? Feat: Add GPT-5.2 & pro ### Type of change - [x] New Feature (non-breaking change which adds functionality)
This commit is contained in:
@ -7,6 +7,20 @@
|
|||||||
"status": "1",
|
"status": "1",
|
||||||
"rank": "999",
|
"rank": "999",
|
||||||
"llm": [
|
"llm": [
|
||||||
|
{
|
||||||
|
"llm_name": "gpt-5.2-pro",
|
||||||
|
"tags": "LLM,CHAT,400k,IMAGE2TEXT",
|
||||||
|
"max_tokens": 400000,
|
||||||
|
"model_type": "chat",
|
||||||
|
"is_tools": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"llm_name": "gpt-5.2",
|
||||||
|
"tags": "LLM,CHAT,400k,IMAGE2TEXT",
|
||||||
|
"max_tokens": 400000,
|
||||||
|
"model_type": "chat",
|
||||||
|
"is_tools": true
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"llm_name": "gpt-5.1",
|
"llm_name": "gpt-5.1",
|
||||||
"tags": "LLM,CHAT,400k,IMAGE2TEXT",
|
"tags": "LLM,CHAT,400k,IMAGE2TEXT",
|
||||||
|
|||||||
@ -54,6 +54,7 @@ class SupportedLiteLLMProvider(StrEnum):
|
|||||||
MiniMax = "MiniMax"
|
MiniMax = "MiniMax"
|
||||||
DeerAPI = "DeerAPI"
|
DeerAPI = "DeerAPI"
|
||||||
GPUStack = "GPUStack"
|
GPUStack = "GPUStack"
|
||||||
|
OpenAI = "OpenAI"
|
||||||
|
|
||||||
|
|
||||||
FACTORY_DEFAULT_BASE_URL = {
|
FACTORY_DEFAULT_BASE_URL = {
|
||||||
@ -78,6 +79,7 @@ FACTORY_DEFAULT_BASE_URL = {
|
|||||||
SupportedLiteLLMProvider.ZHIPU_AI: "https://open.bigmodel.cn/api/paas/v4",
|
SupportedLiteLLMProvider.ZHIPU_AI: "https://open.bigmodel.cn/api/paas/v4",
|
||||||
SupportedLiteLLMProvider.MiniMax: "https://api.minimaxi.com/v1",
|
SupportedLiteLLMProvider.MiniMax: "https://api.minimaxi.com/v1",
|
||||||
SupportedLiteLLMProvider.DeerAPI: "https://api.deerapi.com/v1",
|
SupportedLiteLLMProvider.DeerAPI: "https://api.deerapi.com/v1",
|
||||||
|
SupportedLiteLLMProvider.OpenAI: "https://api.openai.com/v1",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -113,6 +115,8 @@ LITELLM_PROVIDER_PREFIX = {
|
|||||||
SupportedLiteLLMProvider.MiniMax: "openai/",
|
SupportedLiteLLMProvider.MiniMax: "openai/",
|
||||||
SupportedLiteLLMProvider.DeerAPI: "openai/",
|
SupportedLiteLLMProvider.DeerAPI: "openai/",
|
||||||
SupportedLiteLLMProvider.GPUStack: "openai/",
|
SupportedLiteLLMProvider.GPUStack: "openai/",
|
||||||
|
SupportedLiteLLMProvider.OpenAI: "openai/",
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
ChatModel = globals().get("ChatModel", {})
|
ChatModel = globals().get("ChatModel", {})
|
||||||
|
|||||||
@ -487,15 +487,6 @@ class Base(ABC):
|
|||||||
assert False, "Shouldn't be here."
|
assert False, "Shouldn't be here."
|
||||||
|
|
||||||
|
|
||||||
class GptTurbo(Base):
|
|
||||||
_FACTORY_NAME = "OpenAI"
|
|
||||||
|
|
||||||
def __init__(self, key, model_name="gpt-3.5-turbo", base_url="https://api.openai.com/v1", **kwargs):
|
|
||||||
if not base_url:
|
|
||||||
base_url = "https://api.openai.com/v1"
|
|
||||||
super().__init__(key, model_name, base_url, **kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class XinferenceChat(Base):
|
class XinferenceChat(Base):
|
||||||
_FACTORY_NAME = "Xinference"
|
_FACTORY_NAME = "Xinference"
|
||||||
|
|
||||||
@ -1227,6 +1218,7 @@ class LiteLLMBase(ABC):
|
|||||||
"MiniMax",
|
"MiniMax",
|
||||||
"DeerAPI",
|
"DeerAPI",
|
||||||
"GPUStack",
|
"GPUStack",
|
||||||
|
"OpenAI",
|
||||||
]
|
]
|
||||||
|
|
||||||
def __init__(self, key, model_name, base_url=None, **kwargs):
|
def __init__(self, key, model_name, base_url=None, **kwargs):
|
||||||
|
|||||||
Reference in New Issue
Block a user