mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-08 20:42:30 +08:00
Support downloading models from ModelScope Community. (#5073)
This PR supports downloading models from ModelScope. The main modifications are as follows: -New Feature (non-breaking change which adds functionality) -Documentation Update --------- Co-authored-by: Kevin Hu <kevinhu.sh@gmail.com>
This commit is contained in:
@ -32,6 +32,7 @@ export const LocalLlmFactories = [
|
||||
'OpenRouter',
|
||||
'HuggingFace',
|
||||
'GPUStack',
|
||||
'ModelScope',
|
||||
];
|
||||
|
||||
export enum TenantRole {
|
||||
|
||||
@ -21,6 +21,7 @@ const llmFactoryToUrlMap = {
|
||||
Ollama:
|
||||
'https://github.com/infiniflow/ragflow/blob/main/docs/guides/deploy_local_llm.mdx',
|
||||
Xinference: 'https://inference.readthedocs.io/en/latest/user_guide',
|
||||
ModelScope: 'https://www.modelscope.cn/docs/model-service/API-Inference/intro',
|
||||
LocalAI: 'https://localai.io/docs/getting-started/models/',
|
||||
'LM-Studio': 'https://lmstudio.ai/docs/basics',
|
||||
'OpenAI-API-Compatible': 'https://platform.openai.com/docs/models/gpt-4',
|
||||
@ -77,6 +78,9 @@ const OllamaModal = ({
|
||||
{ value: 'speech2text', label: 'sequence2text' },
|
||||
{ value: 'tts', label: 'tts' },
|
||||
],
|
||||
ModelScope: [
|
||||
{ value: 'chat', label: 'chat' },
|
||||
],
|
||||
GPUStack: [
|
||||
{ value: 'chat', label: 'chat' },
|
||||
{ value: 'embedding', label: 'embedding' },
|
||||
|
||||
Reference in New Issue
Block a user