mirror of
https://github.com/infiniflow/ragflow.git
synced 2026-02-02 08:35:08 +08:00
### What problem does this PR solve? Fix: Fixed the styling and logic issues on the model provider page ### Type of change - [x] Bug Fix (non-breaking change which fixes an issue)
273 lines
8.2 KiB
TypeScript
273 lines
8.2 KiB
TypeScript
import { LLMFactory } from '@/constants/llm';
|
|
import { useTranslate } from '@/hooks/common-hooks';
|
|
import { IModalProps } from '@/interfaces/common';
|
|
import { IAddLlmRequestBody } from '@/interfaces/request/llm';
|
|
import {
|
|
Flex,
|
|
Form,
|
|
Input,
|
|
InputNumber,
|
|
Modal,
|
|
Select,
|
|
Space,
|
|
Switch,
|
|
} from 'antd';
|
|
import omit from 'lodash/omit';
|
|
import { useEffect } from 'react';
|
|
|
|
type FieldType = IAddLlmRequestBody & {
|
|
vision: boolean;
|
|
provider_order?: string;
|
|
};
|
|
|
|
const { Option } = Select;
|
|
|
|
const llmFactoryToUrlMap = {
|
|
[LLMFactory.Ollama]:
|
|
'https://github.com/infiniflow/ragflow/blob/main/docs/guides/models/deploy_local_llm.mdx',
|
|
[LLMFactory.Xinference]:
|
|
'https://inference.readthedocs.io/en/latest/user_guide',
|
|
[LLMFactory.ModelScope]:
|
|
'https://www.modelscope.cn/docs/model-service/API-Inference/intro',
|
|
[LLMFactory.LocalAI]: 'https://localai.io/docs/getting-started/models/',
|
|
[LLMFactory.LMStudio]: 'https://lmstudio.ai/docs/basics',
|
|
[LLMFactory.OpenAiAPICompatible]:
|
|
'https://platform.openai.com/docs/models/gpt-4',
|
|
[LLMFactory.TogetherAI]: 'https://docs.together.ai/docs/deployment-options',
|
|
[LLMFactory.Replicate]: 'https://replicate.com/docs/topics/deployments',
|
|
[LLMFactory.OpenRouter]: 'https://openrouter.ai/docs',
|
|
[LLMFactory.HuggingFace]:
|
|
'https://huggingface.co/docs/text-embeddings-inference/quick_tour',
|
|
[LLMFactory.GPUStack]: 'https://docs.gpustack.ai/latest/quickstart',
|
|
[LLMFactory.VLLM]: 'https://docs.vllm.ai/en/latest/',
|
|
[LLMFactory.TokenPony]: 'https://docs.tokenpony.cn/#/',
|
|
};
|
|
type LlmFactory = keyof typeof llmFactoryToUrlMap;
|
|
|
|
const OllamaModal = ({
|
|
visible,
|
|
hideModal,
|
|
onOk,
|
|
loading,
|
|
llmFactory,
|
|
editMode = false,
|
|
initialValues,
|
|
}: IModalProps<IAddLlmRequestBody> & {
|
|
llmFactory: string;
|
|
editMode?: boolean;
|
|
initialValues?: Partial<IAddLlmRequestBody>;
|
|
}) => {
|
|
const [form] = Form.useForm<FieldType>();
|
|
|
|
const { t } = useTranslate('setting');
|
|
|
|
const handleOk = async () => {
|
|
const values = await form.validateFields();
|
|
const modelType =
|
|
values.model_type === 'chat' && values.vision
|
|
? 'image2text'
|
|
: values.model_type;
|
|
|
|
const data = {
|
|
...omit(values, ['vision']),
|
|
model_type: modelType,
|
|
llm_factory: llmFactory,
|
|
max_tokens: values.max_tokens,
|
|
};
|
|
console.info(data);
|
|
|
|
onOk?.(data);
|
|
};
|
|
|
|
const handleKeyDown = async (e: React.KeyboardEvent) => {
|
|
if (e.key === 'Enter') {
|
|
await handleOk();
|
|
}
|
|
};
|
|
|
|
useEffect(() => {
|
|
if (visible && editMode && initialValues) {
|
|
const formValues = {
|
|
llm_name: initialValues.llm_name,
|
|
model_type: initialValues.model_type,
|
|
api_base: initialValues.api_base,
|
|
max_tokens: initialValues.max_tokens || 8192,
|
|
api_key: '',
|
|
...initialValues,
|
|
};
|
|
form.setFieldsValue(formValues);
|
|
} else if (visible && !editMode) {
|
|
form.resetFields();
|
|
}
|
|
}, [visible, editMode, initialValues, form]);
|
|
|
|
const url =
|
|
llmFactoryToUrlMap[llmFactory as LlmFactory] ||
|
|
'https://github.com/infiniflow/ragflow/blob/main/docs/guides/models/deploy_local_llm.mdx';
|
|
const optionsMap = {
|
|
[LLMFactory.HuggingFace]: [
|
|
{ value: 'embedding', label: 'embedding' },
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'rerank', label: 'rerank' },
|
|
],
|
|
[LLMFactory.LMStudio]: [
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'embedding', label: 'embedding' },
|
|
{ value: 'image2text', label: 'image2text' },
|
|
],
|
|
[LLMFactory.Xinference]: [
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'embedding', label: 'embedding' },
|
|
{ value: 'rerank', label: 'rerank' },
|
|
{ value: 'image2text', label: 'image2text' },
|
|
{ value: 'speech2text', label: 'sequence2text' },
|
|
{ value: 'tts', label: 'tts' },
|
|
],
|
|
[LLMFactory.ModelScope]: [{ value: 'chat', label: 'chat' }],
|
|
[LLMFactory.GPUStack]: [
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'embedding', label: 'embedding' },
|
|
{ value: 'rerank', label: 'rerank' },
|
|
{ value: 'speech2text', label: 'sequence2text' },
|
|
{ value: 'tts', label: 'tts' },
|
|
],
|
|
[LLMFactory.OpenRouter]: [
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'image2text', label: 'image2text' },
|
|
],
|
|
Default: [
|
|
{ value: 'chat', label: 'chat' },
|
|
{ value: 'embedding', label: 'embedding' },
|
|
{ value: 'rerank', label: 'rerank' },
|
|
{ value: 'image2text', label: 'image2text' },
|
|
],
|
|
};
|
|
const getOptions = (factory: string) => {
|
|
return optionsMap[factory as keyof typeof optionsMap] || optionsMap.Default;
|
|
};
|
|
return (
|
|
<Modal
|
|
title={
|
|
editMode
|
|
? t('editLlmTitle', { name: llmFactory })
|
|
: t('addLlmTitle', { name: llmFactory })
|
|
}
|
|
open={visible}
|
|
onOk={handleOk}
|
|
onCancel={hideModal}
|
|
okButtonProps={{ loading }}
|
|
footer={(originNode: React.ReactNode) => {
|
|
return (
|
|
<Flex justify={'space-between'}>
|
|
<a href={url} target="_blank" rel="noreferrer">
|
|
{t('ollamaLink', { name: llmFactory })}
|
|
</a>
|
|
<Space>{originNode}</Space>
|
|
</Flex>
|
|
);
|
|
}}
|
|
>
|
|
<Form
|
|
name="basic"
|
|
style={{ maxWidth: 600 }}
|
|
autoComplete="off"
|
|
layout={'vertical'}
|
|
form={form}
|
|
>
|
|
<Form.Item<FieldType>
|
|
label={t('modelType')}
|
|
name="model_type"
|
|
initialValue={'embedding'}
|
|
rules={[{ required: true, message: t('modelTypeMessage') }]}
|
|
>
|
|
<Select placeholder={t('modelTypeMessage')}>
|
|
{getOptions(llmFactory).map((option) => (
|
|
<Option key={option.value} value={option.value}>
|
|
{option.label}
|
|
</Option>
|
|
))}
|
|
</Select>
|
|
</Form.Item>
|
|
<Form.Item<FieldType>
|
|
label={t(llmFactory === 'Xinference' ? 'modelUid' : 'modelName')}
|
|
name="llm_name"
|
|
rules={[{ required: true, message: t('modelNameMessage') }]}
|
|
>
|
|
<Input
|
|
placeholder={t('modelNameMessage')}
|
|
onKeyDown={handleKeyDown}
|
|
/>
|
|
</Form.Item>
|
|
<Form.Item<FieldType>
|
|
label={t('addLlmBaseUrl')}
|
|
name="api_base"
|
|
rules={[{ required: true, message: t('baseUrlNameMessage') }]}
|
|
>
|
|
<Input
|
|
placeholder={t('baseUrlNameMessage')}
|
|
onKeyDown={handleKeyDown}
|
|
/>
|
|
</Form.Item>
|
|
<Form.Item<FieldType>
|
|
label={t('apiKey')}
|
|
name="api_key"
|
|
rules={[{ required: false, message: t('apiKeyMessage') }]}
|
|
>
|
|
<Input placeholder={t('apiKeyMessage')} onKeyDown={handleKeyDown} />
|
|
</Form.Item>
|
|
<Form.Item<FieldType>
|
|
label={t('maxTokens')}
|
|
name="max_tokens"
|
|
rules={[
|
|
{ required: true, message: t('maxTokensMessage') },
|
|
{
|
|
type: 'number',
|
|
message: t('maxTokensInvalidMessage'),
|
|
},
|
|
({}) => ({
|
|
validator(_, value) {
|
|
if (value < 0) {
|
|
return Promise.reject(new Error(t('maxTokensMinMessage')));
|
|
}
|
|
return Promise.resolve();
|
|
},
|
|
}),
|
|
]}
|
|
>
|
|
<InputNumber
|
|
placeholder={t('maxTokensTip')}
|
|
style={{ width: '100%' }}
|
|
onKeyDown={handleKeyDown}
|
|
/>
|
|
</Form.Item>
|
|
{llmFactory === LLMFactory.OpenRouter && (
|
|
<Form.Item<FieldType>
|
|
label="Provider Order"
|
|
name="provider_order"
|
|
tooltip="Comma-separated provider list, e.g. Groq,Fireworks"
|
|
rules={[]}
|
|
>
|
|
<Input placeholder="Groq,Fireworks" onKeyDown={handleKeyDown} />
|
|
</Form.Item>
|
|
)}
|
|
|
|
<Form.Item noStyle dependencies={['model_type']}>
|
|
{({ getFieldValue }) =>
|
|
getFieldValue('model_type') === 'chat' && (
|
|
<Form.Item
|
|
label={t('vision')}
|
|
valuePropName="checked"
|
|
name={'vision'}
|
|
>
|
|
<Switch />
|
|
</Form.Item>
|
|
)
|
|
}
|
|
</Form.Item>
|
|
</Form>
|
|
</Modal>
|
|
);
|
|
};
|
|
|
|
export default OllamaModal;
|