mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-08 20:42:30 +08:00
Updated UI (#4011)
### What problem does this PR solve? ### Type of change - [x] Documentation Update
This commit is contained in:
@ -366,10 +366,7 @@ The above is the content you need to summarize.`,
|
||||
topN: 'Top N',
|
||||
topNTip: `Not all chunks with similarity score above the 'similarity threshold' will be sent to the LLM. This selects 'Top N' chunks from the retrieved ones.`,
|
||||
variable: 'Variable',
|
||||
variableTip: `If you use dialog APIs, the varialbes might help you chat with your clients with different strategies.
|
||||
The variables are used to fill-in the 'System' part in prompt in order to give LLM a hint.
|
||||
The 'knowledge' is a very special variable which will be filled-in with the retrieved chunks.
|
||||
All the variables in 'System' should be curly bracketed.`,
|
||||
variableTip: `Variables can assist in developing more flexible strategies, particularly when you are using our chat assistant management APIs. These variables will be used by 'System' as part of the prompts for the LLM. The variable {knowledge} is a reserved special variable representing your selected knowledge base(s), and all variables should be enclosed in curly braces {}.`,
|
||||
add: 'Add',
|
||||
key: 'Key',
|
||||
optional: 'Optional',
|
||||
@ -381,15 +378,15 @@ The above is the content you need to summarize.`,
|
||||
improvise: 'Improvise',
|
||||
precise: 'Precise',
|
||||
balance: 'Balance',
|
||||
freedomTip: `'Precise' means the LLM will be conservative and answer your question cautiously. 'Improvise' means the you want LLM talk much and freely. 'Balance' is between cautiously and freely.`,
|
||||
freedomTip: `Set the freedom level to 'Precise' to strictly confine the LLM's response to your selected knowledge base(s). Choose 'Improvise' to grant the LLM greater freedom in its responses, which may lead to hallucinations. 'Balance' is an intermediate level; choose 'Balance' for more balanced responses.`,
|
||||
temperature: 'Temperature',
|
||||
temperatureMessage: 'Temperature is required',
|
||||
temperatureTip:
|
||||
'This parameter controls the randomness of predictions by the model. A lower temperature makes the model more confident in its responses, while a higher temperature makes it more creative and diverse.',
|
||||
`This parameter controls the randomness of the model's predictions. A lower temperature results in more conservative responses, while a higher temperature yields more creative and diverse responses.`,
|
||||
topP: 'Top P',
|
||||
topPMessage: 'Top P is required',
|
||||
topPTip:
|
||||
'Also known as “nucleus sampling,” this parameter sets a threshold to select a smaller set of words to sample from. It focuses on the most likely words, cutting off the less probable ones.',
|
||||
'Also known as "nucleus sampling", this parameter sets a threshold for selecting a smaller set of the most likely words to sample from, cutting off the less probable ones.',
|
||||
presencePenalty: 'Presence penalty',
|
||||
presencePenaltyMessage: 'Presence penalty is required',
|
||||
presencePenaltyTip:
|
||||
@ -401,7 +398,7 @@ The above is the content you need to summarize.`,
|
||||
maxTokens: 'Max tokens',
|
||||
maxTokensMessage: 'Max tokens is required',
|
||||
maxTokensTip:
|
||||
'This sets the maximum length of the model’s output, measured in the number of tokens (words or pieces of words).',
|
||||
`This sets the maximum length of the model's output, measured in the number of tokens (words or pieces of words). If disabled, you lift the maximum token limit, allowing the model to determine the number of tokens in its responses. Defaults to 512.`,
|
||||
maxTokensInvalidMessage: 'Please enter a valid number for Max Tokens.',
|
||||
maxTokensMinMessage: 'Max Tokens cannot be less than 0.',
|
||||
quote: 'Show quote',
|
||||
@ -456,7 +453,8 @@ The above is the content you need to summarize.`,
|
||||
profileDescription: 'Update your photo and personal details here.',
|
||||
maxTokens: 'Max Tokens',
|
||||
maxTokensMessage: 'Max Tokens is required',
|
||||
maxTokensTip: `This sets the maximum length of the model's output, measured in the number of tokens (words or pieces of words).`,
|
||||
maxTokensTip:
|
||||
`This sets the maximum length of the model's output, measured in the number of tokens (words or pieces of words). If disabled, you lift the maximum token limit, allowing the model to determine the number of tokens in its responses. Defaults to 512.`,
|
||||
maxTokensInvalidMessage: 'Please enter a valid number for Max Tokens.',
|
||||
maxTokensMinMessage: 'Max Tokens cannot be less than 0.',
|
||||
password: 'Password',
|
||||
|
||||
Reference in New Issue
Block a user