feat: llm show vars

This commit is contained in:
Joel
2024-04-01 11:48:42 +08:00
parent e03367a188
commit 705d765a71
3 changed files with 40 additions and 10 deletions

View File

@ -1,5 +1,6 @@
import { useCallback, useEffect, useRef, useState } from 'react'
import produce from 'immer'
import { flatten, uniqBy } from 'lodash'
import useVarList from '../_base/hooks/use-var-list'
import { VarType } from '../../types'
import type { Memory, ValueSelector, Var } from '../../types'
@ -7,7 +8,9 @@ import { useStore } from '../../store'
import {
useIsChatMode,
useNodesReadOnly,
useWorkflow,
} from '../../hooks'
import { getNodeInfoById } from '../_base/components/variable/utils'
import type { LLMNodeType } from './types'
import { Resolution } from '@/types/app'
import { useModelListAndDefaultModelAndCurrentProviderAndModel, useTextGenerationCurrentProviderAndModelAndModelList } from '@/app/components/header/account-setting/model-provider-page/hooks'
@ -16,11 +19,14 @@ import useNodeCrud from '@/app/components/workflow/nodes/_base/hooks/use-node-cr
import useOneStepRun from '@/app/components/workflow/nodes/_base/hooks/use-one-step-run'
import type { PromptItem } from '@/models/debug'
import { RETRIEVAL_OUTPUT_STRUCT } from '@/app/components/workflow/constants'
import { checkHasContextBlock, checkHasHistoryBlock, checkHasQueryBlock } from '@/app/components/base/prompt-editor/constants'
import { checkHasContextBlock, checkHasHistoryBlock, checkHasQueryBlock, getInputVars } from '@/app/components/base/prompt-editor/constants'
const useConfig = (id: string, payload: LLMNodeType) => {
const { nodesReadOnly: readOnly } = useNodesReadOnly()
const isChatMode = useIsChatMode()
const { getBeforeNodesInSameBranch } = useWorkflow()
const availableNodes = getBeforeNodesInSameBranch(id)
const defaultConfig = useStore(s => s.nodesDefaultConfigs)[payload.type]
const [defaultRolePrefix, setDefaultRolePrefix] = useState<{ user: string; assistant: string }>({ user: '', assistant: '' })
@ -262,8 +268,32 @@ const useConfig = (id: string, payload: LLMNodeType) => {
'#files#': newFiles,
})
}, [runInputData, setRunInputData])
const variables = (() => {
let valueSelectors: ValueSelector[] = []
if (isChatModel) {
valueSelectors = flatten(
(inputs.prompt_template as PromptItem[])
.map(item => getInputVars(item.text)),
)
}
else {
valueSelectors = getInputVars((inputs.prompt_template as PromptItem).text)
}
const varInputs = toVarInputs(inputs.variables)
const variables = uniqBy(valueSelectors, item => item.join('.')).map((item) => {
const varInfo = getNodeInfoById(availableNodes, item[0])?.data
const variable = [...item]
variable[0] = varInfo?.title || availableNodes[0]?.data.title // default start node title
return {
variable: variable.join('/'),
value_selector: item,
}
})
return variables
})()
const varInputs = toVarInputs(variables)
return {
readOnly,