LLM Module
Most fields are compatible with OpenAI API.
LLMConfig
Field's Name | JSON Type (Required/Optional) | Description | Example |
---|---|---|---|
model | "gpt-35-turbo-16k" | "gpt-4-1116-preview" (Required) | Determines the OpenAI language model to be used in LLMConfig. |
temperature | number (Optional) | Controls randomness in the LLMConfig model's responses. | 0.6 |
top_p | number (Optional) | Controls diversity via nucleus sampling in LLMConfig. | 0.8 |
max_tokens | number (Optional) | Determines the maximum length of the model’s response in LLMConfig. | 150 |
presence_penalty | number (Optional) | Influences the likelihood of the model to talk about new topics in LLMConfig. | 0.6 |
frequency_penalty | number (Optional) | Controls how often the model makes use of infrequent words in LLMConfig. | 0.5 |
memory | MemoryItem[] (Optional) | Stores information acquired across multiple rounds of dialog in LLMConfig. | [{ role: 'user', content: 'Hello World' }] |
need_memory | boolean (Optional) | Determines if memory usage is required in LLMConfig. | true |
system_prompt | string | Expression (Required) | Provides system prompt for the user in LLMConfig. |
user_prompt | string | Expression (Required) | Provides user prompt for the system in LLMConfig. |
output_name | string (Required) | Determines the name of the module output in LLMConfig, defaults to 'reply'. | "reply" |
MemoryItem
Field's Name | JSON Type (Required/Optional) | Description | Example |
---|---|---|---|
role | user | assistant (Required) | Specifies the source of the memory item in MemoryItem. |
content | string | Expression (Required) | Specifies the actual value of the memory item in MemoryItem. |
Last updated