# This is an abstract flow, therefore some required fields are not defined (and must be defined by the concrete flow) _target_: flow_modules.aiflows.ChatFlowModule.ChatAtomicFlow.instantiate_from_default_config name: ChatAtomicFlow description: "Flow which uses as tool an LLM though an API" enable_cache: True n_api_retries: 6 wait_time_between_retries: 20 system_name: system user_name: user assistant_name: assistant backend: _target_: aiflows.backends.llm_lite.LiteLLMBackend api_infos: ??? model_name: openai: "gpt-3.5-turbo" n: 1 max_tokens: 2000 temperature: 0.3 top_p: 0.2 stream: True system_message_prompt_template: _target_: aiflows.prompt_template.JinjaPrompt init_human_message_prompt_template: _target_: aiflows.prompt_template.JinjaPrompt template: "{{query}}" input_variables: - "query" human_message_prompt_template: _target_: aiflows.prompt_template.JinjaPrompt template: "{{query}}" input_variables: - "query" input_interface_initialized: - "query" query_message_prompt_template: _target_: aiflows.prompt_template.JinjaPrompt previous_messages: first_k: null # Note that the first message is the system prompt last_k: null output_interface: - "api_output"