This custom visual has the following 2 main configuration sections:
This section covers all the required configurations for the LLM endpoint, along with individual settings for the chatbot, including fine-tuning options for its three modes: General, File, and Dataset. The General mode functions solely as an interface to the LLM, without any additional logic applied.
LLM: This section contains the settings for the LLM endpoint.
Chatbot: This section contains the settings for defining and customizing the behaviors of the Chatbot.
File: This section contains the fine-tuning options for the File mode.
**Dataset:** This section contains the fine-tuning options for the Dataset mode.
LLM host: The host of the LLM endpoint, currently you can choose between OpenAI and Azure OpenAI, depending on this selection the following options change.
This section will be helpful to determine the behavior and appearance of the Chatbot. The
This section provides an overview of the file-mode behavior of the chatbot, which enables users to summarize and ask questions about uploaded files. No external network requests are made, and no data is sent to external entities other than the designated LLM endpoint.
Embeddings model name: This refers to the name of the model used to create embeddings stored in memory. You can select from predefined options such as text-embedding-3-small
, text-embedding-3-large
, text-embedding-ada-002
, or specify a custom model name of your choice using other
Fine-tuning Embeddings: Enables customization of embedding creation
Embedding Dimension: Specifies the size of the vector space, balancing accuracy and computational efficiency.
Embeddings Batch Size: Defines the number of items processed per batch, optimizing memory usage and speed.
Fine-tuning summarization: This allows you to fine-tune the creation of embeddings.
Summarization disable: This allows you to disable the automatic summarization feature on file upload.
Summarization chunk size: Defines the number of characters summarized per chunk before combining results (map-reduce approach).
Summarization chunk overlap: Defines the number of characters overlapping per chunk.
Fine-tuning Question-Answering: Enables customization of embeddings for more precise answers.
Question-Answering Chunk Size: Specifies the number of tokens per embedding.
Question-Answering Chunk Overlap: Defines the number of overlapping tokens between chunks/embeddings.
Question-Answering Vector Search Limit: Sets the maximum number of vectors to query for answers.