

Conversation Opener
Set an opening message that greets users at the start of each conversation, with optional suggested questions to guide them toward what the app does well. You can insert variables into the opening message and suggested questions to personalize the experience.-
In the opening message, type
{or/to insert variables from the picker. -
In suggested questions, type variable names manually in
{{variable_name}}format.


Follow-up
When enabled, the LLM automatically suggests 3 follow-up questions after each response, helping users continue the conversation. This feature is a simple toggle—no additional configuration is needed.Follow-up questions are generated by a separate LLM call using your workspace’s system reasoning model (set in Settings > Model Provider > Default Model Settings), not the model configured in your app.
Text to Speech
Convert AI responses to audio. You can configure the language and voice to match your app’s audience, and enable Auto Play to stream audio automatically as the AI responds.Text to Speech uses your workspace’s text-to-speech model (set in Settings > Model Provider > Default Model Settings).The feature only appears in the Features panel when a default TTS model is configured.
Speech to Text
Enable voice input for the chat interface. When enabled, your end users can dictate messages instead of typing by clicking the microphone button.Speech to Text uses your workspace’s speech-to-text model (set in Settings > Model Provider > Default Model Settings).The feature only appears in the Features panel when a default STT model is configured.
File Upload
Allow end users to send files at any point during a conversation. You can configure which file types to accept, the upload method, and the maximum number of files per message.For self-hosted deployments, you can adjust file size limits via the following environment variables:
UPLOAD_IMAGE_FILE_SIZE_LIMIT(default: 10 MB)UPLOAD_FILE_SIZE_LIMIT(default: 15 MB)UPLOAD_AUDIO_FILE_SIZE_LIMIT(default: 50 MB)UPLOAD_VIDEO_FILE_SIZE_LIMIT(default: 100 MB)
Citations and Attributions
Show the source documents behind AI responses. When enabled, responses that draw from a connected knowledge base display numbered citations linking back to the original documents and chunks.
Content Moderation
Filter inappropriate content in user inputs, AI outputs, or both. Choose a moderation provider based on your needs:- OpenAI Moderation: Use OpenAI’s dedicated moderation model to detect harmful content across multiple categories.
- Keywords: Define a list of blocked terms. Any match triggers the preset response.
- API Extension: Connect a custom moderation endpoint for your own filtering logic.
Annotation Reply
Define curated Q&A pairs that take priority over LLM responses. When a user’s query semantically matches an annotation above the score threshold (how closely a query must match), the curated answer is returned directly without calling the LLM. You can configure the score threshold and the embedding model used for semantic matching. To create and manage your annotations:-
Convert existing conversations into annotations directly from Debug & Preview or Logs by clicking the Add Annotation icon on any LLM response.
Once a message is annotated, the icon changes to Edit, so you can modify the annotation in place.

-
In the Logs & Annotations > Annotations tab, manually add new Q&A pairs, manage existing annotations, and view hit history. Click
...to bulk import or bulk export.
More Like This
Generate alternative outputs for the same input. Once enabled, each generated result includes a button to produce a variation, so you can explore different responses without re-entering your query.