Conversation Assistant
Last updated
Last updated
Conversation applications use a one-question-one-answer mode to have a continuous conversation with the user.
Conversation applications can be used in fields such as customer service, online education, healthcare, financial services, etc. These applications can help organizations improve work efficiency, reduce labor costs, and provide a better user experience.
Conversation applications supports: prompts, variables, context, opening remarks, and suggestions for the next question.
Here, we use a interviewer application as an example to introduce the way to compose a conversation applications.
Click the "Create Application" button on the homepage to create an application. Fill in the application name, and select "Chat App" as the application type.
After the application is successfully created, it will automatically redirect to the application overview page. Click on the button on the left menu: "Orchestrate" to compose the application.
2.1 Fill in Prompts
Prompt phrases are used to guide AI in providing professional responses, making the replies more precise. You can utilize the built-in prompt generator to craft suitable prompts. Prompts support the insertion of form variables, such as {{input}}
. The values in the prompt variables will be replaced with the values filled in by the user.
Example:
Enter the interview scenario command.
The prompt will automatically generate on the right side content box.
You can insert custom variables within the prompt to tailor it to specific needs or details.
For a better experience, we will add an opening dialogue: "Hello, {{name}}. I'm your interviewer, Bob. Are you ready?"
To add the opening dialogue, click the "Add Feature" button in the upper left corner, and enable the "Conversation remarkers" feature:
And then edit the opening remarks:
2.2 Adding Context
If an application wants to generate content based on private contextual conversations, it can use our knowledge feature. Click the "Add" button in the context to add a knowledge base.
2.3 Uploading Documentation File
Some LLMs now natively support file processing, such as Claude 3.5 Sonnet and Gemini 1.5 Pro. You can check the LLMs' websites for details on their file upload capabilities.
Select an LLM that supports file reading and enable the "Documentation" feature. This enables the Chatbot to recognize files without complex configurations.
2.4 Debugging
Enter user inputs on the right side and check the respond content.
If the results are not satisfactory, you can adjust the prompts and model parameters. Click on the model name in the upper right corner to set the parameters of the model:
Debugging with multiple models:
If debugging with a single model feels inefficient, you can utilize the Debug as Multiple Models feature to batch-test the models’ response effectiveness.
Supports adding up to 4 LLMs at the same time.
⚠️ When using the multi-model debugging feature, if only some large models are visible, it is because other large models’ keys have not been added yet. You can manually add multiple models’ keys in “Add New Provider”.
2.4 Publish App
After debugging your application, click the "Publish" button in the top right corner to create a standalone AI application. In addition to experiencing the application via a public URL, you can also perform secondary development based on APIs, embed it into websites, and more. For details, please refer to Publishing.
If you want to customize the application that you share, you can Fork our open source WebApp template. Based on the template, you can modify the application to meet your specific needs and style requirements.