Improving Performance
How to add skills, function calls, and more
Introduction
Brightbot is a technologically advanced AI system designed to enhance productivity and improve the overall user experience. However, like any other system, its performance can be measured and improved upon. In order to do this, we need to understand the various levers that can be used to quantify and enhance its performance.
Measuring Performance
There are four main factors that are used to measure the performance of BrightBot:
Usability
This involves measuring the response time and API reliability of the system.
Accuracy
The percentage of correct answers against benchmarking tasks and the retrieval of the most relevant information from within the warehouse are key indicators of accuracy.
Recall
This refers to the ability of the system to retrieve all relevant tables or documents based on a query.
Creativity
This is a subjective measure of the ability of the bots to create novel ideas and insights for users.
Improving Performance
Once the performance of Brightbot has been measured, there are a number of ways in which it can be improved:
System Engineering
This involves improving API performance and reliability.
Prompt Engineering
Utilizing advanced prompting techniques such as the chain of thought and MEDPrompt can greatly enhance the performance of the system.
RAG refinement
This involves better chunking, indexing, reranking, and multi-part context enhancement to improve system performance.
Function and Skill Augmentation
Adding new function calls and skills to the workflow to offload logic and reasoning from the Language Models (LLMs).
Changing LLMs
Different tasks may require different LLMs, and swapping which agents use which LLMs for which tasks can improve performance.
Fine-tuning
This involves making minor adjustments to the system to improve its overall performance.