Skip to main content

AI Chatbot

Every screen in Perf Agent includes a built-in AI chatbot, accessible via the chat icon in the bottom-right corner of the interface. It has full context of the platform — your builds, scripts, NFR strategies, bugs, and integrations — and can answer questions, surface insights, and help you navigate results without leaving your current screen.

Rather than manually digging through dashboards, you can ask the chatbot in plain English and get structured, data-driven answers instantly.

AI Chatbot panel showing a conversation — user asks "show me list of nfr applications" and the chatbot responds with a table of 11 NFR applications with ID, Status, and Created On columns


Opening the Chatbot

Click the chat icon in the bottom-right corner of any screen in Perf Agent. The chatbot panel slides open as an overlay — it does not navigate you away from your current page.

Each session starts fresh. Use New Chat in the left panel to start a new conversation. Previous sessions are listed under Recent Sessions and can be reopened to review past queries and responses.


Selecting an AI Model

The chatbot icon in the input bar lets you switch the underlying AI model before sending a query.

Model selector showing available models: Qwen-3, claude-3-sonnet (selected), gpt-4o-mini, mistral-7b-instruct, gpt-4

Click the model selector icon (⚙️) next to the input field to see the list of available models. The currently active model is shown with a checkmark. Select any model from the list to switch — your next message will be processed by the selected model.

Available models are configured by your administrator under Settings → AI Settings. See the AI Settings section below for details.

tip

Different models may produce different levels of detail or reasoning depth. If a response feels too brief or lacks context, try switching to a more capable model and re-asking the question.


What the Chatbot Can Answer

The chatbot has access to all data within your Perf Agent workspace. Below are the main categories of questions it can handle.


Performance Metrics

Query response times, TPS (Transactions Per Second), latency, and error rates — for specific services, builds, or endpoints.

Example queries:

  • Show me the average response time for catalogue-service
  • What is the 95th percentile response time?
  • Which service has the highest TPS?
  • Show endpoints with response time greater than 1000ms
  • Show average latency for id-broker service

Latest Build

Ask about the most recent test execution without needing to know the build number.

Example queries:

  • Show me the latest build performance
  • What is the status of the latest build?
  • Show error count in the latest build
  • What are the latest build observations?
  • How is the latest build performing?

Specific Build Analysis

Reference any build by its ID or number to drill into its metrics, errors, observations, and defects.

Example queries:

  • Show me BUILD_20260203_120447
  • What was the performance of build-40?
  • Show observations for build-39
  • What were the Datadog remarks for build-40?
  • Display ADO defects for build-39
  • Compare build-40 with build-39

Trend Analysis

Analyse how performance has changed over time — across builds, services, or specific metrics.

Example queries:

  • Show TPS trend over last 5 builds
  • Display response time trend over last 10 builds
  • What is the error trend over time?
  • Show performance trend for catalogue-service
  • Display performance degradation over builds

Service and Endpoint Comparison

Compare performance across multiple services or endpoints side by side.

Example queries:

  • Compare performance of catalogue-service and workspace-service
  • Which service has the highest response time?
  • Show top 5 services by TPS
  • Which endpoint has the most errors?
  • Show top 10 slowest endpoints
  • Which services are within SLA?

Error Analysis

Investigate error rates, error spikes, and failing endpoints across builds and services.

Example queries:

  • Show error rate for catalogue-service
  • What is the error percentage in latest build?
  • Show endpoints with error rate greater than 5%
  • Show error trend over last 10 builds
  • Which builds had the most errors?
  • Which endpoints are failing?

Build Observations and Defects

Surface observations, Datadog remarks, and ADO defects logged against specific builds.

Example queries:

  • Show observations for latest build
  • What were the Datadog remarks for build-40?
  • Show all builds with ADO defects
  • List all builds that have defects
  • Which builds have Datadog alerts?

Transaction Analysis

Query individual transaction performance — response times, TPS, error rates, and comparisons across builds.

Example queries:

  • Show all transactions
  • Which transaction has the highest error rate?
  • Display top 5 slowest transactions
  • What is the TPS of login transaction?
  • Which transactions degraded?

Execution Status

Check the status of running, queued, or completed test executions.

Example queries:

  • Show all failed executions
  • Which executions are currently running?
  • What is the status of execution-123?
  • How long did the last execution take?
  • Show execution duration for build-40

Script Management

Query the status and details of scripts generated by Auto Script.

Example queries:

  • Show me the latest script
  • List all completed scripts
  • Show scripts for MyApp
  • Which scripts failed?
  • What is the status of script generation?

Bug Analysis

Retrieve open bugs, bug statistics, and bug details from your connected issue tracker.

Example queries:

  • Show me all open bugs
  • Display bugs for Bill Management service
  • List high priority bugs
  • How many critical bugs are open?
  • Show bug count by priority
  • Which service has the most bugs?

SLA Compliance

Check whether builds are meeting the SLA thresholds defined in your NFR strategies.

Example queries:

  • Show builds meeting SLA requirements
  • Which builds violated SLA?
  • Display SLA compliance percentage
  • Show response time SLA violations
  • Which services are within SLA?

Process and Configuration Questions

Ask how to use any part of Perf Agent — Auto Script requirements, NFR generation steps, integration setup, and more.

Example queries:

  • What is needed for autoscript?
  • How to generate NFR document?
  • How to configure Azure DevOps?
  • What credentials are needed for BlazeMeter?
  • How to run a performance test?
  • What are the prerequisites for test run?

Tips for Better Results

The chatbot produces more accurate and useful responses when questions are specific and include context.

Instead ofTry
Show performanceShow TPS trend for catalogue-service over last 5 builds
Show errorsShow error trend for workspace-service
Show response timeShow response time trend over last 10 builds
Show that buildShow BUILD_20260203_120447
Compare buildsCompare build-40 with build-39

Use service names — The chatbot understands service identifiers like catalogue-service, id-broker, and workspace-service directly.

Use build numbers or "latest" — Reference builds by their ID (e.g., build-40, BUILD_20260203_120447) or simply say latest build for the most recent execution.

Specify time ranges — Phrases like last 5 builds, last 10 builds, or last 2 weeks narrow the results and produce cleaner trend charts.

Ask follow-up questions — The chatbot retains context within a session. After asking about the latest build, you can follow up with Why did it fail? or Compare it with build-39 without repeating context.


AI Settings

The models available in the chatbot are configured globally under Settings → AI Settings.

AI Settings page showing a table of configured models: Qwen-3 (Qwen), claude-3-sonnet (anthropic), gpt-4o-mini (openai), mistral-7b-instruct (mistralai), gpt-4 (OpenAI) — each with a Rate Limit column and a delete action

Configured Models

The AI Settings table lists all models available to users in the chatbot, with the following columns:

ColumnDescription
Model NameThe identifier for the model as it appears in the chatbot model selector
ProviderThe AI provider — anthropic, openai, mistralai, qwen, etc.
Rate LimitMaximum number of requests per minute allowed for this model
ActionDelete the model configuration

Adding a Model

Click + Add Model in the top-right corner of the AI Settings page to register a new model. You will need the model name, provider, API key, and rate limit. Once added, the model becomes immediately available in the chatbot's model selector for all users in the workspace.

Removing a Model

Click the delete icon (🗑️) in the Action column to remove a model. Removing a model makes it unavailable in the chatbot selector — any users who had it selected will fall back to the default model.

note

AI Settings are administrator-level controls. Contact your Perf Agent administrator if you need a specific model added or if you are experiencing rate limit errors during chatbot queries.