AI Agents
Quality center
Knowledge agents

Knowledge agents

EARLY ACCESS

Use the Quality center to evaluate and monitor your knowledge agents. Test agent responses with sample questions, analyze real end user traffic, and use the results to identify issues and improve agent performance.

To access, go to AI Agents > Quality center > Knowledge Agents.

Knowledge agents quality center

Task types

  • Evaluate - Test the knowledge agent with a set of sample questions and check whether it responds as expected. Use ground truth evaluation to compare responses against expected answers.

Process overview

  1. Select the knowledge agent you want to evaluate.
  2. Create and run the task.
  3. Check the status of the task.
  4. When the task is complete, download and view the results.
  5. Update the knowledge agent as required.

Evaluate

Test the knowledge agent with sample questions to verify it responds as expected. Upload a .csv file with questions, run the task, and review the results.

Prepare test questions

Create a .csv file with sample questions that represent real end user queries.

Example questions for a car dealership agent:

  • What car models do you have available?
  • Do you sell new and used cars?
  • Can I trade in my old car?
  • What financing options do you provide?

To verify answer accuracy, use ground truth evaluation by adding a reference_answer column alongside each question.

Run the task

  1. On the Infobip web interface, go to AI Agents > Quality center > Knowledge Agents.
  2. In the Select knowledge agent field, select the agent you want to test.
  3. Upload the .csv file that contains the set of questions.
  4. In the Task name field, enter a name for the task.
  5. Select Run task.

Ground truth evaluation (optional)

Ground truth evaluation compares agent responses against expected answers you provide. Add two columns to your .csv file: question and reference_answer.

When the test runs, each generated answer receives one of these labels:

LabelMeaning
GOODThe answer matches the expected answer
INCOMPLETECorrect overall but missing some details from the expected answer
BADContains incorrect or fabricated information

The label and reasoning for each answer are included in the results file.


View task status

You can view all tasks in the Tasks section. This section shows a maximum of the 20 most recent tasks of each type.

StatusDescription
CompletedThe task is complete and the results are ready to download.
FailedThe task failed. Hover over the information icon next to the status to see why.
In progressThe task is running.

View the results

When the task is complete, download the results .csv file from the Tasks section.

Results fields

FieldDescription
questionThe question from the test file.
answerThe response generated by the knowledge agent.
original_contextsRelevant chunks from the knowledge sources that were retrieved, including the source document name.
reranked_contextsContexts after reranking, if enabled. Not available for very short messages.
content_filter_resultsContent filter result for each message. Included only if the content filter is enabled. Contains the status (safe, annotated, blocked), and violation details per category including severity, threshold, configured action, and whether the filter was triggered.
latencyTime taken by the knowledge agent to generate the response. This differs from the chatbot response time, which may include additional processing.
topicThe topic determined by the agent based on the question. Provides insight into end user interests.
is_question_answeredTRUE: agent answered the question. FALSE: agent could not answer (missing content or out of scope). N/A: message was not a question.
is_answer_in_contextIndicates whether the answer is based on the retrieved context.
answer_classificationGOOD, INCOMPLETE, or BAD. Included only for Evaluate tasks when reference_answer is provided.
answer_classification_reasoningExplanation of why the answer received its classification. Included only for Evaluate tasks when reference_answer is provided.

Update the agent based on results

Use the results to identify and fix issues. Investigate the root cause before making changes.

IssueSolution
Agent answered correctly but the answer is not in the contextThe prompt likely contains the answer. If the answer is wrong, check the knowledge source content.
Agent uses an inappropriate toneReview and update the prompt in agent settings.
Agent responds in the wrong languageReview and update the prompt in agent settings.
Agent responses are cut offIncrease the output tokens setting.
Agent hallucinatesThe knowledge source is likely missing relevant content. Add it.
Agent answers out-of-scope questionsTighten the prompt to define scope and limitations in agent settings.

For an overview of testing approaches, see Test knowledge agent.







Need assistance

Explore Infobip Tutorials

Encountering issues

Contact our support

What's new? Check out

Release Notes

Unsure about a term? See

Glossary
Service status

Copyright @ 2006-2026 Infobip ltd.

Service Terms & ConditionsPrivacy policyTerms of use