Unify Agentic AI
Logo
Datasets

Datasets

Logo

6 mins READ

Datasets form the foundation of AI agent testing by providing structured collections of test cases. These test cases enable systematic evaluation of agent responses across various scenarios, from simple question-answer pairs to complex conversational flows.

Dataset Types

The platform accommodates two primary test case categories:

  • LLM Test Cases: Designed for single input-output validation, particularly useful for RAG (Retrieval-Augmented Generation) use cases

  • Conversational Test Cases: Built to test multi-turn dialogues and complete user journeys

    Frame 427319481.png
    Frame 427319481.png

Creating Datasets

Step 1: Name Your Dataset

Begin by providing a descriptive name that reflects the dataset's purpose and scope.

Step 2: Choose Creation Method for test cases

Frame 427319482.png
Frame 427319482.png

The platform offers five flexible options for dataset creation:

  • From Sessions: Import test cases directly from existing agent sessions

    1. Select the specific agent whose sessions you want to import

    2. Define a sampling rate to control the volume of imported cases

    3. Review and potentially filter the imported data for relevance

      Frame 427319483.png
      Frame 427319483.png

  • Upload CSV: Bulk import test cases via CSV files

    Frame 427319484.png
    Frame 427319484.png

  • Generate with Prompts: Auto-generate test cases by defining dataset goals

    1. Define clear, specific goals for your dataset

    2. Select agent for which you want to prepare test cases

    3. Choose desired AI modal and no. of test cases you want that model to generate

    4. Select columns to generate

Frame 427319485.png
Frame 427319485.png
  • Add manually: Manually input questions and expected answers

  • Create OTel Cases: Transform OpenTelemetry trace data from sessions on different platform into test cases

    1. Select “add from otel traces” after clicking on “Add test case” button

    2. You will see no. of traces available in system

    3. Add condition or condition group depending upon your usecase and requirements.

      Frame 427319486.png
      Frame 427319486.png

    4. Click on “create” button. You will see that the dataset has been created.

      Frame 427319487.png
      Frame 427319487.png

Step 3: Define Test Case Fields

Each test case can include up to seven fields:

  • Input (mandatory): The user query or prompt

  • Expected Output (mandatory): The desired agent response

  • Expected Tool: Specific tools the agent should utilise

  • Context: Additional background information

  • You can add custom fields according to your requirement

    Frame 427319488.png
    Frame 427319488.png

This flexible approach ensures datasets can be tailored to specific testing requirements while maintaining consistency across different evaluation scenarios.