Skip to Content
Getting StartedRun Evaluations

Run Evaluations

With your environment set up and your application connected, you are ready to run evaluations. In Rhesis, a group of test cases used to evaluate an application is called a Test Set.

A Test Set is a collection of prompts (with optional context and expected outputs) that will be sent to your Endpoint.

There are three main ways to acquire or create Test Sets in Rhesis:

Generating Tests

Instead of writing test cases manually, you can use Rhesis to automatically generate them based on your application’s requirements or constraints.

Using the UI: You can use the Rhesis Platform to generate tests interactively. Navigate to the Generation menu item, choose the type of test, then choose whether you want to use AI to generate the tests based on a description, or do it manually.

Test Generation

Using the Python SDK: You can also generate tests programmatically using the PromptSynthesizer.

generate_tests.py
from rhesis.sdk.synthesizers.prompt_synthesizer import PromptSynthesizer

synthesizer = PromptSynthesizer(
    "Generate tricky customer support questions about refund policies"
)

# Generate a test set with 10 questions
test_set = synthesizer.generate(num_tests=10)

Importing Existing Tests

If you already have test cases stored in a file (such as a CSV or JSON file from previous evaluations), you can easily import them into Rhesis.

This is useful for regression testing or when you want to evaluate your application against established benchmarks. You can upload files directly through the Platform UI.

Adversarial Security Testing (Garak)

For evaluating the security and robustness of your AI application, Rhesis integrates with specialized tools to generate adversarial tests.

Using our integration with Garak (Generative AI Red-teaming and Assessment Kit), you can automatically probe your application for vulnerabilities like prompt injection, data leakage, and harmful content generation.

Running the Test Set

Once your Test Set is ready and your Endpoint is connected, you can execute a test run.

  1. Navigate to the Test Sets page in the Rhesis UI.
  2. Select your newly created Test Set.
  3. Click the Run Test button.
  4. Select the Endpoint you connected in Step 2.
  5. (Optional) Select any evaluation metrics you want to compute on the results.
  6. Click Execute to start the run.

Rhesis will orchestrate the execution, sending each test case to your application in parallel, collecting the responses, and computing relevant metrics. You can then analyze the results in the interactive dashboards.

For detailed information on configuring execution behavior, concurrency, and handling retries, see the Test Execution Guide.

Congratulations! You’ve completed the Getting Started guide. To dive deeper, explore the Platform Guide or read up on our Core Concepts.