Skip to main content

BLEU Score (bleu)

Contents

Metric description

BLEU measures n-gram overlap between the output and a reference (golden answer). It reflects similarity to the reference, not factual correctness. Smoothing can help when n-grams are missing.

How to interpret the score

  • Closer to 100: stronger match to the reference under BLEU-4 style scoring in this implementation.
  • Closer to 0: little overlap with the reference.

API usage

Prerequisites

After the environment variables are configured, the next step is to create a JSON payload for the custom-runs request. For a field-by-field description of the payload (top-level keys, evaluations, and each row in data), see Custom run request body.

Shortname: bleu

Default threshold: 70

Structural metrics run without an LLM (deterministic checks). Your run may still include model_slug where the API expects it; scoring does not depend on it for this category.

Inputs (each object in data)

  • output (str, required): Candidate text.
  • golden_answer (str, required): Reference text.

metric_args

  • smoothing (boolean optional): Apply smoothing when n-gram counts are zero. Default: true.

  • case_sensitive (boolean optional): Compare with case sensitivity. Default: false.

Eval metadata

Structural metrics do not populate eval_metadata; the field is omitted or ull on the result object.

Example

import json
import os

import requests
from dotenv import load_dotenv

load_dotenv(override=True)

_API_KEY = os.getenv("AEGIS_API_KEY")
_BASE_URL = os.getenv("AEGIS_API_BASE_URL")
_CUSTOM_RUN_URL = f"{_BASE_URL}/runs/custom"


def post_custom_run(payload: dict) -> requests.Response:
"""POST JSON payload to Aegis custom runs; returns the raw response."""
headers = {
"Content-Type": "application/json",
"Authorization": f"Bearer {_API_KEY}",
}
return requests.post(
_CUSTOM_RUN_URL,
headers=headers,
data=json.dumps(payload),
)


if __name__ == "__main__":
data = [
{"output": "hello world", "golden_answer": "hello world"}
]

payload = {
"threshold": 70,
"model_slug": "o4-mini",
"is_blocking": True,
"data_collection_id": None,
"evaluations": [
{
"metrics": [
{
"metric": "bleu",
"metric_args": {"smoothing": True, "case_sensitive": False},
},
],
"threshold": 70,
"model_slug": "o4-mini",
"data": data,
}
],
}

response = post_custom_run(payload)
response.raise_for_status()
print(json.dumps(response.json(), indent=2))