Finding your IDs
Before making API calls, you’ll need these identifiers:| ID | Where to find it |
|---|---|
| Project ID | Settings → Project, or in any URL after /projects/ |
| Judge ID | Click a judge in the dashboard; the ID is in the URL (/judges/{judge_id}) |
| Span ID | In trace details, or returned by your instrumentation code |
Python SDK
Get evaluations by judge
Fetch all evaluations for a specific judge with pagination and optional filters.Get evaluations by span
Fetch all judge evaluations for a specific span (useful when a span has been evaluated by multiple judges).REST API
Use these endpoints directly with your API key in theAuthorization header.
Get evaluations by judge
| Parameter | Type | Description |
|---|---|---|
limit | int | Results per page (1-500, default 100) |
offset | int | Pagination offset (default 0) |
start_date | string | Filter by date (ISO 8601) |
end_date | string | Filter by date (ISO 8601) |
evaluation_result | bool | true for passing, false for failing |
feedback_state | string | with_user_feedback or without_user_feedback |
Get evaluations by span
Response format
Judge evaluations response
Span evaluations response
Evaluation object
| Field | Type | Description |
|---|---|---|
id | string | Unique evaluation ID |
span_id | string | The evaluated span |
evaluation_result | bool | Pass (true) or fail (false) |
evaluation_reason | string | Judge’s reasoning |
confidence_score | float | Model confidence (0-1) |
score | float | null | Numeric score (scored judges only) |
score_min | float | null | Minimum possible score |
score_max | float | null | Maximum possible score |
pass_threshold | float | null | Score required to pass |
model_used | string | LLM model that ran the evaluation |
created_at | string | ISO 8601 timestamp |