Model Providers
OpenRouter
If you use OpenRouter models within your application, you can trace, monitor, and analyze them with Judgment. You can do this by using:
- Judgeval Wrap: Add our wrap around the initialization of your client
OpenRouter provides a unified interface to access multiple LLM providers through a single API. Since OpenRouter uses an OpenAI-compatible API, you can use the OpenAI client with OpenRouter's base URL.
Judgeval Wrap
from openai import OpenAI
from judgeval import Judgeval
client = Judgeval(project_name="openrouter_project")
tracer = client.tracer.create()
openai_client = OpenAI(
api_key="YOUR_OPENROUTER_API_KEY",
base_url="https://openrouter.ai/api/v1",
default_headers={
"HTTP-Referer": "YOUR_OPENROUTER_URL",
"X-Title": "YOUR_OPENROUTER_NAME",
}
)
wrapped_client = tracer.wrap(openai_client) # Use wrapped_client from here onCost Tracking
Once wrapped, you can use the client as normal and all calls will be traced:
response = wrapped_client.chat.completions.create(
model="openai/gpt-5.2",
messages=[
{"role": "user", "content": "Hello, how are you?"}
],
extra_body={"usage": {"include": True}}, # Include this to capture costs
)Last updated on