API Reference

  • Benchmark: Data abstractions for packaging annotated samples into reusable benchmarks and CSV-backed suites.
  • Critico: Orchestrator that aggregates evaluator results and reports back to the RELAI platform.
  • Evaluator: Base classes and built-in evaluators for rubric, format, style, and annotation scoring.
  • Maestro: Optimization engine that tunes agent configs and structure based on evaluation feedback.
  • Mockers: Persona and mock tool definitions for simulating MCP interactions during tests.
  • Simulator: Decorators and simulator runtimes to replay agent flows in controlled environments.
  • Types: Core data models (RELAISample, SimulationTape, logs) shared across simulation and evaluation.