Skip to content

Comparison

Platform provides the ability to compare the registered objects for a specific task using standard and customized metrics. Using this comparison capability one can quickly evaluate a list of candidates to select the best one.

A comparison task typically involves:

  • Current object: Object which is currently selected and needs to be compared with others.
  • Challenger objects: Challenger objects are objects of the same type (e.g., models can be compared with other models, prompts can be compared with prompts, etc.).
  • Data Source: A common data source on which objects would be evaluated.
  • Report & Metrices: Exact evaluation metrics to be compared.

Note: On the platform one can quickly create Copy of current objects and change the definitions, swap in swap out components (Like Models, Prompts, Processing etc.) to create challengers.

How to run a Comparison Task?

  • Register the object and its challenger versions on the platform.
  • Go to the Details page of an object to be compared and click on the Run -> Comparison button.
  • Provide description about the comparison run.
  • Select Dashboard to be evaluated in Dashboard Selection.
  • Select challenger objects which need to be compared in Dependencies Section.
  • Prepare the data in Data Sources which will be used for evaluation.
  • Click on Run at the bottom and wait for job completion.

    • Once a job has been submitted it starts in the NEW status
    • The job will go through the following statuses: COMPILING > QUEUED > RUNNING and finally stop at COMPLETED of FAILED

All comparison tasks are systematically recorded on the platform and displayed on the Jobs page of an object in a structured format. They can also be exported as part of the automated documentation process.

Note: The platform allows customizing reports and dashboards specifically for comparison tasks. Note: Corridor allows running jobs in parallel and multiple threads at a time within a job to expedite the evaluation process.