Get the most from AI while staying safe

Evaluate models using state-of-the-art tools by leading experts. Collaborate on results.

Overview

How's Your AI Performing?

Gage helps you answer the hard questions.

1 – Measure

Apply benchmarks and custom evaluation code to your models.

2 – Publish

Publish results to Gage public or private leaderboards.

3 – Compare

Compare and analyze results. Deepen your understanding.

4 – Collaborate

Get input from team members and other experts on next steps.

Gage Run

Open Source Tools
to Run Evals

Latest Algorithms

Run any code for any framework to leverage the ecosystem of experts.

Trustworthy Results

Track full provenance with industry standard cryptographic methods.

Gage Live

Leaderboard Publishing

Publish evaluations in second to Gage Live leaderboards.

  • Stay up-to-date with the latest from your team and other projects
  • Drill into details and collaborate on next steps for improving your apps
  • Use your own infrastructure for privacy and compliance
About Gage

Gage was founded by the team behind Guild AI, a leading open source experiment tracking toolkit.

Gage has two parts. Gage Run is an open source toolkit for running evaluations and managing results. Gage Live is an online platform for publishing and exploring evaluation results.

Gage is currently in beta and under active development. If you'd like to contribute, request features, or otherwise stay up to date, join the Gage community and say hi!