Skip to content

kevinshowkat/eval-receipts

Eval Receipts — Labs Edition

Eval Receipts ingests AI-lab evaluation claims, reproduces them with pinned, open harnesses, and publishes a shareable receipt per claim with a neutral status.

Eval Receipts verdict marking a claim as likely exaggerated

Status

This repo is a work in progress. Expect rough edges, missing documentation, and breaking changes while we land the core workflows.

Quick Start

git clone https://github.com/kevinshowkat/eval-receipts
cd eval-receipts
cp .env.example .env
# populate IDs for any providers you want to exercise
docker compose up --build

Repo Map

  • apps/web — Next.js client
  • apps/api — FastAPI service + workers
  • packages/harness — evaluation adapters
  • packages/shared — shared TS types
  • docs — method cards and operational notes
  • demo — example claims and artifacts

About

Replays LLM evals; exports audit-ready receipts (latency, cost, traces)

Topics

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published