Skip to Content
ConceptsOverview

Concepts

LaunchGate is built around a simple hierarchy that maps to how teams think about AI quality.

Architecture overview

Organisation └── Project └── Eval Suite (pass threshold: 0.8) ├── Eval Case 1 (scorer: exact_match, threshold: 1.0) ├── Eval Case 2 (scorer: contains, threshold: 1.0) └── Eval Case 3 (scorer: llm_judge, threshold: 0.7)

Core entities

EntityDescription
ProjectTop-level container scoped to your organisation
Eval SuiteA collection of test cases with a pass/fail threshold
Eval CaseA single test — an input, a scorer, and a threshold
ScorerThe evaluation function (exact match, regex, LLM judge, etc.)
RunOne execution of a suite — immutable once complete
BYOK KeyYour encrypted LLM provider key for AI-powered scoring

The eval flow

  1. You call POST /v1/run (or use the SDK/CLI/GitHub Action)
  2. LaunchGate loads the suite and all its active cases
  3. Each case runs its scorer against your output concurrently
  4. Results are aggregated and the pass rate is computed
  5. If pass rate ≥ suite threshold → Cleared for launch
  6. If pass rate < suite threshold → Launch aborted
  7. Results are stored immutably for audit and trend analysis
Last updated on