Skip to content

Model Guide

Guide Maps

graph LR
  train["fit.py"] --> model["model.json"]
  model --> evaluate["evaluate.py"]
  model --> publish["publish/v1/"]
  publish --> review["Release review"]
flowchart LR
  question["What scoring behavior was promoted?"] --> model["Read the promoted model summary"]
  model --> contract["Check feature names, training metadata, and learned weights"]
  contract --> review["Only then compare metrics and threshold policy"]

Use this guide when model.json feels like a technical artifact rather than a review surface. The goal is to make the promoted model file legible enough to support release review without pretending it tells the whole repository story by itself.

What the promoted model owns

Section Meaning Why it matters
feature_names declared model input contract proves the promoted scorer still matches the capstone feature set
weights learned influence per standardized feature exposes the promoted scoring behavior
means and scales standardization contract from training data explains how raw values become model inputs
bias base decision tendency keeps the scoring rule reproducible rather than implied
training rows, iterations, learning rate, l2, and final loss keeps the promoted training story reviewable later

What the promoted model should not answer

  • whether the threshold is the right operational decision policy
  • whether one experiment should replace the baseline
  • whether the publish bundle proves full internal provenance

Use make model-summary when you want the promoted training and weight story rendered into one compact review surface before opening the raw model file.

Best companion guides

  • read CONTROL_SURFACE_GUIDE.md when the next question is how training params changed the model behavior
  • read PUBLISH_CONTRACT.md when the next question is why the model belongs in the promoted release boundary at all
  • read RELEASE_REVIEW_GUIDE.md when the next question is how to combine model review with metrics and report review