Validators
A validator is any entity that runs part of the model on its own hardware, signs its outputs, and stakes the network's token against the correctness of those outputs. Validators are the network's compute substrate.
What validators do
A validator does one or more of:
- Serves inference. Runs forward passes of a published checkpoint against incoming queries.
- Trains. Participates in distributed training runs as a worker in a 3D-parallel job.
- Evaluates. Runs eval suites against new checkpoints before they are promoted.
- Stores data. Holds a shard of the content-addressed dataset and serves it on demand.
- Audits. Re-executes a sample of other validators' outputs and submits challenges on mismatch.
Most validators do a subset, not all. Hardware tiers and the validator's own selection determine which roles it can take.
Hardware tiers
| Tier | Reference hardware | Eligible roles |
|---|---|---|
| Light | Single consumer GPU (24GB+) | Inference (small surfaces), evaluation, storage, audit |
| Standard | Single consumer/prosumer GPU (32GB+) | Inference (all surfaces), evaluation, audit |
| Heavy | Single data-center GPU (80GB) | Inference, evaluation, training (worker), audit |
| Cluster | Multi-node, high-bandwidth interconnect | Training (full), inference (frontier surfaces) |
The tier of a validator is verified at registration by an on-chain attestation that captures hardware identity through TEE-issued certificates. Tier upgrades require re-attestation; tier downgrades happen automatically when a validator fails to meet its tier's latency SLO.
Critical: the Light tier exists because the codex's commitment is that participation should not require institutional capital. A node with a single 24GB GPU can serve at least the language surface in its distilled variant and the audit role for all surfaces.
Staking and slashing
Each validator stakes the network's token. The stake serves two functions: it provides a bond against fraud, and it weights the validator's selection probability in the routing layer.
The slashing conditions are:
- Output divergence. An audited query whose output disagrees with the consensus output beyond the surface's tolerance window. Penalty: configurable, currently of stake per incident, capped at per epoch.
- Attestation forgery. A signed attestation that does not match the validator's actual computation. Penalty: full slash.
- Liveness failure. Sustained failure to respond to dispatched jobs within the latency SLO. Penalty: temporary deregistration, no stake loss.
- Censorship. Selectively dropping queries based on content. Penalty: full slash, after challenge.
Slashing requires a successful challenge through the challenge service. Challenges have a 48-hour window during which any participant can submit a fraud proof. Successful challengers receive a fraction of the slashed stake.
Selection and rewards
Validators are selected for jobs by a probability proportional to
where is stake, is a moving-window quality score from prior audits, and is a stake-weighting parameter set by governance. With the selection is purely stake-weighted; with it is purely quality-weighted. The current operating point is .
Rewards are paid from the treasury per attested job. Reward magnitude is a function of the surface, the latency tier, and the duration of the job. Validators receive rewards in the network token; the relationship between the token and the validator's operating costs is the subject of Token.
Verification mechanics
Validator output is verified statistically, not cryptographically. The mechanisms:
Random redundant dispatch. A fraction of jobs are dispatched to two validators. Mismatches trigger a third-validator resolution and, if confirmed, slashing of the divergent party.
Eval-suite sampling. Periodically, every validator runs a standard eval suite against the active checkpoint. Reported scores are cross-validated; outliers are flagged.
TEE attestation. For training jobs, the worker process runs inside an attested execution environment. The attestation binds the computation to specific hardware and to a specific code binary.
Loss-curve commitment. For training, the validator commits at each checkpoint to a hash of the loss curve. Re-execution must reproduce the curve within an -band.
None of these mechanisms provides cryptographic certainty. Together, with the slashing bond, they provide economic certainty: the expected cost of cheating exceeds the expected gain by a configurable margin.
Economic model
Let be the cost to a validator of an attempted fraud (mostly the slashing risk), be the expected gain (typically negligible for inference, larger for training jobs), and the probability of detection. The honest-validator condition is
The protocol's parameters are set to keep this inequality wide. Specifically:
- is bounded below by , which is configurable per surface.
- is the validator's stake, which is constrained to a minimum proportional to the maximum job value the validator is eligible for.
- is upper-bounded by the protocol's per-job payment.
What validators are not
A validator is not a node in a consensus protocol. The chain achieves consensus through its own validator set; OGI validators provide compute, not consensus. The two notions of "validator" overlap in name only.
A validator is not a customer. Validators do not consume the model's output; they produce it. End users consume through gateways. The two roles are deliberately separate.