Skip to content

Models

Models live in the global registry. Tenant workflows normally use models through assignments, while platform and handler integrations use the direct model routes.

A model stores the human-readable name, description, metadata, handlers, optional handler contracts, and status. Each model has one or more versions. Versions have their own metadata and status.

Use a new version when the runnable behavior changes but the model identity should stay stable.

Handlers describe how training and inference execute. PIE V2 supports:

  • lambda
  • ecs_taskvalue must be <cluster>/<task_definition>, where <cluster> is the ECS cluster short name (no / in the name), and <task_definition> is the task family (or family:revision).
  • in_process

Direct handler configuration is part of model registration and update. Consumers who only run assignments usually do not need to inspect handler details.

Optional handler_contracts describe input and output payloads as JSON Schema Draft 2020-12 documents. When an inference input schema is present, the API can validate direct inference requests before dispatch.

Contracts also drive event-subscription staleness: changing a contract can mark related subscriptions stale so downstream automation does not silently run against incompatible data.

Training and inference are async. Starting either creates a durable job and returns a job_id. Poll the job route for status and output.

Event subscriptions chain model work together. A source model event triggers a target model action using either stored input or the previous job output.

TriggerTarget actionInput sourceTypical use
training_completeinferencestoredRe-run inference after retraining.
inference_completetrainingtrigger_outputFeed inference output into downstream training.
training_completetrainingtrigger_outputChain training artifacts.
inference_completeinferencestored or trigger_outputChain inference models.