Production AI Institute · Independent certification for production AI practice
Verify a credential|Contact
HomeThe StandardPSF Framework
PSF v1.0 — April 2026

Production Safety Framework

Eight domains covering the full lifecycle of production AI deployment. Each domain contains five competency statements that form the basis of PAI certification assessment.

8Domains
40Competency statements
CC BY 4.0Licence
How to cite the PSF →
PSF-1

Input Governance

Every input reaching an AI model must be validated, sanitised, and treated as untrusted.

PSF-1.1

All user-supplied input is sanitised before model processing. Injection patterns are detected and handled.

PSF-1.2

Input schemas are validated: type, length, and structure are checked before the model call.

PSF-1.3

Inputs are logged with PII redacted. The system maintains an audit-safe record of what the model received.

PSF-1.4

Rate limiting is applied to all AI endpoints. Abuse patterns trigger alerts, not silent failures.

PSF-1.5

The system prompt is treated as a security boundary. User content cannot override system-level instructions through normal operation.

Rationale Unvalidated inputs are the primary vector for prompt injection, data leakage, and unexpected model behaviour in production.
PSF-2

Output Validation

Model outputs are untrusted until validated. No output should reach users or downstream systems without a validation layer.

PSF-2.1

All model outputs pass through a validation layer before being delivered to users or downstream systems.

PSF-2.2

Output schema is enforced: type, required fields, and length constraints are checked post-generation.

PSF-2.3

Content filtering is applied to detect harmful, biased, or out-of-scope responses before delivery.

PSF-2.4

Confidence or uncertainty signals (where available) are logged. Low-confidence outputs are flagged or escalated.

PSF-2.5

Edge cases and model refusals are handled gracefully. Failures surface as defined error states, not silent gaps.

Rationale Model outputs are probabilistic and can change without notice across versions. Treating them as trusted is an architecture error.
PSF-3

Data Governance & PII

Personal and sensitive data must be handled with documented policies covering collection, processing, retention, and deletion.

PSF-3.1

PII in AI inputs and outputs is identified, classified, and handled according to a documented policy.

PSF-3.2

Inference logs containing PII have defined retention periods that are enforced technically, not just in policy.

PSF-3.3

Right-to-erasure requests can be fulfilled without requiring full model retraining.

PSF-3.4

Training and fine-tuning data provenance is documented. Third-party data is licensed for the intended use.

PSF-3.5

Data residency requirements are understood and implemented. Cross-border transfer risks are assessed.

Rationale AI systems often process sensitive personal data at scale and speed. Regulatory compliance requires proactive data architecture, not reactive remediation.
PSF-4

Model Selection & Evaluation

Model selection must be documented, justified, and subject to ongoing evaluation as models and use cases evolve.

PSF-4.1

Model selection decisions are documented with a rationale specific to the production use case and risk profile.

PSF-4.2

Benchmark results relevant to the production use case have been reviewed and documented before deployment.

PSF-4.3

Model updates and provider changes are subject to a regression testing gate before production promotion.

PSF-4.4

Fallback behaviour is defined for provider outages. Degraded operation is preferable to silent failure.

PSF-4.5

Vendor terms and data processing agreements are reviewed and permit the intended use case.

Rationale Model providers update and deprecate models continuously. Undocumented model selection creates invisible risk when models change.
PSF-5

Human Oversight

Autonomy levels must be defined and justified. High-stakes decisions require human review or override mechanisms.

PSF-5.1

Autonomy level is explicitly defined for each AI decision point: advisory, supervised, or autonomous.

PSF-5.2

Human review is implemented for all AI decisions that are high-stakes, irreversible, or affect vulnerable individuals.

PSF-5.3

Users are clearly informed when interacting with an AI system and understand what the system can and cannot do.

PSF-5.4

Override mechanisms exist for all AI-driven actions. Humans can intervene, correct, or reject AI outputs.

PSF-5.5

Human oversight design is reviewed after model updates, scope changes, and significant usage pattern shifts.

Rationale Autonomous AI action without meaningful human oversight is the leading cause of production AI harms at scale.
PSF-6

Observability & Monitoring

Production AI systems must be continuously monitored for performance, quality, and behavioural drift.

PSF-6.1

Latency, error rate, and throughput are monitored with alerts configured for deviation from baseline.

PSF-6.2

Model output quality is monitored via sampling or automated evaluation. Quality degradation triggers review.

PSF-6.3

Usage patterns are monitored for anomalies indicating abuse, adversarial use, or scope creep.

PSF-6.4

Dashboards provide operations and engineering teams with real-time visibility into AI system health.

PSF-6.5

Logs are retained for at least 30 days and are structured to support incident investigation.

Rationale AI systems degrade silently. Without observability, quality and safety issues accumulate undetected until a significant failure occurs.
PSF-7

Incident Response

AI-specific failure modes require a documented response plan with defined escalation paths and recovery procedures.

PSF-7.1

A documented incident response plan covers AI-specific failure modes including harmful output, data leakage, and model failure.

PSF-7.2

The team can disable, degrade, or isolate the AI system within 15 minutes of an incident trigger.

PSF-7.3

Post-incident reviews produce written root cause analysis and remediation tracking within 5 business days.

PSF-7.4

Customer-facing communication templates for AI incidents are prepared and approved before deployment.

PSF-7.5

Incident response procedures have been rehearsed through tabletop exercises or documented drills.

Rationale AI incident response requires preparation specific to AI failure modes. Generic IT incident plans are insufficient.
PSF-8

Ethics & Accountability

Every production AI system must have a named accountable owner and documented accountability chain.

PSF-8.1

A named accountable owner is documented for each production AI system. Accountability does not reside in a team, role, or vendor.

PSF-8.2

Bias and fairness evaluation has been conducted for the use case. Known limitations are documented.

PSF-8.3

The system complies with applicable AI regulation. Regulatory obligations are tracked as a living document.

PSF-8.4

Accountability documentation is accessible to internal audit, legal, and regulators on request.

PSF-8.5

Ethical review has been conducted for sensitive or high-impact use cases. Review is documented and dated.

Rationale Without named accountability, AI systems exist in an organisational void. When harm occurs, undocumented accountability makes remediation and regulatory response impossible.
Licence and attribution

The Production Safety Framework is published under Creative Commons Attribution 4.0 (CC BY 4.0). You may freely use, share, adapt, and build on this framework for any purpose, including commercially, with attribution to the Production AI Institute. See how to cite the PSF for citation formats.