Figure 1: Real-time drift analysis triggering the automated circuit breaker.
Figure 2: The AI-model-quality-analyzer Interface showing feature drift quantification and automated decision output.
AI-model-quality-analyzer is an ML reliability control plane that enforces data contracts, detects statistically significant drift, and executes deterministic remediation actions in production-style pipelines.
Rather than treating observability as a dashboarding problem, AI-model-quality-analyzer encodes failure-handling logic directly into the system: invalid data is rejected, biased models are blocked, and high-risk deployments are rolled back automatically.
This project demonstrates applied MLOps practices across data validation, model governance, statistical testing, and operational decisioning—the layers where real-world ML systems typically fail.
The system is implemented as an independent reliability middleware positioned between data ingestion and model inference. This separation ensures that governance and remediation logic remain isolated from model code.
flowchart LR
A[Inbound Data Stream] -->|Validate| B{Data Contract Gate}
B -- Invalid Schema --> C[REJECT: HTTP 400]
B -- Valid Schema --> D[Analysis Engine]
subgraph Core Logic
D --> E[Statistical Tests]
D --> F[Fairness Auditor]
D --> G[Financial Risk Calc]
end
G --> H{Decision Matrix}
H -- Target Drift > 0.1 --> I[CRITICAL: Rollback]
H -- Bias Detected --> J[BLOCK: Compliance]
H -- Weighted Risk > 65 --> K[WARN: Shadow Mode]
H -- Nominal --> L[PASS: Deployment]
AI-model-quality-analyzer shifts the focus from passive observation to active reliability engineering.
| Feature | Traditional Monitoring | AI-model-quality-analyzer |
|---|---|---|
| Logic Model | Passive Observation | Active Deterministic Remediation |
| Alerting | Threshold-based Noise | Business-Impact Weighted |
| Data Quality | Post-Mortem Debugging | Pre-Ingestion Data Contracts |
| Metrics | Aggregate Drift Scores | Target-Aware Risk Scoring |
| Governance | Manual Review | Automated Fairness Circuit Breakers |
The platform converts statistical signals into binary operational actions using the following logic gates:
| Signal Severity | Trigger Condition | System Action | Operational Impact |
|---|---|---|---|
| CRITICAL | Target Drift > 0.1 |
ROLLBACK | Immediate traffic termination to prevent invalid inference. |
| HIGH | DIR < 0.8 |
BLOCK | Deployment halted due to violation of 4/5ths fairness rule. |
| MEDIUM | Risk Score > 65 |
SHADOW | Traffic routed to canary model for parallel evaluation. |
| LOW | Contract Violation |
REJECT | Ingestion API returns 400 Error to upstream producer. |
AI-model-quality-analyzer translates technical drift metrics into estimated financial impact using a heuristic cost-basis model.
Where:
- Volume: Throughput of the current batch.
- AvgCost: Business cost of a False Prediction ($150.00).
- D: Drift Score (0.0 - 1.0).
- Alpha/Beta: Correlation coefficients for feature vs. target drift.
Example: For a batch of 10,000 predictions with a 0.25 target drift score, the estimated revenue risk exceeds $375,000, triggering automatic rollback.
| Component | Technology | Role |
|---|---|---|
| API Server | FastAPI + Uvicorn |
High-concurrency asynchronous ingestion. |
| Validation | Pydantic |
Strict schema enforcement and type checking. |
| Computation | SciPy + NumPy |
Kolmogorov-Smirnov tests and P-Value calculation. |
| Drift Detection | Evidently AI |
Statistical profiling and distance measurement. |
| State Store | SQLite |
Audit logging, versioning, and cooldown management. |
| Frontend | Vanilla JS + CSS3 |
Lightweight, dependency-free visualization layer. |
AI-model-quality-analyzer is intentionally designed as a reliability and governance layer rather than a full inference-serving system.
The platform guarantees:
- Validated, schema-safe data ingestion.
- Statistically justified drift detection.
- Deterministic and explainable remediation decisions.
Note: Inference latency optimization and online serving are intentionally out of scope to preserve auditability, safety, and statistical rigor.
# 1. Clone Repository
git clone https://github.com/SalmaTech-03/Ai-model-quality-analyzer.git
# 2. Setup Environment
python -m venv venv
source venv/bin/activate
pip install -r requirements.txt
# 3. Hydrate Data
python scripts/download_data.py
# 4. Launch Service
uvicorn app.main:app --reloaddocker-compose up --buildAI-model-quality-analyzer exposes an embedded SQL engine for root-cause analysis on ingested batches.
Endpoint: POST /api/sql
SELECT occupation, COUNT(*) as volume
FROM current_table
WHERE income = '>50K'
GROUP BY occupation
ORDER BY volume DESC
LIMIT 5;├── app/
│ ├── api/ # API Route Definitions
│ ├── core/ # Mathematical & Logic Engines
│ ├── static/ # Dashboard Assets
│ └── main.py # App Entry Point
├── data/ # Local Data Storage
├── tests/ # Pytest Suite
├── docker-compose.yml # Container Orchestration
└── requirements.txt # Dependency Manifest
- Data Contracts: Enforced via strict Pydantic schemas.
- Statistical Rigor: Automated drift detection with significance testing (KS test, p-values).
- Model Governance: Fairness constraints via Disparate Impact Ratio.
- Safety Engineering: Deployment safety through circuit breakers and rollback logic.
- Reproducibility: Environments managed using Docker and pinned dependencies.
- Auditability: State and decisions persisted in a relational store.
Developed by Salma S
ML Engineering / MLOps Systems
