Periodic Reporting for period 1 - Deeploy (FIRST MLOPS INTEGRATING REAL-TIME RISK MANAGEMENT, COMPLIANCE AND EXPLAINABILITY WHERE THE AI MODEL RUNS)
Reporting period: 2025-03-01 to 2026-02-28
The strategic significance extends to European technological sovereignty in AI governance, positioning EU-based solutions as alternatives to predominantly US-based AI infrastructure providers. This alignment with broader European digital strategy objectives amplifies the project's potential impact beyond individual organizational benefits to encompass regional competitive positioning in the global AI landscape.
Deeploy's AI governance platform addresses this critical gap by enabling real-time risk management, explainability, and compliance monitoring within operational AI environments. The project emerges at a pivotal moment: organisations face potential fines of up to 7% of global annual turnover under the EU AI Act, while the global AI governance market is projected to reach $5.1 billion by 2028, a critical inflection point where regulatory compliance shifts from best practice to business requirement.
An automated compliance evidence system covers the full AI lifecycle, including inference log archiving, guardrail violation tracking, metadata collection, and automated compliance checks, significantly reducing manual documentation burden.
A system for automated compliance evidence collection was developed, covering the full AI lifecycle from exploration through to retirement. This includes automated event and inference log archiving, guardrail violation tracking, key-value pair metadata collection, and automated binary and calculated compliance checks. This reduces manual documentation burden significantly.
Two Explainable AI (XAI) methods were developed and validated:
(1) a black-box method based on SHAP feature attribution, and
(2) a white-box method based on attention weights and gradient saliency, with specific implementation for generative Hugging Face models. Both methods were validated against EU AI Act requirements in collaboration with external legal and academic experts. A third XAI method targeting closed/inaccessible models is in development. Additionally, a bias detection pipeline was designed covering offline analysis, offline monitoring, online monitoring, and visualisation of sensitive attribute metrics.
A framework-agnostic model onboarding capability enables governance of any AI model (ML, GenAI, Agent) across Kubernetes (KServe), AzureML, and Sagemaker. Enterprise private cloud and hybrid SaaS architectures were validated with TVM, Novo Nordisk, and Healthplus.ai with multi-tenancy, role-based access control, and SSO.
1. Automated compliance evidence across the AI lifecycle. No integrated tooling previously existed to automatically capture, structure, and verify compliance evidence throughout development, validation, and production. Deeploy's evidence pipeline, combining inference logging, guardrail enforcement, lifecycle state tracking, and policy-bound metadata, constitutes a novel operational layer above existing MLOps infrastructure, removing a core bottleneck in regulated AI adoption.
2. Explainability for generative and closed models. Deeploy developed framework-agnostic token-level attribution methods (gradient saliency, attention weights) for generative models. Work is ongoing on XAI for fully closed/external models: a largely unsolved problem. The framework targets practically useful explanations for human oversight, balancing faithfulness against complexity.
3. Governance-native infrastructure. Rather than a post-hoc audit layer, Deeploy integrates control frameworks, evidence collection, explainability, and bias monitoring directly into AI deployment infrastructure — a departure from the current standard of manual, retroactive compliance
Potential impacts include reducing the cost of EU AI Act compliance in high-risk domains (healthcare, finance, pharma, public sector). The platform has already been validated with UMCU, Novo Nordisk, and Healthplus.ai.
Key needs for further uptake include:
(1) stable and specific technical guidance from the EU AI Office on high-risk AI validation requirements, which has been repeatedly delayed;
(2) standardisation of AI governance evidence formats to enable interoperability;
(3) further enterprise-scale demonstration projects in regulated sectors; and
(4) commercial partnerships across EU geographies to localise compliance support (currently active in Benelux, DACH, Nordics, and UK/Ireland).