Periodic Reporting for period 2 - VERGE (AI-powered eVolution towards opEn and secuRe edGe architEctures)
Reporting period: 2024-01-01 to 2025-06-30
This main goal has been achieved through the following specific obejctives:
1.- Design an open and secure edge computing architecture to efficiently support highly demanding XR and IoT-driven applications in terms of computation and connectivity, relying on an edge-cloud continuum that fosters virtualisation of massive heterogeneous edge computing resources across a multi-access RAN.
2.-Design and build the Edge4AI concept that provides the infrastructure, mechanisms and programming tools to enable the development, deployment and orchestration of massively distributed AI processes across heterogeneous computing and memory resources.
3.-Design an AI4Edge framework that encompasses cutting-edge AI solutions for managing and orchestrating the edge computing and the RAN resources towards an optimum performance that satisfies the highly demanding applications requiring extremely low latencies and/or very high-capacity justifying edge processing and computing.
4.-Develop tools that ensure the security, privacy and trustworthiness of the VERGE system.
5. Showcase the VERGE solutions by means of Proof of Concept (PoC) demonstrations.
6. Carry out extensive dissemination, standardisation and exploitation activities.
The project solutions have impacted on different improvement areas, namely reduced latency, enhanced service availability, enhanced service reliability, efficient resource utilization, energy efficiency, flexibility/scalability, enhanced computation, efficient AI model training, AI safety and AI explainability.
• The consolidation of the AI/ML LCM, with intelligent AI4Edge solutions and SPT4AI trustworthiness mechanisms seamlessly integrated into a unified software pipeline.
• The consolidated design of the VERGE open dataspace has been successfully completed. A key achievement is the development of a standardized schema encompassing dataset definitions, metadata specifications, version control mechanisms, and governance policies.
•Accelerated and distributed computing capabilities, coupled with advanced learning paradigms.
• AI-driven hierarchical and modular orchestration, management, and control.
• A portfolio of AI-driven solutions for network slicing and resource optimization.
• AI trustworthiness solutions, addressing security, privacy, robustness, safety, and explainability challenges.
• Unified data access and LCM of all AI-driven pipelines through the VERGE Cognitive Framework.
• The planning, integration, and validation of the final project demonstrations, resulting in the production of seven PoC demonstrations by the end of the project.
• Reduction of up to 79% in the lines of code required to create distributed workflows, significantly enhancing programmability.
•Up to 45% power saving has be achieved by efficient resource allocation and function offloading leveraging FPGA programmability.
• Time-triggered Federated Learning (FL) methodologies ensuring up to 40% reduction in communication cost and faster convergence in resource-constrained environments.
•37% end-to-end latency reduction when the MEC app slice concept is employed to prioritize critical services.
• 36–97% latency reduction has been achieved through autoscaling capabilities of the distributed workflow orchestration (depending on the number of available computing resources), as well as through regression-based approaches for intent-driven edge-site resource allocation and rightsizing.
• Optimal MEC selection and network autoscaling, yielding up to 28% latency reduction and 98.5% service reliability without compromising utilization efficiency.
• Relay management and optimized low-layer functional split selection, leading to up to 23% cost reduction (compared to fixed-splitting solutions).
• Enhanced and generalizable attention-based mechanism for forecasting edge-cloud resource utilization, resulting in 59.4% and 93.4% improvements in prediction accuracy and inference speed, respectively, which could be leveraged by the multi-site orchestrator.
• Homology-based solutions leveraging semantic transmission achieving a 15-fold improvement in resource utilization.
• By employing transfer learning techniques, up to 80% training time reduction was achieved in a Deep Reinforcement Learning (DRL)-based network slicing solution when RAN topologies were expanded.
• Security mitigations for two AI-driven tasks (beam selection and power control) in a Distributed Multiple Input Multiple Output (D-MIMO) environment, decreasing the attack success rate to below 1% and mitigating poisoning attacks while maintaining data validity.
• A selective knowledge transfer approach reducing the attack success rate to 7% in collaborative hazard detection scenarios in untrusted transportation environments.
• A homomorphic encryption approach to avoid exposing private data during model aggregation reduced the leakage attack success rate to less than 10%, while maintaining model accuracy within a 3% margin.
• A framework for trustworthy RL was proposed, ensuring reliable decision-making with a strong trade-off between accuracy (92%) and robustness (94%) under previously unseen conditions.
• Causal RL methodologies demonstrated over 40% better sample efficiency than traditional methods and improved generalization.
• Formal verification framework enabling probabilistic estimation of whether key safety, reliability, and resilience properties are fulfilled.
• AI feature explainability for DRL models was supported through a Shapley Value-based approach.
• 75.6% reduction in total pipeline time for training and validation of an ML model was achieved through CF services, thanks to enhanced pipeline orchestration and parallel execution