Skip to main content

Header

SAFEXPLAIN

Safe and Explainable Critical Embedded Systems based on AI

The SAFEXPLAIN project will provide certified and economically viable solutions for European industry to enable fully-autonomous Critical Systems (e.g. cars, trains, satellites). It will bring increased efficiency of these systems and allow the European Critical Autonomous Systems industry to benefit from deep learning functionalities so they can remain competitive in the future while still being trustworthy.

Business Categories
Duration
-

Vision

Deep Learning (DL) techniques are key for most future advanced software functions in Critical Autonomous AI-based Systems (CAIS) in cars, trains and satellites. Hence, those CAIS industries depend on their ability to design, implement, qualify, and certify DL-based software products under bounded effort/cost. There is a fundamental gap between Functional Safety (FUSA) requirements of CAIS and the nature of DL solutions needed to satisfy those requirements. The lack of transparency (mainly explainability and traceability), and the data-dependent and stochastic nature of DL software clash against the need for deterministic, verifiable and pass/fail test based software solutions for CAIS.

SAFEXPLAIN tackles this challenge by providing a novel and flexible approach to allow the certification – hence adoption – of DL-based solutions in CAIS by

    - Architecting transparent DL solutions that allow explaining why they satisfy FUSA requirements, with end-to-end traceability, with specific approaches to explain whether predictions can be trusted, and with strategies to reach (and prove) correct operation, in accordance with certification standards.
   - Devising alternative and increasingly complex FUSA design safety patterns for different DL usage levels (i.e. with varying safety requirements) that will allow using DL in any CAIS functionality, for varying levels of criticality and fault tolerance.

 

Assets related to SAFEXPLAIN