Explainable AI: Beware What You Wish For

2nd September 2022

Can ‘Explainable AI’ provide a way to apply DO-178C for the safety assurance of neural networks in aviation applications? We think not, but there are a few subtle misconceptions to be unwrapped.

However, there does exist a different approach for ensuring the reliable performance of machine-learned (‘AI’) components for a new generation of flight control instruments: demonstrating Machine Learning Generalization and quantifying so-called “Domain Gaps”.

This 8-minute video provides a synopsis of the concepts for anyone who has a basic understanding of high school-level math.

00:19 How DO-178 works for the classical software systems in aviation
01:00 What is that ‘AI’ you speak of
03:44 Certainty and uncertainty
04:56 Traceability: Beware of what you wish for
06:16 How to design a system with a component working with uncertainty
06:56 Will it work ‘in vivo’ as well as it worked ‘in vitro’?


Visit Supplier

Visit Supplier Website

Contact Daedalean


Daedalean AG
Albisriederstrasse 199
8047 Zürich

More News

Get in touch

    We'd love to send you the latest news and information from the world of Future Transport-News. Please tick the box if you agree to receive them.

    For your peace of mind here is a link to our Privacy Policy.

    By submitting this form, you consent to allow Future Transport-News to store and process this information.