This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revision | |||
| en:safeav:softsys:vaicomp [2026/04/24 09:42] – raivo.sell | en:safeav:softsys:vaicomp [2026/04/29 16:12] (current) – raivo.sell | ||
|---|---|---|---|
| Line 1: | Line 1: | ||
| + | ====== Open issues of validating AI components ====== | ||
| + | |||
| + | |||
| + | ===== AI COMPONENT VALIDATION ===== | ||
| + | |||
| + | Both the automotive and airborne spaces have reacted to AI by viewing it as “specialized Software” in standards such as ISO 8800 [14] and [13]. This approach has the great utility of leveraging all the past work in generic mechanically safety and past work in software validation. However, now, one must manage the issue of how to handle the fact that we have a data generated “code” vs conventional programming code. In the world of V&V, this difference is manifested in three significant aspects: coverage analysis, code reviews, and version control. | ||
| + | |||
| + | ^ V&V Technique ^ Software ^ AI/ML ^ | ||
| + | | Coverage Analysis | Code Structure provides basis of coverage | No structure | | ||
| + | | Code Reviews | Crowd source expert knowledge | No Code to Review | | ||
| + | | Version Control | Careful construction/ | ||
| + | |||
| + | These differences generate an enormous issue for intelligent test generation and any argument for completeness. | ||
| + | 1) Training Set Validation: | ||
| + | 2) Robustness to Noise: | ||
| + | Overall, developing robust methods for AI component validation is quite an active and unsolved research topic for “fixed” function AI components. That is, AI components where the function is changing with active version control. | ||
| + | |||
| + | ===== AI SPECIFICATION ===== | ||
| + | |||
| + | For well-defined systems with an availability of system level abstractions, | ||
| + | 1) Anti-Spec | ||
| + | In these situations, the only approach left is to specify correctness through an anti-spec. The simplest anti-spec is to avoid accidents. Based on some initial work by Intel, there is a standard, IEEE 2846, “Assumptions for Models in Safety-Related Automated Vehicle Behavior” [18] which establishes a framework for defining a minimum set of assumptions regarding the reasonably foreseeable behaviors of other road users. For each scenario, it specifies assumptions about the kinematic properties of other road users, including their speed, acceleration, | ||
| + | 2) AI-Driver | ||
| + | While IEEE 2846 comes from a bottom-up technology perspective, | ||
| + | |||
| + | a) Full Driving Capability: The AI driver must handle the entire driving task, including perception (sensing the environment), | ||
| + | b) Safety Assurance: Koopman stresses that AVs need rigorous safety standards, similar to those in industries like aviation. This includes identifying potential failures, managing risks, and ensuring safe operation even in the face of unforeseen events. | ||
| + | c) Human Equivalence: | ||
| + | d) Ethical and Legal Responsibility: | ||
| + | e) Testing and Validation: Koopman emphasizes the importance of robust testing, simulation, and on-road trials to validate AI driver systems. This includes covering edge cases, long-tail risks, and ensuring that systems generalize across diverse driving conditions. | ||
| + | Overall, it is a very ambitious endeavor and there are significant challenges to building this specification of a reasonable driver. First, the idea of a “reasonable” driver is not even well encoded on the human side. Rather, this definition of “reasonableness” is built over a long history of legal distillation, | ||
| + | Currently, the state-of-art for specification is relatively poor for both ADAS and AV. ADAS systems, which are widely proliferated, | ||
| + | |||