AI systems work with large amounts of data. How can this data be measured and evaluated to develop reliable quality standards? What data do you need to collect?
Schwabe: We will look at specific reference examples for different data classifications. First, we will study which aspects of data can be analyzed using simple statistical methods, for example. This ranges from the simple presence or absence of data points, to examining distribution functions, to class balance analysis, for example.
For example, we want to examine the distribution of disease classification in the dataset. How does AI address rare diseases or "cases" of the disease? How reliably does an AI technology work on a data set where disease classification may be underrepresented? Do we have to introduce certain minimum requirements in this setting? Do we need to set a lower limit for the number of data points that must be available for the AI technology to reliably detect a disease?
What would an AI platform look like that does not comply with the approval processes?
Schwabe: There are many ways to make mistakes, even as it pertains to general aspects such as security issues, data protection, or aspects that were not adequately captured during development. Another concern is utilizing the right data and tools for the targeted area of application. You must show that the data selection is suitable.
One must examine the AI models to draw conclusions about how they reach their decisions. Famous examples in research have shown that AI detected tumors in X-ray images. In other words, AI has acquired a very high detection rate.