Artificial intelligence (AI) can be trained to detect whether or not a tissue picture contains a tumour. However, until recently, it has remained a mystery as to how it makes its judgement. A team from Ruhr-Universitat Bochum’s Research Center for Protein Diagnostics (PRODI) is working on a new approach that will make an AI’s judgement clear and hence trustworthy. The researchers led by Professor Axel Mosig describe the approach in the journal Medical Image Analysis. For the study, bioinformatics scientist Axel Mosig cooperated with Professor Andrea Tannapfel, head of the Institute of Pathology, oncologist Professor Anke Reinacher-Schick from the Ruhr-Universitat’s St. Josef Hospital, and biophysicist and PRODI founding director Professor Klaus Gerwert. The group developed a neural network, i.e. an AI, that can classify whether a tissue sample contains tumour or not. To this end, they fed the AI a large number of microscopic tissue images, some of which contained tumours, while others were tumour-free. “Neural networks are initially a black box: it’s unclear which identifying features a network learns from the training data,” explains Axel Mosig. Unlike human experts, they lack the ability to explain their decisions. “However, for medical applications in particular, it’s important that the AI is capable of explanation and thus trustworthy,” adds bioinformatics scientist David Schuhmacher, who collaborated on the study. The Bochum team’s explainable AI is therefore based on the only kind of meaningful statements known to science: on falsifiable hypotheses. If a hypothesis is false, this fact must be demonstrable through an experiment. Artificial intelligence usually follows the principle of inductive reasoning: using concrete observations, i.e. the training data, the AI creates a general model on the basis of which it evaluates all further observations. The underlying problem had been described by philosopher David Hume 250 years ago and can be easily illustrated: No matter how many white swans we observe, we could never conclude from this data that all swans are white and that no black swans exist whatsoever. Science therefore makes use of so-called deductive logic. In this approach, a general hypothesis is the starting point. For example, the hypothesis that all swans are white is falsified when a black swan is spotted. Source: ANI
Related Articles
Indian Americans
British Indian doctor’s rock band embarks on UK-wide tour
LONDON (TIP): A 25-year-old British Indian doctor whose passion for music resulted in a band with some of his fellow medical colleagues has embarked on his first UK-wide tour, with the launch show in London […]
NY/NJ/CT
T-Mobile Data Breach Sparks Alert From New York AG James
The breach impacted more than 53 million customers, including more than 4 million New Yorkers. NEW YORK (TIP): In an “urgent message,” New York Attorney General Letitia James confirmed that stolen sensitive customer information was […]
Indian Americans
Indian man killed in Singapore workplace accident
SINGAPORE (TIP): A 33-year-old Indian worker died after being struck by a reversing vehicle at a worksite in Jurong West industrial region here. The man was preparing his tipper truck for unloading when he was […]

Be the first to comment