As artificial intelligence continues to seep slowly into medical practices around the world, how can we bridge the gap between the systems being developed by research and industry, and the clinics, where take-up is not yet widespread? A team of University of Amsterdam researchers at Amsterdam Science Park looking at the use of AI in ophthalmology believes the key lies in the trustworthiness of the AI, as well as in involving all relevant stakeholders at every stage of the production process. Their study, already available in an open access version, will soon appear in the prestigious publication Progress in Retinal and Eye Research.
In ophthalmology there are currently only a small number of systems regulated and even those are very seldom used. Despite achieving performance close to or even superior to that of experts, there is a critical gap between the development and integration of AI systems in ophthalmic practice. The research team looked at the barriers preventing use and how to bring them down. They concluded that if the systems were finally to see widespread use in actual medical practice, the main challenge was to ensure trustworthiness. And that to become trustworthy they need to satisfy certain key aspects: they need to be reliable, robust and sustainable over time.
Study author Cristina González Gonzalo: ‘Bringing together every relevant stakeholder group at each stage remains the key. If each group continues to work in silos, we’ll keep ending up with systems that are very good at one aspect of their work only, and then they’ll just go on the shelf and no one will ever use them.’
Stakeholders for AI in ophthalmology include AI developers, reading centres, healthcare providers, healthcare institutions, ophthalmological societies and working groups or committees, patients, regulatory bodies, and payers. With the interests of so many groups to take into account, the team created an ‘AI design pipeline’ (see image) in order to gain the best overview of the involvement of each group in the process. The pipeline identifies possible barriers at the various stages of AI production and shows the necessary mechanisms to address them, allowing for risk anticipation and avoiding negative consequences during integration or deployment.
Among the various challenges involved, the team realised ‘explainability’ would be an one of the most important elements in achieving trustworthiness. The so-called ‘black box’ around AI needed opening up. ‘The black box’ is a term used to describe the impenetrability of much AI. The systems are given data at one end and the output is taken from the other, but what happens in between is not clear. González Gonzalo: ‘For example, a system that gives a binary answer – ‘Yes, it’s a cyst’ or ‘No, it’s not a cyst’ – won’t be easily trusted by clinicians, because that’s not how they’re trained and not how they work in daily practice. So we need to open that out. If we provide clinicians with meaningful insight into how the decision has been made, they can work in tandem with the AI and incorporate its findings in their diagnosis.’
This publication has been reproduced with permission from the University of Amsterdam. Source
If you want to read more about AI research at Amsterdam Science Park, go to our page with everything about AI & Data science
Looking for partners to collaborate. Or looking for a certain expertise? Or would you like to locate your business in the Amsterdam Science Park? Drop us a line and we help you to find a perfect match.
Subscribe to our LinkedIn-newsletter and we will keep you updated on all that our park has to offer. It will arrive to your inbox five times a year and you can unsubscribe easily at any time.
Subscribe