Image Credit: gopixa/Getty Images
So the argument goes. But I embrace the legislation and the risk-based approach the EU has taken.
Why should I care? I live in the UK, and my company, Healx, which uses AI to help discover new treatment opportunities for rare diseases, is based in Cambridge.
EU General Data Protection Regulation (GDPR) laws required just about every company with a website either side of the Atlantic to react and adapt to them when they were rolled out in 2016. It would be naive to think that any company with an international outlook won’t run up against these proposed rules too. If you want to do business in Europe, you will still have to adhere to them from outside it.
And for areas like health, this is incredibly important. The use of artificial intelligence in healthcare will almost inevitably fall under the “high risk” label. And rightly so: Decisions that affect patient outcomes change lives.
That’s why the legislation’s focus on reducing bias in AI, and setting a gold standard for building public trust, is vital for the industry. If an AI system is fed patient data that does not accurately represent a target group (women and minority groups are often underrepresented in clinical trials), the results can be skewed.
That damages trust, and trust is crucial in healthcare. A lack of trust limits effectiveness. That’s part of the reason such large swathes of people in the West are still declining to get vaccinated against COVID. The problems that’s causing are plain to see.
AI breakthroughs will mean nothing if patients are suspicious of a diagnosis or therapy produced by an algorithm, or don’t understand how conclusions have been drawn. Both result in a damaging lack of trust.
In 2019, Harvard Business Review found that patients were wary of medical AI even when it was shown to out-perform doctors, simply because we believe our health issues to be unique. We can’t begin to shift that perception without trust.
The next step won’t just be to build on these breakthroughs but to build trust so that they can be implemented safely, without disregarding vulnerable groups, and with clear transparency, so worried individuals can understand how a decision has been made.
This is something that will always, and should always, be monitored. That’s why we should all take notice of the spirit of the EU’s proposed AI legislation, and embrace it, wherever we operate.
Tim Guilliams is a co-founder and CEO of drug discovery startup Healx.