It was a medical scoop for the Benelux countries. Anaesthetists at the UZ Brussels hospital made use of new artificial intelligence (AI) technology during an operation.
An algorithm, developed by an American supercomputer, analyses all sorts of parameters and predicts a drop in blood pressure, fifteen minutes before the event possibly occurs. The new technology is a telling example of how artificial intelligence can be a priceless asset in healthcare and medicine.
AI might become as significant to the 21st century, as was the Industrial Revolution to the 19th century. It is difficult to give a proper definition of AI, since it concerns a wide range of methods, applications and goals. Key to those is the development of an internal decision-making system, whose logic is almost impossible to translate into human language and reasoning. So AI is all about systems that develop autonomous thought-processes. Those thought-processes are deemed to be more objective and precise, as they are generated autonomously and contain more information than human brains process.
No Science Fiction
The AI-revolution is already underway. New technologies and applications are constantly being developed. Yet most people don’t even have the faintest idea of what it involves. And most politicians don’t have the faintest intention of prioritizing a technology that seems so impalpable.
Many tend to think of AI as something that belongs to the realm of science-fiction. But it’s real and it is happening now. All around us, also in Brussels. The first operations using the new AI-technology in UZ Brussels were successful, and it speaks volumes because everybody involved is enthusiastic about the prospects. When blood pressure fall occurs during an operation, it is no minor inconvenience. It can have long-term consequences and, in particular, have detrimental effects on the brain and the kidneys.
Blind Optimism and Luddism
Yet as always, with great possibilities often come great possible problems. And so it is with AI. One of the main challenges we face is finding the right balance between sufficiently stimulating technological developments that will have clear benefits, and having sufficient regulation in place to offset the risks.
We should not fear future developments, as the Luddites once did. But blind optimism that things will work out fine and regulate themselves is not justified either. We need political action and clear policies. The sooner the better.
A case in point is the application of AI for military purpose or security reasons. Weaponised AI will play a key role in how future wars are fought, and how peace can be maintained. But it raises major issues about where the responsibility in the decision-making process lies, and who is to be held accountable when drones launch a fatal attack that goes wrong.
In matters of security and surveillance, AI offers unique opportunities to foster peace and stability, to detect possible dangers even before they arise, to prevent terror attacks and criminal acts long before they are committed. But at what cost does such security come? How much monitoring can be permitted? Where is the tipping point, when maintaining free movement and expression in the public sphere and pursuing security and public safety have simply become impossible?[…]