The more artificial intelligence becomes an indispensable part of our lives, the more risks we face regarding our privacy. While some experts say AI should be regulated, others point at bigger problems behind its rise in popularity.

AI was a better target for regulators than the more abstract concept of privacy, said Sean McGregor, a researcher who chronicles tech failures for the AI Incident Database.
AI was a better target for regulators than the more abstract concept of privacy, said Sean McGregor, a researcher who chronicles tech failures for the AI Incident Database. (Violeta Stoimenova / Getty Images)

Artificial intelligence (AI) has seeped into every aspect of modern lives – from "intelligent" vacuum cleaners and self-driving cars to advanced methods for diagnosing diseases.

Its promoters reckon it is revolutionising human experience, but critics stress that the technology risks putting machines in charge of life-changing decisions.

Regulators in Europe and North America are worried.

The European Union is likely to pass legislation next year –– the AI Act –– aimed at reining in the age of the algorithm.

The United States recently published a blueprint for an AI Bill of Rights. And Canada is also mulling similar legislation.

For instance, China's use of biometric data, facial recognition and other technology to build a powerful system of control is looming large in the debates.

Gry Hasselbalch, a Danish academic who advises the EU on the controversial technology, argued that the West was also in danger of creating "totalitarian infrastructures".

"I see that as a huge threat, no matter the benefits," she told AFP.

But before regulators can act, they face the daunting task of defining what AI actually is.

'Mug's game'

Suresh Venkatasubramanian of Brown University, who co-authored the AI Bill of Rights, said trying to define AI was "a mug's game".

Any technology that affects people's rights should be within the scope of the bill, he tweeted.

The 27-nation EU is taking the more tortuous route of attempting to define the sprawling field.

Its draft law lists the kinds of approaches defined as AI and includes pretty much any computer system involving automation.

The problem stems from the changing use of the term AI.

For decades, it described attempts to create machines that simulated human thinking.

But funding largely dried up for this research –– known as symbolic AI –– in the early 2000s.

The rise of the Silicon Valley titans saw AI reborn as a catch-all label for their number-crunching programs and the algorithms they generated.

This automation allowed them to target users with advertising and content, helping them to make hundreds of billions of dollars.

"AI was a way for them to make more use of this surveillance data and to mystify what was happening," Meredith Whittaker, a former Google worker who co-founded New York University's AI Now Institute, told AFP.

So the EU and US have both concluded that any definition of AI needs to be as broad as possible.

'Too challenging'

But from that point, the two Western powerhouses have primarily gone their separate ways.

The EU's draft AI Act runs to more than 100 pages.

Among its most eye-catching proposals are the complete prohibition of certain "high-risk" technologies –– the kind of biometric surveillance tools used in China.

It also drastically limits the use of AI tools by migration officials, police and judges.

Hasselbach said some technologies were "simply too challenging to fundamental rights".

On the other hand, the AI Bill of Rights is a brief set of principles framed in aspirational language, with exhortations like "you should be protected from unsafe or ineffective systems".

The bill was issued by the White House and relied on existing law.

Experts reckon no dedicated AI legislation is likely in the United States until 2024 at the earliest because Congress is deadlocked.

'Flesh wound'

Opinions differ on the merits of each approach.

"We desperately need regulation," Gary Marcus of New York University told AFP.

He points out that "large language models" –– the AI behind chatbots, translation tools, predictive text software and much else –– can be used to generate harmful disinformation.

Whittaker questioned the value of laws aimed at tackling AI rather than the "surveillance business models" that underpin it.

"If you're not addressing that at a fundamental level, I think you're putting a band-aid over a flesh wound," she said.

But other experts have broadly welcomed the US approach.

AI was a better target for regulators than the more abstract concept of privacy, said Sean McGregor, a researcher who chronicles tech failures for the AI Incident Database.

But he said there could be a risk of over-regulation.

"The authorities that exist can regulate AI," he told AFP, pointing to the likes of the US Federal Trade Commission and the housing regulator HUD.

But where experts broadly agree is the need to remove the hype and mysticism surrounding AI technology.

"It's not magical," McGregor said, likening AI to a highly sophisticated Excel spreadsheet.

READ MORE: Can an invention enabled by artificial intelligence be patented?

READ MORE: Can an AI-led Danish party usher in an age of algorithmic politics?

READ MORE: Rigged AI: How algorithms fuel anti-Muslim hatred in Europe and beyond

Source: TRTWorld and agencies