Amandeep Singh Gill, the United Nations tech policy chief, speaks during an interview, Friday, Sept. 22, 2023, at UN headquarters. Gill fears corporate interests may undermine the push to rein in artificial intelligence, exacerbating social divisions and encroaching on human rights. THE CANADIAN PRESS/AP/Mary Altaffer
MONTREAL – The United Nations’ top tech official fears that corporate interests may undermine the push to rein in artificial intelligence, exacerbating social divisions and encroaching on human rights.
Countries could feel pressure to accommodate business demands for greater leeway rather than curbing industry excess, said Amandeep Gill in an interview ahead of a global AI conference in Montreal that begins Wednesday.
“I’m quite worried, frankly,” he said.
Human rights and democratic values are at stake, said Gill, the UN secretary-general’s tech envoy.
Researchers and political leaders have highlighted concerns ranging from biased data sets and widening global inequality to existential threats around sweeping cyber attacks and AI-developed bioweapons.
Artificial intelligence pioneer Yoshua Bengio, who founded the Mila AI institute, has sounded the alarm bells on immediate dangers such as “counterfeiting humans” using AI-driven bots.
Academics, advocates, business leaders and policymakers were in Montreal for the three-day gathering titled “Protecting Human Rights in the Age of AI” and hosted by Mila.
Power consolidation, prejudice and privacy are three of its core themes.
“As potentially there is more concentration of wealth and tech power in a few companies, then that has implications for social equity, for our social contract,” said Gill.
The impulse to get a leg up in the global technology race could conflict with the need to curtail the risks around rapid AI advances through laws and regulations, he said. Gill also stressed that AI oligopolies or concentration in a handful of countries would disadvantage smaller firms and developing nations.
Further polarization within and between states is another possible outcome of AI run amok.
“If AI is leveraged for deep fakes, for misinformation, disinformation at scale, that could undermine the legitimacy of political processes in our society,” Gill said.
Built-in bias is also an ongoing problem in the sector as algorithms that draw on reams of data sometimes magnify existing prejudices rather than sifting them out. “Those could be perpetuated over even worse,” he said, “whether it’s decisions being made on housing, on parole, in the judicial system or allocation of social benefits.”
Discriminatory outcomes are not limited to the national level, Gill added, noting that information fed into machine learning models stems largely from North America or Western Europe.
“That means a vast majority of world cultures, languages and contexts are not properly reflected in these data sets.”
Meanwhile, the power to track the online activity of citizens and social media users risks veering into violations of privacy rights.
In spite of the urgency to control cutting-edge AI, Gill said the United Nations must take a “modest” approach to establishing rules to encourage as many states as possible to sign on.
Last month, a UN advisory body released a preliminary report laying out the guiding principles for a framework on AI governance, stressing that no country be “left behind” as the pace of innovation nears light speed.
In Canada, the federal government introduced legislation placing guardrails around AI use in June 2022, but it has languished at the committee stage for nearly 10 months.
Big Tech executives said last week the Artificial Intelligence and Data Act is currently too vague, arguing that it fails to adequately distinguish between high- and low-risk AI systems.
The Liberals have said they will amend the act to introduce new rules, including requiring companies responsible for generative AI systems – the algorithmic engine behind chatbots such as OpenAI’s ChatGPT, which can spit out anything from math problems to marriage advice – to take steps toward ensuring their content is identifiable as AI-made.
The legislation still aims for a more general, principles-based approach to AI governance that allows for agility amid the technology’s constant evolution, leaving most details to a later date.
Ottawa has said the act known as Bill C-27 will come into force no sooner than 2025.
This report by The Canadian Press was first published Feb. 14, 2024.