Latest News

AI will match human intelligence by 2062, UNSW professor says

The idea that Artificial Intelligence (AI) will learn unique human traits like adaptability, creativity and emotional intelligence is something that many in society consider to be an unlikely or distant possibility.

But Toby Walsh, Scientia professor of AI at UNSW Sydney, has put a date on would could be reality.

He considers 2062 the year that artificial intelligence will match human intelligence, although a fundamental shift has already occurred in the world as we know it.

Speaking at the Festival of Dangerous Ideas, Walsh argued that we are already experiencing the risks of artificial intelligence that seem to be so far in the future.

READ: Study shows how AI and blockchain could revolutionise housing

“Even without machines that are very smart, I’m starting to get a little bit nervous about where it’s going and the important choices we should be making,” said Walsh.

The key challenge will be to avoid the apocalyptic rhetoric of AI and to determine how to move forward in the new age of information.

Privacy concerns about the collection of personal data is nothing new.

Citing the Cambridge Analytica scandal, Walsh argues that we should be more sceptical about how data is misused by tech companies.

“A lot of the debate has focused on how personal information was stolen from people, and we should be rightly outraged by that,” said Walsh.

“But there is another side to the story that I’m surprised hasn’t gotten as much attention from the media, which is that the information was used very actively to manipulate how people were going to vote.”

Information is the currency of today’s tech giants, and there is a growing fear that many people are in denial, or even complicit, in just how much data is collected about themselves on a daily basis.

Walsh said breaches of data privacy will occur more often and are becoming increasingly normalised.

“Many of us have smartwatches that are monitoring our vital signs; our blood pressure, our heartbeat, and if you look at the terms of service, you don’t own that data,” he said.

“We’re giving up our analogue privacy, the most personal things about us. Just think what you could do as an advertiser if you could tell how people really respond to your adverts.

“You can lie about your digital preferences, but you can’t lie about your heartbeat,” said Walsh.

Untangling the ethics of machine accountability will be the second fundamental shift in the world as we know it, said Walsh.

“Fully autonomous machines will radically change the nature of warfare and will be the third revolution in warfare.”

But using autonomous machines as weapons of war poses an ethical dilemma – can you hold a machine accountable for death?

“Machines have no moral compass, they are not sentient, they don’t suffer pain and they can’t be punished,” he said.

“This takes us into interesting new legal territory of who should be held responsible, and there is no simple answer.”

AI is developed by learning from examples – therefore the key driver of its behaviour is the environment that it is exposed to, more so than the programmer.

Walsh believes the issue is creating machines that are aligned with human values, which is currently a problem on other platforms driven by AI.

“Facebook is an example of the alignment problem, it is optimised for your attention, not for creating political debate or for making society a better place,” he said.

But it’s not all doom and gloom, according to Walsh. Artificial Intelligence isn’t necessarily heading towards an apocalyptic scenario.

“The future is not fixed. There is this idea that technology is going to shape our future and that we are going to have to deal with it, but this is the wrong picture to think of because society gets to push back and change the technology,” he said.

Instead of being proponents of technological determinism, Walsh argued that we need to push for societal determinism, ensuring that we build trustworthy systems with distinct lines of accountability

Send this to a friend