Could artificial intelligence really wipe out humanity?

Could artificial intelligence really wipe out humanity?

Many fear that artificial intelligence will be the end of humankind – here’s the truth according to experts.

By now, most people around the world use some sort of AI-utilizing device that is integrated into their daily lives.

They use Siri to check the weather, or ask Alexa to turn off their smart lights – these are all forms of AI that many people don’t realize.

However, despite the widespread (and relatively harmless) use of this technology in nearly every facet of our lives, some people still seem to believe that machines could one day wipe out humanity.

This apocalyptic ideal has been perpetuated through various texts and movies over the years.

Even staple figures in the field of science such as Stephen Hawking and Elon Musk have been vocal about technology’s threat against humanity.

In 2020, Musk told the New York Times that AI would grow vastly smarter than humans and would overtake the human race by 2025, adding that things would get “unstable or weird.”

Despite Musk’s prediction, most experts in the field say humanity has nothing to worry about when it comes to AI – at least, not yet.

Most AI is “narrow”

The fear of AI taking over has developed from the idea that machines will somehow gain consciousness and turn on their creators.

In order for AI to achieve this, it would not only need to possess human-like intelligence, but it would also need to be able to predict the future or plan ahead.

As it stands, AI is not capable of doing either.

When prompted with the question “Is AI an existential threat to humanity,” Matthew O’Brien, a robotics engineer from the Georgia Institute of Technology wrote on Metafact: “The long-sought goal of a ‘general AI’ is not on the horizon. We simply do not know how to make a general adaptable intelligence, and it’s unclear how much more progress is needed to get to that point”.

The facts of the matter are that machines generally operate how they’re programmed to and we are a long way from developing the ASI (artificial superintelligence) needed for this “takeover” to even be feasible.

At present, most of the AI technology utilized by machines is considered “narrow” or “weak,” meaning it can only apply its knowledge towards one or a few tasks.

“Machine learning and AI systems are a long way from cracking the hard problem of consciousness and being able to generate their own goals contrary to their programming,” George Montanez, a data scientist at Microsoft, wrote under the same Metafact thread.

Read more