AI researchers called upon to sign open letter for safe development

The Future of Life Institute (FLI) has published an open letter to AI engineers, hoping research is done in an ethical manner.

Published
Updated
44 seconds read time

Research into artificial intelligence (AI) continues to evolve, and there is growing concern that uncontrolled AI could have a significant impact on mankind. To prevent this from happening, the Future of Life Institute (FLI) wants AI researchers to sign an open letter to protect humans from intelligent machines.

AI researchers called upon to sign open letter for safe development | TweakTown.com

"We recommend expanded research aimed at ensuring that increasingly capable AI systems are robost and beneficial: our AI systems must do what we want them to do," the letter reads. "The attached research priorities document gives many examples of such research directions that can help maximize the societal benefit of AI. This research is by necessity interdisciplinary, because it involves both society and AI. It ranges from economics, law and philosophy to computer security, formal methods and, of course, various branches of AI itself."

AI is being used in autonomous weapons systems, robots and humanoids, and in autonomous vehicles - raising serious ethical questions that must be answered.

Notable letter signees include Tesla founder Elon Musk, Professor Stephen Hawking, Microsoft research director Eric Horvitz, and numerous other leading professors, engineering students, and private sector leaders.

An experienced tech journalist and marketing specialist, Michael joins TweakTown to cover everything from cars & electric vehicles to solar and green energy topics. A former Staff Writer at DailyTech, Michael is now the Cars & Electric Vehicles News Reporter and will contribute news stories on a daily basis. In addition to contributing here, Michael also runs his own tech blog, AlamedaTech.com, while he looks to remain busy in the tech world.

Newsletter Subscription

Related Tags