Technology content trusted in North America and globally since 1999
8,225 Reviews & Articles | 62,328 News Posts

AI researchers called upon to sign open letter for safe development

The Future of Life Institute (FLI) has published an open letter to AI engineers, hoping research is done in an ethical manner
By: Michael Hatamoto | Science, Space & Robotics News | Posted: Jan 12, 2015 7:42 am

Research into artificial intelligence (AI) continues to evolve, and there is growing concern that uncontrolled AI could have a significant impact on mankind. To prevent this from happening, the Future of Life Institute (FLI) wants AI researchers to sign an open letter to protect humans from intelligent machines.




"We recommend expanded research aimed at ensuring that increasingly capable AI systems are robost and beneficial: our AI systems must do what we want them to do," the letter reads. "The attached research priorities document gives many examples of such research directions that can help maximize the societal benefit of AI. This research is by necessity interdisciplinary, because it involves both society and AI. It ranges from economics, law and philosophy to computer security, formal methods and, of course, various branches of AI itself."


AI is being used in autonomous weapons systems, robots and humanoids, and in autonomous vehicles - raising serious ethical questions that must be answered.


Notable letter signees include Tesla founder Elon Musk, Professor Stephen Hawking, Microsoft research director Eric Horvitz, and numerous other leading professors, engineering students, and private sector leaders.


Related Tags

Got an opinion on this news? Post a comment below!