Home / Gaming / Test your ethics around killer AI cars with this Moral Machine game

Test your ethics around killer AI cars with this Moral Machine game



Should a self-driving car full of old folks crash to avoid puppies in the cross-walk? Is it OK to run over two criminals if you save one doctor? Whose lives are worth more, seven-year-olds or senior citizens?

This new game called the “Moral Machine” from MIT’s researchers lets you make the calls in the famous “trolley problem” and see analytics about your ethics. Thinking about these tough questions is more important than ever since engineers are coding this type of decision making into autonomous vehicles right now. Who should be responsible for these choices? The non-driving passenger, the company who made the AI or no one?

ethics-results

The Moral Machine poses basic choices like whether AI should intervene at all if it will save more lives, or if it should stay passive instead of actively changing events in a way that makes it responsible for someone’s death.

But the scenarios also raise more nuanced quandaries. Should pedestrians be saved instead of passengers who knowingly got into an inherently dangerous speeding metal box? Should we rely on the air bags and other safety features in a crash instead of swerving into unprotected pedestrians?

If you think these decisions are tough with clear-cut situations, imagine how tough it will be for self-driving cars amidst chaotic road conditions.



Source link

About admin

Check Also

landscape06.png

Games developer Supersolid raises $4M Series A round led by Index Ventures

Supersolid, a London-based mobile games developer with more than 50 million downloads under its belt, ...

Leave a Reply

Your email address will not be published. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <s> <strike> <strong>