Share it with your friends Like

Thanks! Share it with your friends!


Day 1 Session 3: Building Morality into Machines

:00 – Matthew Liao Opening Remarks
1:52 – Stephen Wolfram “How to Tell AIs What to Do (and What to Tell Them)”
38:20 – Francesca Rossi “Ethical Embodied Decision Making”
1:13:30 – Peter Railton “Machine Morality: Building or Learning?”
1:47:00 – Speaker panel

More info:

On October 14-15, 2016, the NYU Center for Mind, Brain and Consciousness in conjunction with the NYU Center for Bioethics hosted a conference on “The Ethics of Artificial Intelligence”.

Recent progress in artificial intelligence (AI) makes questions about the ethics of AI more pressing than ever. Existing AI systems already raise numerous ethical issues: for example, machine classification systems raise questions about privacy and bias. AI systems in the near-term future raise many more issues: for example, autonomous vehicles and autonomous weapons raise questions about safety and moral responsibility. AI systems in the long-term future raise more issues in turn: for example, human-level artificial general intelligence systems raise questions about the moral status of the systems themselves.

This conference will explore these questions about the ethics of artificial intelligence and a number of other questions, including:

What ethical principles should AI researchers follow?
Are there restrictions on the ethical use of AI?
What is the best way to design AI that aligns with human values?
Is it possible or desirable to build moral principles into AI systems?
When AI systems cause benefits or harm, who is morally responsible?
Are AI systems themselves potential objects of moral concern?
What moral framework and value system is best used to assess the impact of AI?



Write a comment


DARPA SUPERHIT 2021 Play Now!Close


(StoneBridge Mix)

Play Now!