🚨BREAKING NEWS ALERT 🚨This new search engine is amazing!🔥🔥🔥🔥 BOOM🔥...😎👉Click here!!! 🚨🚀🚀🚀🚀🚀🚀❤👋
From a young age, humans are capable of developing moral competency and autonomy through experience. We begin life by constructing sophisticated moral representations of the world that allow for us to successfully navigate our way through complex social situations with sensitivity to morally relevant information and variables. This capacity for moral learning allows us to solve open-ended problems with other persons who may hold complex beliefs and preferences. As AI systems become increasingly autonomous and active in social situations involving human and non-human agents, AI moral competency via the capacity for moral learning will become more and more critical. On this episode of the AI Alignment Podcast, Peter Railton joins us to discuss the potential role of moral learning and moral epistemology in AI systems, as well as his views on metaethics. Topics discussed in this episode include: -Moral epistemology -The potential relevance of metaethics to AI alignment -The importance of moral learning in AI systems -Peter Railton’s, Derek Parfit’s, and Peter Singer’s metaethical views You can find the page for this podcast here: https://futureoflife.org/2020/08/18/peter-railton-on-moral-learning-and-metaethics-in-ai-systems/ Have any feedback about the podcast? You can share your thoughts here: www.surveymonkey.com/r/DRBFZCT Follow the podcast on: Spotify: https://open.spotify.com/show/2Op1WO3gwVwCrYHg4eoGyP Apple Podcasts: https://podcasts.apple.com/us/podcast/future-of-life-institute-podcast/id1170991978?mt=2 SoundCloud: https://soundcloud.com/futureoflife Timestamps: 0:00 Intro 3:05 Does metaethics matter for AI alignment? 22:49 Long-reflection considerations 26:05 Moral learning in humans 35:07 The need for moral learning in artificial intelligence 53:57 Peter Railton’s views on metaethics and his discussions with Derek Parfit 1:38:50 The need for engagement between philosophers and the AI alignment community [More]
In the first of our 2021 Thomas Telford Lecture Series, Management, Procurement and Law editor, Dr Simon Smith talks to Philip McAleenan about his award winning paper on the moral responsibility and action in the use of artificial intelligence in construction. It has been said that AI is with us and that there is no reversing progress in this field. Whether the latter is true or not, it is certainly true that AI has entered the construction industry and there are technological developments occurring in various realms of the industry from the design process to production and logistics, and to the on-site monitoring of the speed of construction, efficiency and worker and plant interactions. The developments are promoted not only as a benefit to the industry but necessary to its development into the future. This presentation will not negate such benefits, but will instead give consideration to the unintended consequences of a system that is capable and becoming more capable all the time of making decisions independently of human operators. It will examine the potential for harmful outputs and make the case for the retention of human control over all decision making impacting societal, industry and worker needs, including decisions about the design and development of AI systems. This paper won a 2021 Parkman medal, best paper published by the Institution in the previous year on the practical aspects of the control or management, including project management of the design and/or construction of a specific scheme. Read the paper for [More]
It is common to see headlines about Artificial Intelligence (AI) – good things, like AI that can detect cancer better than doctors, or bad things, like a racist or sexist algorithm. In this talk, Rumman Chowdhury asserts that when considering moral dilemmas around AI, it is important not to detach humans from the machines they program and control. In doing this, she further breaks down how we can all individually shape AI for the better. Dr. Chowdhury’s passion lies at the intersection of artificial intelligence and humanity. She is an internationally recognized speaker on Artificial Intelligence, the future society, and ethics. She is the Global Lead for Responsible AI at Accenture Applied Intelligence, where she makes practical and ethical AI solutions for her clients. She is a data scientist and social scientist, holding two undergraduate degrees from MIT, a master’s from Columbia University, and a PhD from the University of California, San Diego. She has been recognized as one of Silicon Valley’s 40 under 40, one of the BBC’s 100 women, and is a fellow at the Royal Society of the Arts. This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at https://www.ted.com/tedx
Would a robot kill a human being? Many people ask themselves this question. Sophia is one of the most famous AI worldwide. We have her as a guest and ask her the trolley problem. How will the AI answer this moral question? How would you solve the trolley problem? Write it in the comments… ▸ subscribe to our channel: https://goo.gl/UupwgM #SophiaTheRobot #TrolleyProblem #Robot
Episode 44. Rebecca is a PhD candidate in Machine Ethics, and consultant in Ethical AI at Oxford Brookes University, Institute for Ethical Artificial Intelligence. Her PhD research is entitled ‘Autonomous Moral Artificial Intelligence’, and as a consultant she specialises in looking at developing practical approaches to embedding ethics in AI Products. Her background is primarily philosophy. She completed her BA, then MA in philosophy at The University of Nottingham in 2010, before working in analytics for several different industries. As an undergraduate she had a keen interest in logic, metametaphysics, and the topic of consciousness, spurring her to come back into academia in 2017 to undertake a further qualification in psychology at Sheffield Hallam University, before embarking on her PhD. She hopes she can combine her diverse interests to solving the challenge of creating moral machines. In her spare time she can be found playing computer games, running, or trying to explore the world.
http://www.biohackersummit.com Kaj Sotala is a researcher and author interested in both the risks of advanced artificial intelligence, and in developing communities that support the growth and flourishing of all their members. He has worked for the Machine Intelligence Research Institute and published several papers on the societal consequences of artificial general intelligence. Kaj’s grant proposal on developing AI systems capable of understanding human morality was chosen to receive funding from Elon Musk’s $10M donation for advancing long-term AI safety, one of the few proposals to be chosen from among 300 candidates. He is a long-time board member of the Finnish Transhumanist Association, and is currently involved in building a new self-development and peer-oriented life coaching community founded on scientifically vetted principles.
Jean-Francois Bonnefon, Toulouse School of Economics, held a keynote, “The Moral Machine Experiment”, at IJCAI-ECAI 2018, the 27th International Joint Conference on Artificial Intelligence and the 23rd European Conference on Artificial Intelligence, the premier international gathering of researchers in AI.
The success of any human-crewed interstellar mission depends on the existence of effective human-machine relationships. We anticipate that machines during such a mission won’t simply play the part of a supporting, background role, like an autopilot. Instead, navigating the demands of such a mission means that machines need to be equal ethical partners with humans, making decisions under conditions of irreducible uncertainty, in scenarios with potentially grave consequences. The objective of our work is to identify the salient factors that would either encourage or discourage effective partnerships between humans and machines in mission-critical scenarios. Our hypothesis is that there needs to be ethical congruence between human and machine: specifically, machines must not only understand the concept of moral responsibility; they must be able to convey to humans that they will make decisions accordingly. Recorded November 11, 2019
ACHLR ‘The Ethics of Artificial Intelligence: Moral Machines’ Public Lecture Learn more: https://www.qut.edu.au/law/research
Artificial Intelligence and Experience Series (AIEX): “Do People Perceive Machines as Moral Agents?” Bert F. Malle Department of Cognitive, Linguistic, and Psychological Sciences Brown University October 25, 2018
Denise Howell, J. Michael Keyes and Amanda Levendowski discuss the Moral Machine, an MIT Media Lab platform “for gathering a human perspective on moral decisions made by machine intelligence.” For the full episode, visit https://twit.tv/twil/356
Prof. Edmond Awad (Institute for Data Science and Artificial Intelligence at the University of Exeter) Abstract:  I describe the Moral Machine, an internet-based serious game exploring the many-dimensional ethical dilemmas faced by autonomous vehicles. The game enabled us to gather 40 million decisions from 3 million people in 200 countries/territories. I report the various preferences estimated from this data, and document interpersonal differences in the strength of these preferences. I also report cross-cultural ethical variation and uncover major clusters of countries exhibiting substantial differences along key moral preferences. These differences correlate with modern institutions, but also with deep cultural traits. I discuss how these three layers of preferences can help progress toward global, harmonious, and socially acceptable principles for machine ethics. Finally, I describe other follow up work that build on this project. Bio: Edmond Awad is a Lecturer (Assistant Professor) in the Department of Economics and the Institute for Data Science and Artificial Intelligence at the University of Exeter. He is also an Associate Research Scientist at the Max Planck Institute for Human Development, and is a Founding Editorial Board member of the AI and Ethics Journal, published by Springer. Before joining the University of Exeter, Edmond was a Postdoctoral Associate at MIT Media Lab (2017-2019). In 2016, Edmond led the design and development of Moral Machine,  a website that gathers human decisions on moral dilemmas faced by driverless cars. The website has been visited by over 4 million users, who contributed their judgements on 70 million dilemmas. Another [More]
Speaker: Toby Walsh The AI Revolution will transform our political, social and economic systems. It will impact not just the workplace, but many other areas of our society like politics and education. There are many ethical challenges ahead, ensuring that machines are fair, transparent, trustworthy, protective of our privacy and respect many other fundamental rights. Education is likely to be one of the main tools available to prepare for this future. Toby Walsh, Scientia Professor of Artificial Intelligence at Data61, University of New South Wales will argue that a successful society will be one that embraces the opportunity that these technologies promise, but at the same time prepares and helps its citizens through this time of immense change. Join him in this session, aiming to stimulate debate and discussion about AI, education and 21st century skill needs. www.oeb.global
A self-driving car has a split second to decide whether to turn into oncoming traffic or hit a child who has lost control of her bicycle. An autonomous drone needs to decide whether to risk the lives of busload of civilians or lose a long-sought terrorist. How does a machine make an ethical decision? Can it “learn” to choose in situations that would strain human decision making? Can morality be programmed? We will tackle these questions and more as the leading AI experts, roboticists, neuroscientists, and legal experts debate the ethics and morality of thinking machines. This program is part of the Big Ideas Series, made possible with support from the John Templeton Foundation. Subscribe to our YouTube Channel for all the latest from WSF. Visit our Website: http://www.worldsciencefestival.com/ Like us on Facebook: https://www.facebook.com/worldsciencefestival Follow us on twitter: https://twitter.com/WorldSciFest Original Program Date: June 4, 2016 MODERATOR: Bill Blakemore PARTICIPANTS: Fernando Diaz, Colonel Linell Letendre, Gary Marcus, Matthias Scheutz, Wendell Wallach Can Life and Death Decisions Be Coded? 00:00 Siri… What is the meaning of life? 1:49 Participant introductions 4:01 Asimov’s Three Laws of Robotics 6:22 In 1966 ELIZA was one of the first artificial intelligence systems. 10:20 What is ALPHAGO? 15:43 TAY Tweets the first AI twitter bot. 19:25 Can you test learning Systems? 26:31 Robots and automatic reasoning demonstration 30:31 How do driverless cars work? 39:32 What is the trolley problem? 49:00 What is autonomy in military terms? 56:40 Are landmines the first automated weapon? 1:10:30 Defining how artificial [More]
The William G. McGowan Charitable Fund is a philanthropic family foundation established in 1993 to perpetuate William McGowan’s tradition of compassionate philanthropy and ethical leadership. Today, the Fund preserves the legacy of William McGowan while embodying his tireless spirit and determined optimism. To this end, the Fund promotes, nurtures, and supports initiatives in three program areas: Education, Human Services, and Healthcare & Medical Research. Through the McGowan Fellows Program, the Fund supports and inspires emerging business leaders. In partnership with the nation’s leading graduate business programs, the Fund aims to imbue these future leaders with a framework for ethical decision-making and establish an ongoing dialogue on the importance of ethical practices.
Chair: Ned Block Links to panelists talks: S. Matthew Liao (NYU, Bioethics) https://www.youtube.com/watch?v=qPIqZ1rs-j8 Eric Schwitzgebel (UC Riverside, Philosophy) and Mara Garza (UC Riverside, Philosophy) https://www.youtube.com/watch?v=54-FI4qpwa8 John Basl (Northeastern, Philosophy) and Ronald Sandler (Northeastern, Philosophy) https://www.youtube.com/watch?v=m4OUitBEoiw
Machine ethics is an emerging discipline that enables ethical problems to be refined into something computational, that machines and humans can both understand rationally. New technologies can make ethical decisions calculable and transactional for the first time. Furthermore, Artificial Moral Advisors can help inform human beings of the potential trade-offs and repercussions of their decisions, and help people live more. Nell Watson believes these new capabilities self-reinforce each other, and have the potential of reshaping the moral fabric of our society within a generation. – This talk was given at a TEDx event using the TED conference format but independently organized by a local community. Learn more at https://www.ted.com/tedx
How do different cultures value human life? To find out, researchers created a viral online experiment to gather data from millions of participants across the world. Some values generalised across cultures, but others came as a surprise. Find out more in this Nature Video. Take part in the experiment here: http://moralmachine.mit.edu Find the original research paper here: https://www.nature.com/articles/s41586-018-0637-6 *CORRECTION* The spelling of Massachusetts at 00:38 is incorrect. We regret the error.
SixArticles: Issue #8 AI and FutureOfWork
By pairing the power of AI systems and human wisdom, scientists at Duke University hope to offer a tool for strengthening our moral capacities. Learn more at http://diverseintelligences.com/.
VR Robots like a Human. really??? Designing a Moral Machine
WATCH FULL EPISODE: https://youtu.be/NYNN87txLWQ .@SamHarrisOrg on how @WestworldHBO crosses uncanny valley of robotics & raises moral issues & questions about humanity-w/@jason-THX @wistia Today’s guest is Sam Harris, philosopher, neuroscientist and best-selling author of books including “Waking Up,” “The End of Faith,” “Letter to a Christian Nation,” and “The Moral Landscape.” Jason and Sam explore a wide range of topics, including the ethics of robots, the value of meditation, Trump’s lies, and his most recent obsession AI, which stemmed from an initial conversation with Elon Musk. Sam argues that the threat of uncontrolled AI is one of the most pressing issues of our time and poses the question: Can we build AI without losing control over it? The two then discuss why meditation is so important for entrepreneurs and business people. Sam has built his brand and fan base around radical honesty and authenticity, so the conversation naturally segues to Trump and his lies. This is only the first of two parts, so stay tuned for much more. For full show notes, subscribe to http://thisweekinstartups.com/about/#allsubscribe