<--- Back to Details
First PageDocument Content
Artificial intelligence / Technology / Motivation / Metaphysics / Game theory / Futurology / Philosophy of artificial intelligence / Choice modelling / Utility / Expected utility hypothesis / Intelligent agent / Friendly artificial intelligence
Date: 2017-01-30 22:06:06
Artificial intelligence
Technology
Motivation
Metaphysics
Game theory
Futurology
Philosophy of artificial intelligence
Choice modelling
Utility
Expected utility hypothesis
Intelligent agent
Friendly artificial intelligence

The AI Alignment Problem: Why It’s Hard, and Where to Start Eliezer Yudkowsky Machine Intelligence Research Institute May 5, 2016

Add to Reading List

Source URL: intelligence.org

Download Document from Source Website

File Size: 278,17 KB

Share Document on Facebook

Similar Documents

Futurepol Conference April

Futurepol Conference April

DocID: 1rtd4 - View Document

BRIEFER No. 02 | July 20, 2011 The Inadequate U.S. Response to a Major Security Threat: Climate Change Francesco Femia, Christine Parthemore and Caitlin E. Werrell

BRIEFER No. 02 | July 20, 2011 The Inadequate U.S. Response to a Major Security Threat: Climate Change Francesco Femia, Christine Parthemore and Caitlin E. Werrell

DocID: 1rt7u - View Document

|  Natasha Vita-More, PhD CV Design / Media Arts / Theory

| Natasha Vita-More, PhD CV Design / Media Arts / Theory

DocID: 1rqyK - View Document

“Lowballing” in Analysts’ Forecasts  Gilles Hilary  HKUST Charles Hsu

“Lowballing” in Analysts’ Forecasts Gilles Hilary HKUST Charles Hsu

DocID: 1rmgi - View Document

Harnessing the Wisdom of Crowds* Zhi Da†, and Xing Huang‡ This Draft: July 2016 Abstract We examine the negative information externality associated with herding on a crowd-based

Harnessing the Wisdom of Crowds* Zhi Da†, and Xing Huang‡ This Draft: July 2016 Abstract We examine the negative information externality associated with herding on a crowd-based

DocID: 1rkxV - View Document