<--- Back to Details
First PageDocument Content
Future / Artificial intelligence / Futurology / Transhumanists / Computational neuroscience / Friendly artificial intelligence / Eliezer Yudkowsky / Strong AI / Agent-based model / Singularitarianism / Science / Time
Date: 2015-03-20 21:19:24
Future
Artificial intelligence
Futurology
Transhumanists
Computational neuroscience
Friendly artificial intelligence
Eliezer Yudkowsky
Strong AI
Agent-based model
Singularitarianism
Science
Time

Aligning Superintelligence with Human Interests: An Annotated Bibliography Nate Soares Machine Intelligence Research Institute

Add to Reading List

Source URL: intelligence.org

Download Document from Source Website

File Size: 140,21 KB

Share Document on Facebook

Similar Documents

Futurepol Conference April

Futurepol Conference April

DocID: 1rtd4 - View Document

BRIEFER No. 02 | July 20, 2011 The Inadequate U.S. Response to a Major Security Threat: Climate Change Francesco Femia, Christine Parthemore and Caitlin E. Werrell

BRIEFER No. 02 | July 20, 2011 The Inadequate U.S. Response to a Major Security Threat: Climate Change Francesco Femia, Christine Parthemore and Caitlin E. Werrell

DocID: 1rt7u - View Document

|  Natasha Vita-More, PhD CV Design / Media Arts / Theory

| Natasha Vita-More, PhD CV Design / Media Arts / Theory

DocID: 1rqyK - View Document

“Lowballing” in Analysts’ Forecasts  Gilles Hilary  HKUST Charles Hsu

“Lowballing” in Analysts’ Forecasts Gilles Hilary HKUST Charles Hsu

DocID: 1rmgi - View Document

Harnessing the Wisdom of Crowds* Zhi Da†, and Xing Huang‡ This Draft: July 2016 Abstract We examine the negative information externality associated with herding on a crowd-based

Harnessing the Wisdom of Crowds* Zhi Da†, and Xing Huang‡ This Draft: July 2016 Abstract We examine the negative information externality associated with herding on a crowd-based

DocID: 1rkxV - View Document