{"version": "1.0", "type": "rich", "title": "Listen the thing about the argument that AI is dangerous because we can\u2019t control it is that it\u2019s not necessarily wrong it\u2019s...", "author_name": "kontextmaschine", "author_url": "https://kontextmaschine.com", "provider_name": "kontextmaschine", "provider_url": "https://kontextmaschine.com", "url": "https://kontextmaschine.com/post/140184024413/", "html": "<p><a href=\"http://antinegationism.tumblr.com/post/140090479956/antinegationism-listen-the-thing-about-the\" class=\"tumblr_blog\" target=\"_blank\">antinegationism</a>:</p>\n\n<blockquote><p><a class=\"tumblr_blog\" href=\"http://antinegationism.tumblr.com/post/140090265511\" target=\"_blank\">antinegationism</a>:</p>\n<blockquote>\n<p>Listen the thing about the argument that AI is dangerous because we can\u2019t control it is that it\u2019s not necessarily <i>wrong </i>it\u2019s just that it kind of distracts from the more immediate concern of all of the ways AI could be dangerous because we <i>can </i>control it.</p>\n</blockquote>\n\n<p>Not that there\u2019s much purpose to asking these questions. I think collectively we as a society have already sort of decided that, despite all of the ways things could go horribly wrong, AIs are just way too cool and sci-fi to not make.</p></blockquote>"}