{"id":1044,"date":"2023-07-03T10:09:45","date_gmt":"2023-07-03T10:09:45","guid":{"rendered":"https:\/\/hello.inherentknowledge.org\/2024\/2023\/07\/03\/microsoft-pushes-for-government-regulation-of-ai-should-we-trust-it\/"},"modified":"2023-07-03T10:09:45","modified_gmt":"2023-07-03T10:09:45","slug":"microsoft-pushes-for-government-regulation-of-ai-should-we-trust-it","status":"publish","type":"post","link":"https:\/\/hello.inherentknowledge.org\/2024\/2023\/07\/03\/microsoft-pushes-for-government-regulation-of-ai-should-we-trust-it\/","title":{"rendered":"Microsoft pushes for government regulation of AI. Should we trust it?"},"content":{"rendered":"<p>By now, virtually everyone agrees that powerful <a href=\"https:\/\/www.computerworld.com\/article\/3698191\/governments-worldwide-grapple-with-regulation-to-rein-in-ai-dangers.html\">generative AI needs to be regulated<\/a>. In its various forms, it presents a <a href=\"https:\/\/www.computerworld.com\/article\/3695073\/qa-at-mit-tom-siebel-labels-the-consequences-of-ai-as-terrifying.html\">variety of potential dangers<\/a>: helping authoritarian regimes, thanks to its ability to create misinformation; allowing Big Tech firms to establish monopolies; eliminating millions of jobs; taking over vital infrastructure; and \u2014 in the worst case \u2014 becoming an existential threat to humankind.<\/p>\n<p class=\"jumpTag\"><a href=\"https:\/\/www.computerworld.com\/article\/3700969\/microsoft-pushes-for-government-regulation-of-ai-should-we-trust-it.html#jump\">To read this article in full, please click here<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>By now, virtually everyone agrees that powerful generative AI needs to be regulated. In its various forms, it presents a variety of potential dangers: helping authoritarian regimes, thanks to its ability to create misinformation; allowing Big Tech firms to establish monopolies; eliminating millions of jobs; taking over vital infrastructure; and \u2014 in the worst case [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-1044","post","type-post","status-publish","format-standard","hentry","category-uncategorized"],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts\/1044","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/comments?post=1044"}],"version-history":[{"count":0,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts\/1044\/revisions"}],"wp:attachment":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/media?parent=1044"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/categories?post=1044"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/tags?post=1044"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}