{"id":1170,"date":"2024-03-05T22:16:45","date_gmt":"2024-03-05T22:16:45","guid":{"rendered":"https:\/\/hello.inherentknowledge.org\/2024\/2024\/03\/05\/researchers-legal-experts-want-ai-firms-to-open-up-for-safety-checks\/"},"modified":"2024-03-05T22:16:45","modified_gmt":"2024-03-05T22:16:45","slug":"researchers-legal-experts-want-ai-firms-to-open-up-for-safety-checks","status":"publish","type":"post","link":"https:\/\/hello.inherentknowledge.org\/2024\/2024\/03\/05\/researchers-legal-experts-want-ai-firms-to-open-up-for-safety-checks\/","title":{"rendered":"Researchers, legal experts want AI firms to open up for safety checks"},"content":{"rendered":"<p>More than 150 leading artificial intelligence (AI) researchers, ethicists and others have signed an\u00a0<a href=\"https:\/\/sites.mit.edu\/ai-safe-harbor\/\" target=\"_blank\" rel=\"noopener\">open letter<\/a>\u00a0calling on generative AI (genAI) companies to submit to independent evaluations of their systems, the lack of which has led to concerns about basic protections.<\/p>\n<p>The letter,\u00a0drafted by researchers from MIT, Princeton, and Stanford University, called for legal and technical protections for\u00a0<a href=\"https:\/\/substack.com\/redirect\/34e4ef63-73a8-4559-8c91-968546804f9d?j=eyJ1IjoiMmN3eXlpIn0.-XionA57hnMoMt9ryxqfe913wQYanc0bgQWTzBRe3Ow\" target=\"_blank\" rel=\"noopener\">good-faith\u00a0research on genAI models<\/a>, which they said is hampering safety measures that could help protect the public.<\/p>\n<p class=\"jumpTag\"><a href=\"https:\/\/www.computerworld.com\/article\/3714180\/researchers-legal-experts-want-ai-firms-to-open-up-for-safety-checks.html#jump\">To read this article in full, please click here<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>More than 150 leading artificial intelligence (AI) researchers, ethicists and others have signed an\u00a0open letter\u00a0calling on generative AI (genAI) companies to submit to independent evaluations of their systems, the lack of which has led to concerns about basic protections. The letter,\u00a0drafted by researchers from MIT, Princeton, and Stanford University, called for legal and technical protections [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-1170","post","type-post","status-publish","format-standard","hentry","category-uncategorized"],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts\/1170","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/comments?post=1170"}],"version-history":[{"count":0,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/posts\/1170\/revisions"}],"wp:attachment":[{"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/media?parent=1170"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/categories?post=1170"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hello.inherentknowledge.org\/2024\/wp-json\/wp\/v2\/tags?post=1170"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}