<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Safety & Governance Newsletter]]></title><description><![CDATA[A Newsletter Highlighting Emerging Topics and News on AI Safety & Governance From Around the World, Curated by Victor Akinwande.]]></description><link>https://www.safeai.news</link><image><url>https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png</url><title>AI Safety &amp; Governance Newsletter</title><link>https://www.safeai.news</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 10:59:43 GMT</lastBuildDate><atom:link href="https://www.safeai.news/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Victor Akinwande]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[safeintelligence@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[safeintelligence@substack.com]]></itunes:email><itunes:name><![CDATA[Victor Akinwande]]></itunes:name></itunes:owner><itunes:author><![CDATA[Victor Akinwande]]></itunes:author><googleplay:owner><![CDATA[safeintelligence@substack.com]]></googleplay:owner><googleplay:email><![CDATA[safeintelligence@substack.com]]></googleplay:email><googleplay:author><![CDATA[Victor Akinwande]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Issue #6 of the AI Safety & Governance Newsletter]]></title><description><![CDATA[U.S. Senate AI roadmap, South Africa's AI hubs, U.S.-China AI collaboration, Colorado's AI law, Sony Music's AI restrictions, AI governance in Africa, implications of AI taking over basic life skills.]]></description><link>https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Mon, 27 May 2024 17:49:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thank you for reading AI Safety &amp; Governance Newsletter. This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p>Consumers in many African countries are not lagging behind in adopting new AI technologies. Charles Asiegbu and Chinasa T. Okolo for Brookings write that <strong>existing data protection regulations can be integrated into the AI governance frameworks that African countries seek to develop</strong> (<a href="https://www.brookings.edu/articles/how-ai-is-impacting-policy-processes-and-outcomes-in-africa/">link</a>).</p><p></p><p>In a new report titled &#8220;Driving U.S. Innovation in Artificial Intelligence: A Roadmap for Artificial Intelligence Policy in the United States Senate,&#8221; the Bipartisan Senate AI Working Group, led by U.S. Senator Chuck Schumer, recommends <strong>at least $32 billion per year for AI innovation.</strong> The report is a culmination of a series of AI &#8220;insight&#8221; forums involving AI experts from industry and academia (<a href="https://www.young.senate.gov/wp-content/uploads/Roadmap_Electronic1.32pm.pdf">link</a>).</p><p>In addition to funding to enable cross-agency R&amp;D, semiconductor R&amp;D, AI Grand Challenges, and the National AI Research Resource, the report emphasizes the need for workforce up-skilling and improved immigration for STEM talent while <strong>ensuring AI systems comply with existing laws, especially in consumer protection and civil rights.</strong></p><p>Critics argue that the report prioritizes industry and defense interests over public benefit, civil rights, and corporate accountability, lacking concrete measures to address AI's potential harms, such as bias and discrimination. See <a href="https://www.techpolicy.press/us-senate-ai-report-meets-mostly-disappointment-condemnation/">Tech Policy article</a>.</p><p>While the report advocates for a risk-based approach to AI, comprehensive data privacy laws, and measures to manage national security threats from AI, including export controls and safeguards against bioweapons, it is not legislation and thus not a concrete regulatory intervention. Personally, I found the written statements from the expert briefings to be the most informative.</p><p></p><p>Kate Bartlett writes for VOA News about <strong>South Africa&#8217;s investment in AI expertise through the establishment of sector-focused AI hubs</strong>. More broadly, the article highlights the increasing interest of the U.S. and China in investing in AI on the African continent (<a href="https://www.voanews.com/a/ai-becomes-latest-frontier-in-china-us-race-for-africa/7605069.html?trk=feed_main-feed-card_feed-article-content">link</a>). </p><p><strong>An emerging theme of U.S.-China collaboration on AI is capacity development for developing countries</strong>. Xue Lan, director of Tsinghua University's Institute for AI International Governance, <a href="https://www.scmp.com/news/china/diplomacy/article/3263546/us-china-can-work-together-ai-despite-barriers-expert-tells-forum">recently said</a>, &#8220;Another potential area that the U.S. and China can work together on AI is how to increase, to improve the capacity of developing countries in accessing and using AI technologies.&#8221;</p><p></p><p>The <strong>UK AI Safety Institute released a new tool for evaluating large language models</strong>. I&#8217;ve written about the AI Safety Institute in previous newsletters, and the newly published tool, Inspect, is designed to enhance safety evaluations by providing a standardized, open-source platform for assessing AI models' capabilities and security (<a href="https://www.gov.uk/government/news/ai-safety-institute-releases-new-ai-safety-evaluations-platform">link</a>).</p><p></p><p>The Council of Europe has been proactively addressing the challenges posed by advancements in AI, particularly concerning human rights, democracy, and the rule of law. This effort culminated in forming the Committee on Artificial Intelligence (CAI) to develop a Framework Convention to ensure AI activities comply with these fundamental principles.</p><p>The resulting Framework Convention is heralded as the &#8220;<strong>first-ever international legally binding treaty&#8221; aimed at ensuring the respect for human rights, the rule of law, and democracy standards in the use of AI systems</strong>. The treaty was recently adopted in Strasbourg and aligns with existing international and domestic legal obligations in Europe, aiming to mitigate risks such as discrimination, undermining democratic processes, and violations of human dignity. See the <a href="https://www.coe.int/en/web/portal/-/council-of-europe-adopts-first-international-treaty-on-artificial-intelligence">press release</a> for more details.</p><p></p><p>Helen Toner and Tasha McCauley, <strong>former board members of OpenAI, in The Economist, argue that private companies cannot be expected to align their profit motives with the public good when developing technologies like AI</strong>, necessitating effective governmental regulation (<a href="https://www.economist.com/by-invitation/2024/05/26/ai-firms-mustnt-govern-themselves-say-ex-members-of-openais-board">link</a>). They explain how OpenAI's experiment in self-governance, despite its innovative structure, ultimately failed to uphold its public-interest mission, highlighting the need for external oversight.</p><p></p><p>Benjamin Cedric Larsen and Sabrina K&#252;spert for Brookings <strong>compare and contrast the EU and U.S. approaches to regulating AI</strong> (<a href="https://www.brookings.edu/articles/regulating-general-purpose-ai-areas-of-convergence-and-divergence-across-the-eu-and-the-us/">link</a>). They argue that the EU's approach has been more comprehensive and impactful. For example, the EU AI Act imposes legally binding regulations on providers of general-purpose AI within the EU, whereas the U.S. executive order primarily sets guidelines for federal agencies and imposes limited regulations on private entities. They claim that the voluntary nature of such guidelines aligns with the U.S.'s preceding preference for industry self-governance. However, as international efforts come into focus, <strong>cooperation between the EU and the U.S., together with other G7 countries, could serve as a model for aligning international efforts.</strong></p><p></p><p><strong>A newly enacted law in Colorado mandates that developers and deployers of high-risk AI systems use reasonable care to avoid algorithmic discrimination</strong>, with a rebuttable presumption of compliance if they meet specific requirements. Developers must disclose information about their AI systems, provide documentation for impact assessments, publicly summarize risks, and inform the attorney general and deployers of any risks within 90 days of discovery (<a href="https://fpf.org/blog/colorado-enacts-first-comprehensive-u-s-law-governing-artificial-intelligence-systems/">link</a>, <a href="https://www.techpolicy.press/colorado-makes-history-with-the-nations-first-comprehensive-ai-act">link</a>).</p><p></p><p>Sony Music in a recent statement prohibited unauthorized text or data mining, web scraping, or similar uses of their content for AI training or commercialization (<a href="https://www.sonymusic.com/sonymusic/declaration-of-ai-training-opt-out/">link</a>).</p><p></p><p>At the heart of the current wave of innovation in generative AI is massive amounts of data collected from the web. In a recent article for the NY Times, Cade Metz discusses how companies like OpenAI, Google, and Meta have scoured every corner of the internet for reputable English-language text to train their AI models. As these companies face a growing data shortage, they resort to various methods to gather more data, including transcribing YouTube videos and exploring the use of synthetic data. Metz highlights <strong>the legal and ethical dilemmas many AI companies encounter as they push the boundaries of data usage and copyright law to stay ahead in the AI race</strong> (<a href="https://www.nytimes.com/2024/04/06/technology/tech-giants-harvest-data-artificial-intelligence.html">link</a>).</p><p></p><p>Similarly, <strong>obtaining high-quality training data for AI models in non-English languages is incredibly challenging.</strong> A recent discovery following the release of a new model from OpenAI revealed that the model&#8217;s Chinese token library was polluted with phrases from spam websites. Zeyi Yang <a href="https://www.technologyreview.com/2024/05/22/1092763/openais-gpt4o-chinese-ai-data/">discussed this</a> in more detail in MIT Tech Review.</p><p></p><p><strong>U.S. Secretary of Commerce Gina Raimondo recently announced plans to launch a global scientific network for AI safety, as well as establish the U.S. AI Safety Institute (AISI)</strong>. According to the vision statement, </p><blockquote><p>AISI &#8220;will enable more rigorous assessment of AI risk; more effective internal and external safeguards for AI models, systems, and agents; greater public confidence; and ultimately wider and more responsible development and adoption of AI.&#8221;</p></blockquote><p></p><p>Technology companies do not seem to be shying away from military applications of AI. Following the inaugural &#8220;AI Expo for National Competitiveness,&#8221; hosted by the think tank created by former Google CEO Eric Schmidt, Caroline Haskins writes for The Guardian about <strong>how tech companies are discussing the applications of AI in military settings</strong> (<a href="https://www.theguardian.com/technology/article/2024/may/17/ai-weapons-palantir-war-technology">link</a>).</p><p></p><p><strong>The big AI risk,</strong> according to Tyler Austin Harper, isn't the sci-fi scenario of robots becoming smarter than us and taking over. It is something closer to home and happening right now: <strong>the way we're letting AI take over basic life skills</strong>. Harper points out that while AI tools like dating apps, meal-kit services, and personalized recommendations are sold to us as ways to make life easier, they actually make us more dependent on algorithms for things we used to do ourselves&#8212;like socializing, thinking, and creating. This, he says, could lead to a future where we struggle to have normal human interactions without relying on AI, which would fundamentally change what it means to be human. <a href="https://archive.ph/5sz1m#selection-1143.0-1143.19">Link to article on The Atlantic.</a></p><p></p><p>In an insightful interview, Linda Moore, CEO of TechNet&#8212;a tech lobbying organization&#8212;highlights the <strong>importance of educating both the public and policymakers on AI</strong> while ensuring that new regulations do not stifle innovation or drive businesses away (<a href="https://www.semafor.com/article/05/10/2024/as-states-take-on-ai-techs-top-lobbyist-fights-back">link</a>).</p><p></p><p>In a powerful personal account, former content moderator for Facebook and Instagram in Kenya, Jatani Hussein, shares <strong>the traumatic experiences of reviewing thousands of graphic and violent posts daily.</strong> Employed through outsourcing companies, Jatani and his fellow moderators faced severe mental health challenges while scrubbing harmful content off the internet. Last week, Kenyan President William Ruto visited the White House. Jatani urges consideration of the risks faced by content moderators&#8212;a role that continues to be in demand as more companies try to ensure their AI models are used safely.</p><p></p><p>Thanks for reading!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.safeai.news/p/issue-6-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Issue #5 of the AI Safety & Governance Newsletter]]></title><description><![CDATA[Guidance reports by NIST on generative AI, enforcing voluntary compliance to pre-release safety testing of AI models, UN AI governance principles, AI's impact on perceptions of linguistic authenticity]]></description><link>https://www.safeai.news/p/issue-5-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-5-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Mon, 06 May 2024 17:41:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the 5th issue of the Artificial Intelligence (AI) Safety &amp; Governance Newsletter. Thank you for subscribing! Let&#8217;s dive right in.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.safeai.news/subscribe?"><span>Subscribe now</span></a></p><p></p><h3>Key Reads</h3><p>When a new technology emerges, there are opportunities and risks. Generative AI is no different. The National Institute of Standards and Technology (NIST) published a report describing a set of 12 distinct risks for the widespread use of generative AI technology, including the spread of misinformation, data privacy violations, and the potential for AI to generate harmful content or biased outputs (<a href="https://airc.nist.gov/docs/NIST.AI.600-1.GenAI-Profile.ipd.pdf">link</a>).</p><p>The report is one of <strong>four &#8220;guidance&#8221; reports published by NIST</strong>. Another report focuses on guidelines for documenting the origin and history of content, using methods like digital watermarking and recording metadata (<a href="https://airc.nist.gov/docs/NIST.AI.100-4.SyntheticContent.ipd.pdf">link</a>).  The third advocates for collaborative efforts with international partners to develop and implement AI-related consensus standards (<a href="https://airc.nist.gov/docs/NIST.AI.100-5.Global-Plan.ipd.pdf">link</a>), and the final report is a specialized extension of the Secure Software Development Framework (SSDF) (<a href="https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-218A.ipd.pdf">link</a>) tailored specifically for generative AI and foundation model development. This extension provides a set of augmented recommendations and considerations that align with established SSDF practices (<a href="https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-218A.ipd.pdf">link</a>).</p><p>See Tech Policy for a summary of each report (<a href="https://www.techpolicy.press/nist-unveils-ai-draft-guidance-reports/">link</a>).</p><p></p><p>A <strong>new bill</strong> introduced by U.S. Senators Mark R. Warner and Thom Tillis, co-chairs of the Senate Cybersecurity Caucus, aims to <strong>improve the tracking and processing of security and safety incidents and risks associated with AI</strong>, including enhancing information sharing between the federal government and private companies (<a href="https://www.theverge.com/2024/5/1/24146566/ai-security-bill-warner-tillis-senate-redteam-safety">link</a>, <a href="https://www.warner.senate.gov/public/index.cfm/pressreleases?id=86DBB0A3-F402-4594-AD76-AF38B38CA068">link</a>).</p><p></p><p>A new Florida <strong>law mandates that political advertisements using generative AI to</strong> depict unreal events prominently <strong>display a disclaimer</strong>, with violations treated as first-degree misdemeanors, effective July 1, 2024 (<a href="https://www.dataguidance.com/news/florida-governor-signs-bill-ai-use-political">link</a>).</p><p></p><p><strong>Jeremy Howard, CEO of Answer.AI, critiques California&#8217;s SB-1047</strong> (<a href="https://leginfo.legislature.ca.gov/faces/billNavClient.xhtml?bill_id=202320240SB1047">link</a>), arguing it could stifle AI innovation by imposing harsh regulations on open-source development and small businesses. He suggests focusing on regulating AI applications in high-risk areas rather than development (<a href="https://www.answer.ai/posts/2024-04-29-sb1047.html">link</a>).</p><p></p><p>A new <strong>Colorado law aims to protect the privacy of neural data</strong>, increasingly collected by consumer neuro-technologies (<a href="https://www.nytimes.com/2024/04/17/science/colorado-brain-data-privacy.html">link</a>).</p><p></p><p><strong>Tina Stowell,</strong> a member of the House of Lords in the UK Parliament, expressed approval for the <strong>UK government&#8217;s commitment to AI safety</strong> and support for startups but <strong>criticized its handling of copyright issues and lack of clear policies on AI standards</strong>. She emphasized the need for more decisive actions to protect the integrity of British businesses and innovation in a recent letter. See longer discussion in The Register (<a href="https://www.theregister.com/2024/05/03/uk_lords_llm_copyright/">link</a>). </p><p></p><p>The non-profit privacy group noyb, led by activist Max Schrems, has filed a complaint with the Austrian data protection authority <strong>accusing OpenAI&#8217;s ChatGPT of violating the EU's General Data Protection Regulation (GDPR).</strong> The complaint centers on ChatGPT&#8217;s generation of inaccurate personal data, specifically incorrect birthdays, and its failure to correct or delete such data upon request (<a href="https://www.politico.eu/article/chatgpts-hallucinations-get-eu-privacy-complaint/">link</a>). This issue is part of broader concerns in Europe, as evidenced by Italy&#8217;s temporary ban on ChatGPT and the formation of a task force by the European Data Protection Board to address these challenges.</p><p></p><p>At the AI safety summit at Bletchley Park, technology companies initially agreed to allow the UK&#8217;s AI Safety Institute (AISI) to conduct pre-release safety testing on new AI models. However, six months later, this agreement appears to have largely failed to materialize, according to a new article by Vincent Manancourt, Gian Volpicelli, and Mohar Chatterjee for Politico (<a href="https://www.politico.eu/article/rishi-sunak-ai-testing-tech-ai-safety-institute/">link</a>). Despite the UK government&#8217;s claims of commencing such testing, most new AI models, like Meta&#8217;s Llama-3, have not been subjected to <strong>pre-release checks by the AISI</strong>. This situation highlights the challenges of <strong>enforcing voluntary compliance in the absence of stringent legal requirements</strong> and the reluctance of AI firms to share sensitive technologies due to fears of competition and jurisdictional issues.</p><p></p><p>Nonprofits backed by technology billionaires, such as the Center for AI Policy and Center for AI Safety, have begun <strong>lobbying in Washington to address the existential risks of AI,</strong> writes Brendan Bordelon for Politico (<a href="https://www.politico.com/news/2024/02/23/ai-safety-washington-lobbying-00142783">link</a>). According to Brendan, these nonprofits advocate for regulations that could impose severe liabilities on AI developers and potentially halt high-risk AI projects. However, critics argue that <strong>these efforts might stifle smaller AI companies by raising entry barriers and focusing too much on catastrophic scenarios.</strong></p><p></p><p>The use of AI to create compelling propaganda is a key risk. A recent survey experiment conducted with US participants examined the <strong>persuasiveness of news articles produced by foreign propagandists compared to content generated by a large language model</strong> (GPT-3) (<a href="https://academic.oup.com/pnasnexus/article/3/2/pgae034/7610937?login=false">link</a>). The study found that both the original propaganda articles and those generated by GPT-3 were highly persuasive. While GPT-3-generated content was slightly less persuasive on average than the original propaganda, human involvement through editing prompts and curating output improved its persuasiveness.</p><p></p><p>During the recent 2024 China-Africa Internet Development and Cooperation Forum, <strong>collaboration between African countries and China on AI governance</strong> was one of the key themes. The call to action includes strengthening policy dialogue, promoting technology research and application, fostering industrial cooperation, enhancing talent exchanges and capacity building, and bolstering safeguards for cyber and data security (<a href="https://www.cac.gov.cn/2024-04/03/c_1713731793842754.htm">link</a>).</p><p></p><p>An article by Alex Hern for The Guardian discusses <strong>the emergence of distinct linguistic patterns, dubbed &#8220;AI-ese&#8221;, in responses generated by AI chatbots</strong> (<a href="https://www.theguardian.com/technology/2024/apr/16/techscape-ai-gadgest-humane-ai-pin-chatgpt">link</a>). Alex highlights the telltale signs of AI-generated content, such as exaggerated politeness and aversions to brevity, alongside idiosyncratic linguistic tendencies like the overuse of words like &#8220;delve&#8221; (a recent source of controversy when a famous venture capitalist, Paul Graham, alluded to ignoring a message that included the word &#8220;delve&#8221;).</p><p>Labelling for AI chatbots, including large language models is often outsourced to cheaper labor markets. This implies their linguistic patterns can mimic those of the labelers. For instance<strong>, the overuse of &#8220;delve&#8221; in ChatGPT&#8217;s responses may reflect the linguistic norms of Nigerian English</strong>, where the word is more prevalent. Alex&#8217;s article underscores how the blurred lines between AI-generated content and human speech pose <strong>implications for language dynamics and societal perceptions of linguistic authenticity</strong>. Shout out to my senior colleague Kush Varshney for sending this article my way.</p><p></p><p>A new analysis by the Brookings Institute shows a <strong>1500% increase in the potential value of US Department of Defense contracts related to AI</strong> between August 2022 and August 2023 (<a href="https://www.brookings.edu/articles/the-evolution-of-artificial-intelligence-ai-spending-by-the-u-s-government/">link</a>).</p><p></p><p>Belinda Cleeland, Maxime Stauffer, and Malou Estier provide a summary and review of the Global Digital Compact (GDC) Zero Draft. <strong>The Zero Draft of the GDC is a set of principles and proposed actions for global governance of emerging technologies</strong>, particularly AI, set to be adopted by member states of the United Nations (<a href="https://www.simoninstitute.ch/blog/post/response-to-the-zero-draft-of-the-global-digital-compact/">link</a>).</p><p></p><h3>Opportunities<br></h3><p>A new program by NIST called the GenAI evaluation program is seeking approaches from academia and industry on <strong>ways to discriminate between synthetic and human-generated content</strong> in text-to-text and text-to-image modalities. Registration is currently open (<a href="https://ai-challenges.nist.gov/genai">link</a>).</p><p>Can wisdom and philosophy be automated? AI Impacts announced an essay <strong>competition focusing on the automation of wisdom and philosophy</strong>. The aim is to explore the potential impact of advanced AI on decision-making processes and philosophical understanding. There is a total prize pool of $25,000 (<a href="https://blog.aiimpacts.org/p/essay-competition-on-the-automation">link</a>).</p><p></p><p>Thank you for reading!</p><p>Victor.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/p/issue-5-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.safeai.news/p/issue-5-of-the-ai-safety-and-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Issue #4 of the AI Safety & Governance Newsletter]]></title><description><![CDATA[Acceptable use policies for foundation models, China's AI regulation stance, AI deployment and security best practices, AI safety benchmark]]></description><link>https://www.safeai.news/p/issue-4-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-4-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Sun, 21 Apr 2024 23:42:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the 4th issue of the Artificial Intelligence (AI) Safety &amp; Governance Newsletter. Thank you for subscribing!</p><p>Let's dive right into the key reads and news I've come across in the past two weeks.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>Key Reads</strong></h3><p>A recent article in MIT Tech Review by Zeyi Yang argues that <strong>China's approach to AI regulation is currently in a phase of lax regulation</strong>, due to the Chinese government's interest in the development of China&#8217;s AI industry (<a href="https://www.technologyreview.com/2024/04/09/1091004/china-tech-regulation-harsh-zhang/">link</a>). The article posits that Chinese AI regulations are looser than those in the US and Europe when it comes to preventing AI from causing harm, except in areas related to freedom of speech and content control. However, if there is a significant misuse of AI that &#8220;threatens social stability,&#8221; Angela Huyue Zhang, a law professor at Hong Kong University who is quoted, predicts that the Chinese regulatory pendulum will swing to the harsh side again, and it will be quick and unexpected.</p><p></p><p>Elections are being held in <a href="https://time.com/6550920/world-elections-2024/">at least 64 countries</a> globally this year, and the use of AI for spreading misinformation is top of mind. <strong>Rest of World has launched an AI elections tracker to track incidents of AI being used in elections globally</strong>&#8212;for campaigning, spreading misinformation, and creating memes (<a href="https://restofworld.org/2024/elections-ai-tracker/">link</a>).</p><p></p><p>MLCommons is a consortium of organizations in the technology industry and academia, built on a philosophy of open collaboration to make AI systems better for everyone. The <strong>MLCommons AI Safety working group has published a new benchmark to evaluate the safety of large language models (LLMs)</strong> (<a href="https://mlcommons.org/benchmarks/ai-safety/">link</a>, <a href="https://arxiv.org/html/2404.12241v1">link</a>). The benchmark is designed to measure when LLMs respond affirmatively when prompted according to seven hazard categories, including crimes, child sexual exploitation, indiscriminate weapons, suicide, self-harm, and hate. This is an important milestone that coincides with the release of Llama-3, a <a href="https://ai.meta.com/blog/meta-llama-3/">new LLM by Meta</a>, which is touted as the most capable openly available LLM to date.</p><p></p><p>In May, <strong>the South Korean capital, Seoul, will host the next AI Safety Summit</strong>, following the inaugural event that took place in the UK last November. The country recently announced a $6.8 billion commitment to boost AI chip development and production (<a href="https://aibusiness.com/verticals/south-korea-commits-6-8b-to-boost-ai-chip-development-production">link</a>).</p><p></p><p>Acceptable use policies for a digital product or service are often used to articulate how the service provider may restrict what users can and cannot do with their products and services. A recent article by <strong>Kevin Klyman catalogs the acceptable use policies for large-scale AI models (foundation models) by 30 technology companies.</strong> The article highlights the differences in how developers articulate and enforce acceptable use policies, which affects what content is prohibited and how the use of foundation models may be restricted in certain industries (<a href="https://crfm.stanford.edu/2024/04/08/aups.html">link</a>).</p><p></p><p>A recent panel discussion on the newly published <a href="https://aiindex.stanford.edu/wp-content/uploads/2024/04/HAI_AI-Index-Report-2024.pdf">500-page AI Index report </a>highlights <strong>emerging topics around AI governance and policy.</strong> The AI Index is an independent initiative at the Stanford Institute for Human-Centered AI that tracks, collates, distills, and visualizes data related to AI. The discussion highlighted how the industry continues to dominate frontier AI research (over academia), the shifting geopolitical dynamics of AI governance towards constraining or controlling AI, AI's environmental impact, and the role of open-source development in countries seeking to maintain leadership in AI development (<a href="https://www.cfr.org/event/next-chapter-artificial-intelligence">link</a>).</p><p></p><p><strong>The Trade Union Congress (TUC), a federation of trade unions in the UK, recently published a bill to encourage the UK government to take a firmer stance on regulating AI.</strong> Key provisions include mandatory transparency for AI systems, including the ability for them to explain their decision-making processes, along with detailed risk assessments before and after deployment (<a href="https://tuc.org.uk/research-analysis/reports/ai-bill-project">link</a>).</p><p>The bill also seeks to enforce strong consultation rights for workers and unions, establish a register of AI systems, and guarantee the right to human review. Additionally, it addresses data rights, bans harmful emotion recognition technology, and includes anti-discrimination measures by reversing the burden of proof for AI-related discrimination claims.</p><p></p><p>A newly introduced bill by Rep. Adam Schiff of California would require AI companies to disclose which copyrighted works were used to train their models (<a href="https://www.theguardian.com/technology/2024/apr/09/artificial-intelligence-bill-copyright-art">link</a>).</p><p></p><p>A new report authored by a collaboration of prominent national security and cybersecurity organizations from several countries provides &#8220;<strong>Best Practices for Deploying Secure and Resilient AI Systems</strong>.&#8221; Organizations involved in the report include the US Federal Bureau of Investigation (FBI), the Australian Signals Directorate&#8217;s Australian Cyber Security Centre (ACSC), the United Kingdom&#8217;s National Cyber Security Centre (NCSC-UK), and their counterparts in Canada and New Zealand. The guidelines include securing the deployment infrastructure for AI systems, monitoring models for unauthorized changes or unexpected modifications, and limiting access to the model weights, among other practices (<a href="https://media.defense.gov/2024/Apr/15/2003439257/-1/-1/0/CSI-DEPLOYING-AI-SYSTEMS-SECURELY.PDF">link</a>).</p><p></p><p>The Federal Office for Information Security in Germany also published a report titled &#8220;<strong>Generative AI Models: Opportunities and Risks for Industry and Authorities</strong>&#8221;. Opportunities include the use of language models to examine existing software for known security vulnerabilities, among others. The report also highlights 28 different risks, with examples including information extraction attacks&#8212;where a malicious user attempts to reconstruct an LLM or gain information about its training data, evasion attacks&#8212;where a malicious user modifies the input to an LLM in such a way that the LLM's behavior is deliberately manipulated, and poisoning attacks that aim to induce a malfunction or performance degradation by poisoning the targeted model (<a href="https://www.bsi.bund.de/SharedDocs/Downloads/EN/BSI/KI/Generative_AI_Models.pdf?__blob=publicationFile&amp;v=4">link</a>).</p><p></p><p>Thank you for reading! </p><p>Victor.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Issue #3 of the AI Safety & Governance Newsletter ]]></title><description><![CDATA[Policy frameworks for governing AI in Africa, AI licensing and accountability structures, requirements for U.S. agencies to prevent discriminatory use of AI, Canadian AI Safety Institute.]]></description><link>https://www.safeai.news/p/issue-3-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-3-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Sun, 07 Apr 2024 23:04:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the 3rd issue of the Artificial Intelligence (AI) Safety &amp; Governance Newsletter. Thank you for being a subscriber!</p><h3><strong>Key Reads</strong></h3><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Nigeria&#8217;s Ministry of Communications, Innovation and Digital Economy, is <strong>bringing together 120 AI researchers to co-create a national AI strategy</strong> (<a href="https://twitter.com/bosuntijani/status/1775468537554854019">link</a>). <a href="https://mitci.govmu.org/Documents/Strategies/Mauritius%20AI%20Strategy.pdf">Mauritius</a> was the first country in Africa to publish a National AI Strategy in 2018, and Rwanda and <a href="https://mcit.gov.eg/Upcont/Documents/Publications_672021000_Egypt-National-AI-Strategy-English.pdf">Egypt</a> have followed suit. Several other countries, including <a href="https://thefuturesociety.org/stakeholder-consultation-workshops-drive-insights-for-national-ai-strategies-in-tunisia-and-ghana/">Ghana and Tunisia</a>, are in various stages of the process.</p><p>Microsoft recently published a report on <strong>&#8220;policy frameworks&#8221; for governing AI in Africa,</strong> which highlights several of these efforts and proposes a legal and regulatory blueprint for AI governance on the continent. Key themes of the proposal include: the need to define a class of high-risk AI systems based on their capability, autonomy, and potential for large-scale harm; the role of governments in mandating mechanisms in such systems designed to detect, avoid, and disengage from unintended consequences; and the proposal that such AI systems should only be deployed on licensed AI infrastructure, thereby offering an additional layer of human oversight (<a href="https://drive.google.com/file/d/1NdulnulQFxMCur69jJfXM61EpvJtAFWe/view?usp=sharing">link</a>).</p><p></p><p>Most generative AI models are trained on text, image, and video data aggregated from the web. <strong>Christo Buschek and Jer Thorp investigate the curation and labeling process of LAION-5B</strong>, one of the largest and most commonly used datasets, and reveal several critical issues. The presence of illegal material in LAION-5B highlights the legal and ethical challenges in managing large datasets. Due to its sheer size, human curation is impractical, leading to reliance on automated processes that often introduce biases. These biases can be amplified when algorithmic data filters, as is commonly the case, are used (<a href="https://knowingmachines.org/models-all-the-way">link</a>).</p><p></p><p>The European Parliament released a briefing titled &#8220;<strong>What Think Tanks are Thinking</strong>&#8221; that references several position articles on AI and AI governance from a broad range of think tanks (<a href="https://drive.google.com/file/d/1w_daiccCwn8AEPVFcbE_2L3JX4YxB6eA/view?usp=sharing">link</a>). Here are a couple of them I found interesting:</p><p><strong>Licensing AI is not the answer&#8212;but it contains the answers (Tom Wheeler, Brookings Institution):</strong> While licensing AI efforts above a certain scale of capabilities may address some concerns, it is insufficient for effective oversight; instead, establishing standards and a new federal agency to oversee and enforce these standards is essential for governing AI in the public interest, ensuring safety, and fostering innovation and competition across all scales of AI development (<a href="https://www.brookings.edu/articles/licensing-ai-is-not-the-answer-but-it-contains-the-answers/">link</a>).</p><p><strong>The AI Election Year How to Counter the Impact of Artificial Intelligence (Katja Mu&#241;oz, German Council on Foreign Relations)</strong>: The strategic manipulation of information through generative AI poses a significant threat to democracy, particularly in an &#8220;AI election year&#8221; like 2024 with over 70 elections worldwide. This necessitates an assertive defense strategy and collective efforts from tech companies (e.g via content watermarking), social media platforms (e.g via better identity verification), Germany, the EU, and NATO to implement solutions, legislative measures, and tactics to combat these threats (<a href="https://dgap.org/en/research/publications/ai-election-year">link</a>).</p><p></p><p><strong>The U.S. government is implementing new, binding requirements for U.S. agencies to prevent discriminatory use of AI,</strong> including measures such as allowing travelers to refuse facial recognition scans at airports, establishing transparency and oversight for AI tools used by the government, and aiming to influence the AI industry through federal procurement policies (<a href="https://www.cnn.com/2024/03/28/tech/vp-kamala-harris-agencies-ai-technology/index.html">link</a>).</p><p></p><p><strong>The U.S. National Telecommunications and Information Administration (NTIA) published a report and recommendations on AI Accountability Policy</strong> (<a href="https://www.ntia.gov/issues/artificial-intelligence/ai-accountability-policy-report">link</a>). The report addresses several topics regarding the safety and governance of AI. I&#8217;ll highlight a key section on AI evaluation, edited for clarity:</p><blockquote><p>The use of audits within AI accountability structures should depend on the risk level, deployment sector, maturity of evaluation methodologies, and available resources. Audits are appropriate for high-risk applications or models, at least for validating claims about system performance, limitations, and governance controls. Audits should employ replicable, standardized, and transparent methods. We recommend that audits be required for high-risk AI systems and applications, with government support for an independent evaluation ecosystem. Audits should incorporate applicable standards recognized by federal agencies. Designating AI as high-risk outside of specific contexts is challenging; however, the Office of Management and Budget has provided guidance on rights-impacting and safety-impacting AI systems, with exemptions as necessary.</p></blockquote><p></p><p><strong>The Utah Senate Bill 149, known as the Artificial Intelligence Policy Act, has been signed into law.</strong> The new law introduces regulations on the use of generative AI in commercial communications and healthcare, requiring clear disclosures about AI use, establishing fines for violations, and creating a state Office of Artificial Intelligence Policy to oversee AI practices and innovation (<a href="https://www.jdsupra.com/legalnews/utah-enacts-first-ai-law-a-potential-7421259/">link</a>).</p><p></p><p>In The Economist, Madeleine I.G. Daepp and Robert Osazuwa Ness discuss how <strong>generative AI is being used to spread targeted disinformation in Taiwan,</strong> highlighting how fact-checkers and government agencies are struggling to keep up, and the need to develop new strategies for monitoring and countering disinformation at scale (<a href="https://www.economist.com/by-invitation/2024/03/26/video-will-kill-the-truth-if-monitoring-doesnt-improve-argue-two-researchers">link</a>).</p><p></p><p>Most AI researchers and companies agree that <strong>the government should not impose limitations on open-source AI, including blocking access to the weights of large-scale models.</strong> The NTIA received over 300 comments from the public, which will help inform a report to the President containing the NTIA&#8217;s findings and policy recommendations (<a href="https://www.ntia.gov/federal-register-notice/2024/ntia-receives-more-300-comments-open-weight-ai-models">link</a>). <a href="https://www.threads.net/@nickclegg/post/C5Bs5kqPAYO/?xmt=AQGz7zu2SOT7t6pGQvpgGpuxOQR8jvcGpOoraJgArS8KEA">Meta</a>, <a href="https://openai.com/global-affairs/openai-s-comment-to-the-ntia-on-open-model-weights">OpenAI</a>, and a group of researchers affiliated with the <a href="https://www.ias.edu/stsv-lab/aipolicy">Institute for Advanced Study </a>in Princeton have published their comments.</p><p></p><p>The Emerging Technology Observatory, a project of the Center for Security and Emerging Technology at Georgetown University published an analysis that shows that <strong>AI safety research constitutes 2% of all AI research.</strong> Despite its smaller proportion, AI safety research is highly cited indicating its impact within the broader AI research community (<a href="https://eto.tech/blog/state-of-global-ai-safety-research/">link</a>).</p><p></p><p><strong>The U.S. and UK recently signed a Memorandum of Understanding to collaborate closely on AI safety research and testing,</strong> leveraging shared scientific approaches and expert exchanges to enhance global AI safety standards and address emerging risks (<a href="https://www.commerce.gov/news/press-releases/2024/04/us-and-uk-announce-partnership-science-ai-safety">link</a>).</p><p></p><p>Achieving global governance of AI is complicated due to differences in countries' approaches to development, safety, and application, including in military contexts. Despite these challenges, Xue Lan, <strong>Chinese public policy expert, Professor and the Dean of Schwarzman College at Tsinghua University advocates for inclusive dialogue, technology sharing to bridge the intelligence gap, and the formation of globally inclusive governance mechanisms</strong> to ensure the ethical and safe advancement of AI technology (<a href="https://www.inewsweek.cn/people/2024-03-19/21498.shtml">link</a>; article needs to be translated from Chinese).</p><p></p><p>Content moderation involves screening and monitoring user-generated content on online platforms to ensure it adheres to the platform's rules, guidelines or legal requirements. Alyssa Boicel discusses in Tech Policy Press <strong>the challenges and limitations of using large language models (LLMs) for content moderation</strong>, particularly in the context of their inability to fully grasp cultural nuances, evolving language, and the reappropriation of derogatory terms (<a href="https://www.techpolicy.press/using-llms-to-moderate-content-are-they-ready-for-commercial-use/">link</a>).</p><p></p><p>U.S. Congress members, led by Co-Chair of the House AI Caucus, Anna G. Eshoo, and Congressman Neal Dunn, have introduced the <strong>Protecting Consumers from Deceptive AI Act</strong> to combat the challenges posed by AI-generated deepfakes and other deceptive content. The legislation aims to develop standards for identifying and labeling AI-generated content, requiring disclosures from AI developers and online platforms (<a href="https://eshoo.house.gov/media/press-releases/rep-eshoo-introduces-bipartisan-bill-label-deepfakes">link</a>).</p><p></p><p>The Canadian government plans to allocate $2.4 billion to enhance Canada's AI capabilities, and maintain its competitive edge globally. Key initiatives include investing in computing capabilities and technological infrastructure, supporting AI startups, aiding small and medium-sized businesses in integrating AI, and allocating funds to train workers in industries that could be potentially disrupted by AI. Additionally, Canada plans to establish a <strong>Canadian AI Safety Institute</strong> (<a href="https://www.pm.gc.ca/en/news/news-releases/2024/04/07/securing-canadas-ai-advantage">link</a>).</p><p></p><p><strong>The use of automated systems, including AI, by private and public entities can potentially result in unlawful discrimination and violate federal laws</strong>, given issues like biased datasets, lack of transparency, and incorrect assumptions in their development. Various U.S. federal agencies, including the CFPB, DOJ, EEOC, FTC, and others, in a recent statement, assert their commitment to enforcing civil rights, non-discrimination, and consumer protection laws, emphasizing that <strong>these laws apply to automated systems just as they do to traditional practices</strong> (<a href="https://www.eeoc.gov/joint-statement-enforcement-civil-rights-fair-competition-consumer-protection-and-equal-opportunity">link</a>).</p><p></p><h3>Opportunities</h3><p><strong>USAID is requesting comments as it seeks to develop a global AI research agenda</strong>. The key questions focus on identifying effective practices, guidelines, and frameworks to ensure that AI research is conducted safely, ethically, and with consideration for global contexts (<a href="https://www.federalregister.gov/documents/2024/03/14/2024-05357/global-ai-research-agenda">link</a>).</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI]]></title><description><![CDATA[Some of the sentences on this newsletter are produced by AI exclusively as succinct paraphrases of content from the original sources.]]></description><link>https://www.safeai.news/p/ai</link><guid isPermaLink="false">https://www.safeai.news/p/ai</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Tue, 26 Mar 2024 22:40:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Some of the sentences on this newsletter are produced by AI exclusively as succinct paraphrases of content from the original sources. Every word is carefully reviewed by myself for clarity and coherence.</p>]]></content:encoded></item><item><title><![CDATA[The /ai page]]></title><description><![CDATA[The AI page is a documentation of the use of AI on a website.]]></description><link>https://www.safeai.news/p/the-ai-page</link><guid isPermaLink="false">https://www.safeai.news/p/the-ai-page</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Tue, 26 Mar 2024 22:36:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The AI page is a documentation of the use of AI on a website. Generative AI is now capable of producing content that is indistinguishable from those created by humans. As we entire this new era of communication, transparency about the content we share with others is paramount for building trust and credibility in our work. Daramola Morenikeji came up with the idea of AI page.</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:142228725,&quot;url&quot;:&quot;https://www.bydamo.la/p/ai-manifesto&quot;,&quot;publication_id&quot;:18950,&quot;publication_name&quot;:&quot;More! by Damola&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5235c98f-7a96-4ce7-bb49-761c3553a7d9_600x600.png&quot;,&quot;title&quot;:&quot;The /ai 'manifesto'&quot;,&quot;truncated_body_text&quot;:&quot;Large language models have become good at crafting texts, audio, still, and moving images. We are getting to a point where to retain the trust others have in you and your work, tell us how you made use of AI in your craft. Create a page on your website that reflects how you use generative AI. Let&#8217;s call it the /ai page. That is, yourwebsite/ai.&quot;,&quot;date&quot;:&quot;2024-03-04T11:00:08.934Z&quot;,&quot;like_count&quot;:0,&quot;comment_count&quot;:0,&quot;bylines&quot;:[{&quot;id&quot;:4419615,&quot;name&quot;:&quot;Damola Morenikeji&quot;,&quot;handle&quot;:&quot;more&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e170ec4-b38f-4a4c-a381-dda03cad17bd_600x600.png&quot;,&quot;bio&quot;:&quot;Learner at damolamorenikeji.com&quot;,&quot;profile_set_up_at&quot;:null,&quot;publicationUsers&quot;:[],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:false,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.bydamo.la/p/ai-manifesto?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!frPy!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5235c98f-7a96-4ce7-bb49-761c3553a7d9_600x600.png"><span class="embedded-post-publication-name">More! by Damola</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The /ai 'manifesto'</div></div><div class="embedded-post-body">Large language models have become good at crafting texts, audio, still, and moving images. We are getting to a point where to retain the trust others have in you and your work, tell us how you made use of AI in your craft. Create a page on your website that reflects how you use generative AI. Let&#8217;s call it the /ai page. That is, yourwebsite/ai&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">2 years ago &#183; Damola Morenikeji</div></a></div><p>The AI page for this website is at https://www.safeai.news/ai</p>]]></content:encoded></item><item><title><![CDATA[Issue #2 of the AI Safety & Governance Newsletter]]></title><description><![CDATA[Welcome to the 2nd issue of the Artificial Intelligence (AI) Safety & Governance Newsletter.]]></description><link>https://www.safeai.news/p/issue-2-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-2-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Sun, 24 Mar 2024 23:55:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the 2nd issue of the Artificial Intelligence (AI) Safety &amp; Governance Newsletter. Thank you for being a subscriber! </p><p>AI is not a new technology, but it is increasingly becoming an important part of the way many people around the world live. <strong>This newsletter is a curation of content and news pertaining to how technologists, policymakers, and everyone in between are working towards ensuring the impacts of AI are beneficial.</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3>Key Reads</h3><p>All 193 members of the United Nations General Assembly have unanimously adopted a <strong>resolution aimed at governing artificial intelligence (AI)</strong>, marking a global consensus for the development of safe, secure, and trustworthy AI systems (<a href="https://news.un.org/en/story/2024/03/1147831">link</a>). The resolution emphasizes the need for capacity building to close digital divides, discourages the use of AI to undermine peace or repress human rights, and calls for the responsible development and deployment of AI technologies. It highlights the mutual reinforcement of innovation and regulation, and the importance of education, transparency, and knowledge sharing on a voluntary basis.</p><p>Although non-binding, the resolution is seen as a step toward using AI to tackle existential threats and seize opportunities across various sectors, including healthcare, education, agriculture, and democracy promotion. It sets the stage for future collaboration within the UN and beyond. Representatives from Japan, Morocco, Singapore, and the UK, in a subsequent media stakeout (<a href="https://www.youtube.com/watch?v=YwO701_MxJ0&amp;t=919s">link</a>), expressed their support and outlined their views on the resolution's impact.</p><p></p><p>While there is no consensus on the likelihood and severity of risks from near-future AI systems, the White House's executive order on AI, issued on October 30, requires companies developing &#8220;dual-use foundation models&#8221; of AI&#8212;which could pose significant safety and security risks&#8212;to report various related information to the government. In a recent article, Thomas Woodside of The Center for Security and Emerging Technology at Georgetown University argues for <strong>the importance of the U.S. Congress focusing on establishing an early warning system for severe risks posed by advanced AI systems</strong> as part of their legislative agenda on AI (<a href="https://cset.georgetown.edu/article/keeping-up-with-the-frontier/">link</a>). He asserts that legislative action is necessary to solidify and enhance these provisions beyond the limitations of an executive order, which relies on the older Defense Production Act and may be rescinded by future administrations.</p><p></p><p>The cost of developing cutting-edge AI technology is increasingly becoming prohibitive for academics, as large-scale compute clusters demand substantial financial resources that typical research labs may not be able to cover with research grants. Consequently, academics are becoming more dependent on major technology companies, thus pushing AI researchers to orient their work towards commercial applications. Fei-Fei Li of Stanford <strong>advocates for the creation of a national AI repository to democratize access to computing resources and data</strong> (<a href="https://www.washingtonpost.com/technology/2024/03/10/big-tech-companies-ai-research/">link</a>). This initiative aims to counter the dominance of tech giants and preserve the independence of AI research. Conversely, Meredith Whittaker, President of Signal and Chief Advisor to The AI Now Institute, recently stated in a tweet that &#8220;...proposals to alleviate this imbalance largely exacerbate it, insofar as they amount to the government paying Big Tech to provide academics access to Big Tech resources&#8221; (<a href="https://twitter.com/mer__edith/status/1766871907821883402">link</a>).</p><p></p><p>In a recent panel discussion, experts from The RAND Corporation were unified in recognizing the importance of independent research to understand AI's risks and implications for public policy (<a href="https://www.rand.org/pubs/commentary/2024/03/is-ai-an-existential-risk-qa-with-rand-experts.html">link</a>).&nbsp; An analogy of <strong>AI as a slow-moving catastrophe, akin to climate change, was drawn, emphasizing incremental harms</strong> that could gradually erode the quality of life and essential societal structures. The discussion also touched upon existential threats, with varying opinions on whether AI poses an unrecoverable harm to humanity. While some panelists viewed AI as a magnifier of existing problems rather than an existential threat, others raised concerns about AI's potential to disrupt meaningful human activity and connections, positing a more profound existential risk.&nbsp; Policy solutions, such as transparency, oversight, and diversity in AI development, were suggested as means to mitigate these risks.</p><p></p><p>A recently introduced bill in the UK&#8217;s House of Lords seeks to establish a comprehensive regulatory framework for AI in the UK (<a href="https://drive.google.com/file/d/1gu4NnRGvsfeeGY4zuYyELw4VpFATXExt/view?usp=sharing https://www.bloomberglaw.com/public/desktop/document/ONanetalvDatabricksIncetalDocketNo324cv01451NDCalMar082024CourtDo?doc_id=X78MN4GVSRC9F4OTESFJ7KII697">link</a>). Among many important stipulations, the bill mandates that <strong>individuals involved in AI training must document and report the use of third-party data and intellectual property (IP) to an authority,</strong> ensuring all data and IP are utilized with informed consent and in compliance with applicable laws. Copyright and IP issues are increasingly becoming contentious within the AI industry, exemplified by several high-profile lawsuits, such as a recent case involving book authors and the software company Databricks.</p><p></p><p>A new series of articles published by the AI Now Institute &#8220;<strong>survey the nationalist narratives around AI and emergent industrial policies being proposed by governments with differing economic and geopolitical motivations</strong>&#8221;. I&#8217;ve only read the first article and it was deeply insightful.</p><p>The first article, written by Susannah Glickman, delves into the deep historical intertwining of the tech industry, government intervention, and the development of AI and semiconductors within the United States. It highlights how government policies and funding have played pivotal roles in the evolution of technology, dispelling the myth of the lone tech entrepreneur and emphasizing the importance of state support and intervention (<a href="https://ainowinstitute.org/publication/ai-and-tech-industrial-policy-from-post-cold-war-post-industrialism-to-post-neoliberal-re-industrialization">link</a>).</p><p>The narrative begins in the mid-1980s, tracing the emergence of 'tech' as a category and its subsequent growth, which was fueled by government intervention. It showcases how the tech industry has been closely linked with national security, government priorities, and economic policy, illustrating this through various administrations' approaches to industrial policy, such as the CHIPS Act and executive orders focusing on semiconductor manufacturing. The history of AI is deeply intertwined with government funding and its strategic importance to national security. The development of AI is shown to be reliant on advances in semiconductor technology, which, in turn, benefits from massive government support and global cooperation.</p><p>Some key highlights include the Cold War's influence on tech development, the strategic importance placed on semiconductors and AI during different administrations, the impact of global competition (notably with Japan), and the shift towards a fabless model in semiconductor manufacturing. The article also discusses the recent re-engagement of government with the semiconductor industry in response to global challenges and supply chain issues, underscoring the ongoing significance of government involvement in sustaining the tech industry.</p><p>The narrative culminates in a discussion on the current state and future prospects of AI and semiconductors, emphasizing how intertwined these fields remain with governmental policies and the broader political economy. It asserts that the history of government intervention in tech is not just a series of isolated events but a fundamental aspect of how the tech industry has evolved and operates within the global economy.</p><p>The other articles in the series examine the policies and narratives in other countries, including South Africa, the UK, India, and the UAE (<a href="https://ainowinstitute.org/ai-nationalisms">link</a>).</p><p></p><p>Former IBM CEO Ginni Rometty expressed strong support for up-skilling in a recent podcast, discussing how <strong>AI can assist in job performance and provide opportunities for those with less formal education</strong> or credentials due to rapidly changing skills. In the same podcast, Drago&#537; Tudorache, a member of the European Parliament, explores the potential impact of AI on the workforce (<a href="https://www.washingtonpost.com/podcasts/post-live/global-efforts-to-build-guardrails-around-artificial-intelligence---/">link</a>).</p><p></p><p>A recent investigation conducted by the AI Democracy Projects tested five leading AI text models (OpenAI's GPT-4, Anthropic's Claude, Google's Gemini, Meta's Llama 2, and Mistral's Mixtral) for their ability to provide accurate and reliable election-related information. Despite AI companies' pledges to promote information integrity, the study found that the <strong>AI models often delivered responses that were inaccurate, misleading, incomplete, or harmful, with all models struggling to correctly answer queries about election rules,</strong> such as the prohibition of campaign-related apparel at Texas polling places. Among the models, GPT-4 performed best in terms of accuracy, but the overall findings highlighted significant gaps in the models' utility for providing public election information (<a href="https://www.proofnews.org/seeking-election-information-dont-trust-ai/">link</a>).</p><p></p><p>Melissa Heikkil&#228; wrote a comprehensive summary in MIT Technology Review (<a href="https://www.technologyreview.com/2024/03/19/1089919/the-ai-act-is-done-heres-what-will-and-wont-change/?utm_source=engagement_email&amp;utm_medium=email&amp;utm_campaign=wklysun&amp;utm_term=03.24.24.nonsubs_eng&amp;mc_cid=f23ceeee78&amp;mc_eid=2e319ffde4">link</a>) about how the <strong>development and use of AI within the EU will be influenced by the recently enacted EU AI Act.</strong>&nbsp;</p><p></p><p>The first <strong>global summit on deepfake sexual abuse</strong>, organized by My Image My Choice and partners, including Bumble, convened experts across fields to discuss the burgeoning issue of deepfake-generated pornography. Deepfakes are 'synthetic media that have been digitally manipulated to replace one person's likeness convincingly with that of another.' This form of abuse, notably targeting women, has seen a dramatic increase in instances due to easily accessible AI tools. The experts called for stronger enforcement of terms of service by tech giants, legal reforms including adjustments to Section 230 of the US Communications Decency Act, and cautioned against the inadequacy of AI watermarking (<a href="https://www.techpolicy.press/ai-experts-officials-and-survivors-talk-policy-solutions-in-first-ever-global-summit-on-deepfake-abuse/">link</a>).</p><p></p><p>In a recent article, Arvind Narayanan and Sayash Kapoor argue that <strong>AI safety largely depends on the context and environment in which the AI is deployed.</strong> The article posits that focusing solely on making AI models &#8220;safe&#8221; overlooks the potential for misuse in varying contexts, as models lack the necessary information to discern between harmful and benign uses. They recommend shifting the focus of safety efforts away from model alignment alone and advocate for the development of external defenses against misuse. They also suggest a more nuanced assessment of marginal risk - risks that would otherwise not be present, a reorientation of red teaming towards identifying early warnings of misuse, and third-party-led red teaming to ensure unbiased risk assessments (<a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">link</a>).</p><p></p><h3>Opportunities</h3><p>The <strong>UN Secretary-General's AI Advisory Body is seeking input on its recently launched Interim Report on AI Governance.</strong> The key questions for further discussion in the next phase of work by the Advisory Body focus on creating data protection frameworks and common standards for data labeling, promoting shared compute resources and determining the financing and capacity-building mechanisms necessary for effective international governance, among others. There's an emphasis on enhancing AI talent through international exchanges, joint academic programs, and cross-domain development, alongside leveraging international collaboration to support scientific research. Key challenges include reaching consensus on AI risks, especially concerning autonomous weapons, and establishing enforceable guidelines for ethical boundaries.</p><p>Link: <a href="https://www.un.org/en/ai-advisory-body">https://www.un.org/en/ai-advisory-body</a>&nbsp;&nbsp;<br><br></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Issue #1 of the AI Safety & Governance Newsletter]]></title><description><![CDATA[Welcome to the 1st issue of the Artificial Intelligence (AI) Safety & Governance Newsletter.]]></description><link>https://www.safeai.news/p/issue-1-of-the-ai-safety-and-governance</link><guid isPermaLink="false">https://www.safeai.news/p/issue-1-of-the-ai-safety-and-governance</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Mon, 11 Mar 2024 02:15:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TAP-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6e9f903-4d9a-4fe2-94ad-ab2f650372dc_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the 1st issue of the Artificial Intelligence (AI) Safety &amp; Governance Newsletter. AI is not a new technology, but it is increasingly becoming an important part of the way many people around the world live. <strong>This newsletter is a curation of content and news pertaining to how technologists, policymakers, and everyone in between are working towards ensuring the impacts of AI are beneficial.</strong></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>Key reads</strong></h3><p>The African Union Development Agency has published (<a href="https://drive.google.com/file/d/1FNQnv_14-zRBySdN1ySjC9vT9DmtZ3P0/view?usp=sharing">link</a>) an <strong>AI Continental Roadmap</strong> for the continent with proposals for measures African countries should adopt for an &#8220;inclusive and sustainable AI-enabled socioeconomic transformation&#8221;. The strategic pillars focus on developing human capital, leveraging infrastructure and data, creating a supportive environment, establishing an economic climate, building partnerships, and enhancing monitoring and evaluation capacities to strengthen AI systems and strategies in African countries.</p><p></p><p><strong>Can AI be an inventor?</strong> In the US, the answer seems to be No (<a href="https://public-inspection.federalregister.gov/2024-02623.pdf">link</a>), but in Brazil, a newly introduced Bill (<a href="https://ipwatchdog.com/2024/02/29/brazilian-lawmaker-introduces-bill-allow-ai-inventor/id=173809/#">link</a>) seeks to recognize AI as patent inventors.</p><p></p><p>Several biology-technology (Bio-tech) researchers from around the world are articulating <strong>values and principles to guide the responsible development of AI technologies in the field of protein design</strong>. They emphasize safety, security, equity, international collaboration, openness, responsibility, and societal benefit. The commitments focus on conducting beneficial research, preparing for health emergencies, adhering to biosecurity in DNA synthesis, improving DNA screening, continuously evaluating software risks, promoting scientific openness, and updating practices as needed to ensure AI developments are safe, secure, and beneficial for all. See: <a href="https://responsiblebiodesign.ai">https://responsiblebiodesign.ai</a></p><p></p><p>In a recent analysis (<a href="https://www.bloomberg.com/graphics/2024-openai-gpt-hiring-racial-discrimination/">link</a>) on Bloomberg, AI <strong>models from OpenAI were found to exhibit clear signs of name-based discrimination</strong>: resumes with names distinct to Black-Americans were the least likely to be ranked as the top candidate for a financial analyst role, compared to resumes with names associated with other races and ethnicities.</p><p></p><p>Valentin Hofmann tweets &#8220;when LLMs are asked to pass judgment on defendants who committed murder, they choose the death penalty more often when the defendants speak African American English rather than Standardized American English &#8230; without being overtly told that they are African American&#8221; (<a href="https://twitter.com/vjhofmann/status/1764687442047570053">link</a>). This is one of several analyses done in a recent paper on <strong>dialect prejudice when AI is used to assess employability, criminality and character</strong>. Paper <a href="https://arxiv.org/abs/2403.00742">link</a>.</p><p></p><p>In a recent study (<a href="https://arxiv.org/abs/2402.17861">link</a>), researchers from Brown, CMU, Data &amp; Society and Mozilla Foundation <strong>survey the landscape of AI audit tooling to understand the challenges to accountability and opportunities for tool development and research</strong>. AI audits are independent evaluations of performance, risks and safety of AI systems. Standards and regulations of AI systems around the world are largely still in early stages. The study on AI audit tools reveals a rich landscape but with notable gaps particularly in harm discovery, audit communication, and advocacy, emphasizing the need for tools that support comprehensive evaluation and genuine accountability. The authors call for a focus on legal clarity, community engagement, and interdisciplinary standards to enhance AI accountability.</p><p></p><p>A team of ethical hackers identified <strong>critical vulnerabilities in Google's AI and Cloud services</strong>, including unauthorized access to user images via an <a href="https://cheatsheetseries.owasp.org/cheatsheets/Insecure_Direct_Object_Reference_Prevention_Cheat_Sheet.html">Insecure direct object reference</a> (IDOR) in Bard's Vision feature, a DoS vulnerability through GraphQL directive overloading, a security issue with a hardcoded signature key, and the most significant finding: exfiltrating sensitive Google Workspace data via Bard (now Gemini) by injecting custom markdown. Security challenges in AI systems are constantly evolving as evidenced by these discoveries for which they were awarded significant bounties (<a href="https://www.landh.tech/blog/20240304-google-hack-50000">link</a>).</p><p></p><p>The <strong>US Department of Veterans Affairs is leveraging AI to improve veterans' healthcare</strong> by enhancing diagnostic accuracy, predicting cancer risks, and optimizing treatment paths. Through an initiative launched in 2017, they identified over 117,000 veterans at high suicide risk, leading to increased outpatient attendance, new safety plans, and reduced mental health admissions and suicide attempts. In addition, a tool called STORM assists in managing opioid risk through predictive models. Another FDA-authorized system, GI Genius, has been deployed in 106 facilities, enhancing the detection of precancerous polyps during colonoscopies. Listen to a recent oversight hearing (<a href="https://veterans.house.gov/calendar/eventsingle.aspx?EventID=6371">link</a>).</p><p></p><p>ASCII art, a technique that uses characters from the <a href="https://www.ascii-code.com/">ASCII set </a>to create images, has been identified as a vulnerability in the security of large language models (LLMs) like GPT-3.5, GPT-4, Gemini, Claude, and Llama2 (<a href="https://arxiv.org/abs/2402.11753">link</a>). Researchers have discovered that these <strong>state-of-the-art models are susceptible to "jailbreaks" through ASCII art</strong>, allowing individuals to circumvent safety protocols. This vulnerability enables the elicitation of undesired behaviors, such as producing instructions for making bombs if the prompts are cleverly disguised as ASCII art. This security loophole arises because current safety measures for LLMs primarily focus on interpreting the semantic content of text, overlooking the potential for ASCII art to convey complex and potentially harmful instructions beyond simple semantics.</p><p></p><p>Compute governance is an emerging tool for AI governance. Researchers from a cross-section of institutions in a recent paper (<a href="https://arxiv.org/abs/2402.08797">link</a>), emphasize that computing power (compute) is crucial for AI progress, with a 350 million-fold increase in compute used for training leading AI systems over thirteen years. <strong>Compute governance involves regulating the access and usage of computing power to achieve policy goals,</strong> enhance safety, and manage the development and deployment of AI technologies. The paper proposes that compute governance can serve three main purposes: enhancing visibility into AI development and usage, allocating resources to foster desirable AI projects, and enforcing rules to ensure safe and lawful AI deployment. It also explores innovative enforcement mechanisms, including physical limitations on hardware and multiparty control protocols to mitigate risks associated with large-scale AI systems.</p><p></p><p>In a recent paper (<a href="https://knightcolumbia.org/blog/a-safe-harbor-for-ai-evaluation-and-red-teaming">link</a>), researchers from multiple institutions including CMU, MIT and Stanford, <strong>advocate for legal and technical safe harbors to protect researchers conducting good-faith evaluations of proprietary AI models,</strong> highlighting the importance of independent assessments in identifying vulnerabilities and ensuring AI safety. They propose voluntary commitments from AI companies to facilitate public interest research without fear of legal repercussions or account termination to create an equitable, transparent, and accountable AI research ecosystem. In the same light, calls for broader access to these AI models i.e access to their Weights, Activations, or Gradients&nbsp;were made by another group of researchers (<a href="https://arxiv.org/abs/2401.14446">link</a>).</p><p></p><p>Anna Makanju, VP of Global Affairs at OpenAI, on a recent World Economic Forum Podcast (<a href="https://www.weforum.org/podcasts/radio-davos/episodes/ai-governance-2024/">link</a>), <strong>advocates for a global regime that includes every country to address catastrophic risks of AI</strong>, suggesting a model similar to the International Atomic Energy Agency (IAEA) for AI governance.</p><p></p><p>The EU AI Act (<a href="https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai">link</a>), potentially the first comprehensive legal framework for AI globally, aims to ensure that AI systems are trustworthy by mandating that they respect fundamental rights, safety, and ethical principles. It introduces a risk-based approach to regulation, categorizing AI applications into four levels of risk and applying corresponding requirements, ranging from transparency obligations for low-risk AI to strict controls and obligations for high-risk applications. <strong>The European Parliament is scheduled to vote on the AI Act on 13 March</strong>. This vote is one of the last steps before the law is enacted, but enforcement will not begin until 2025.</p><p></p><p>Amba Kak, a co-director of the AI Now Institute writes in Foreign Policy (<a href="https://foreignpolicy.com/2024/02/12/ai-public-private-partnerships-task-force-nairr/">link</a>) that we need to <strong>critically examine whether large, and increasingly consolidated AI companies have interests that align with broader society.</strong>&nbsp; She writes&nbsp; &#8220;In an overwhelming focus on AI-driven harms, we&#8217;ve missed a key piece of the puzzle: demanding that firms articulate, clearly and with evidence to back it, what the benefits of AI are to the public &#8230; we haven&#8217;t asked enough of AI firms to show us their homework, instead permitting them to coast on shallow assertions that AI will inevitably lead us down the path of technological innovation.&#8221;</p><p></p><h3><strong>Opportunities</strong></h3><p>The DARPA AI Cyber Challenge (AIxCC) is a competition to develop AI-enabled systems for real-time identification and remediation of software vulnerabilities. Registration is currently open and there is a $4 million prize pool.</p><p><a href="https://www.darpa.mil/news-events/2023-12-14">https://www.darpa.mil/news-events/2023-12-14</a>&nbsp;</p><p>The data.org Privacy Enhancing Technologies (PETs) for Public Health Challenge is currently open. The challenge involves demonstrating how financial transaction data, while ensuring privacy, can be leveraged to inform and support epidemiological policy decisions and public health responses during pandemics.</p><p><a href="https://data.org/initiatives/pets-challenge/">https://data.org/initiatives/pets-challenge</a></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.safeai.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Safety &amp; Governance Newsletter! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>