Harnessing AI Risk Initiative

The convergence of accelerating AI innovation and unregulated digital communications has brought us to what may be the most critical juncture in human history.

We can still turn the dawn of this new era into the greatest opportunity for humanity, but only if we come together globally like never before to govern its immense risks and opportunities.

Fact Sheet

Mission and Goals:

  • Enable humanity to manage AI's immense global risks for safety and concentration of power, and realize and equitably share its opportunities, while affirming the subsidiarity principle.

  • Facilitate an open, timely, democratic and efficient treaty-making process to build a global federal intergovernmental organization for AI governance.

  • Inspired by the 1946 Baruch Plan proposal for international control of nuclear weapons/energy.

Strategy:

  • Aggregate a critical mass of globally diverse NGOs and states to design and launch an open, expert and participatory constituent process.

  • Use an intergovernmental constituent assembly treaty-making model, inspired by the process that led to the US Constitution in 1787.

  • Aim to include both AI superpowers jointly. Smaller states are incentivized for joining early and by the massive economic advantage of a Global Public Interest AI Lab, to be built even before superpowers join.

Current Status:

  • Convened by the Trustless Computing Association and announced on September 10th 2024 by its six founding NGO partners.

  • Actively onboarding additional NGO partners and experts, states and donors.

Roadmap:

  • Short-term goals focus on expanding the coalition, engaging states and experts.

  • 2nd Pre-Summit on October 24th, 2024 in Geneva/Online

  • 1st Harnessing AI Risk Summit in November 28-29th 2024 in Geneva/Online

  • Aim to onboard progressively 3-7 highly diverse states by December 2024 and then at least 15-30 by November 2025.

  • Long-term goal of establishing the intergovernmental organization for AI governance.

Positioning and Challenges:

  • Recognizes the need to engage AI superpowers (US, China) while maintaining neutrality.

  • Positions itself as more inclusive and democratic than other AI governance initiatives.

  • Emphasizes urgency due to rapid AI development and potential risks.

  • The coalition is in its early stages, actively seeking partners, funding, and engagement from states and experts to advance its ambitious goals for global AI governance.


ExeCutive Summary

The Harnessing AI Risk Initiative is aggregating an open coalition of NGOs and a critical mass of globally diverse states to facilitate a democratic, expert-led treaty-making process to create a democratic, federal and expert-led global intergovernmental organization - including a stat-of-the art AI lab - that can reliably stave off the AI immense risks for human safety and for unaccountable concentration of power, while realizing and sharing equitably its benefits and control. Starting September 10th 2024, the Initiative will be advanced primarily through an open and participatory Coalition for a Baruch Plan of AI, centered on advancing its core goals and methods.

Risks and Opportunities of AI 

It only took eighteen months since the release of ChatGPT for a vast majority of experts, world leaders, and citizens to become cognizant of the immense risks of AI to human safety, global concentration of power, and potential distortion of human nature. Many experts believe both risks could materialize unpredictably soon due to radical algorithmic innovations.  At the same time, the advantages of AI are becoming increasingly apparent, posing extraordinary opportunities to accelerate scientific advancement, generate abundance, eradicate poverty, and improve education and the quality of life itself. Reflecting the urgency and scale of this issue, hundreds of AI experts and influential figures have called via AItreaty.org for a robust and binding global treaty for AI, and so has the Pope. Such an idea is even shared by 77 percent of U.S. citizens. 

Shortcomings of Current Global AI Governance Initiatives

Historically, the security agencies of a handful of countries have safeguarded humanity against the dangers and misuse of powerful technologies. Although we have averted major catastrophes, the risks associated with nuclear and bioweapons are greater today than ever before. While the recent Council of Europe treaty on AI and other initiatives for international coordination, such as the AI Summit series, are praiseworthy, they fall massively short in addressing the principal risks for safety and concentration of power and lack mechanisms for their future improvement. The current dominant path toward treaty-making and global governance of AI is centered on a tightly US-led coalition of "liberal democracies" and rich autocracies pitched against​ dozens of "undemocratic" states, including China. This is inadequate because: (1) Reliably preventing dangerous AI proliferation will require unprecedented global compliance; (2) If successful, it is likely to amount to immense undemocratic concentrations of power and wealth in very few states, firms, ​and/or agencies; (3) The so-called "undemocratic" states represent a rich variety of human ​political and cultural models that​ are entrenched, legitimate, ​and often comparable in levels of democraticness​ with the "West" one. Hence, we may need to consider a much more inclusive and globally democratic approach that faces head-on the unprecedented levels of global coordination that are needed.  

Aims of the Initiative and Coalition

As detailed on the Coalition's home page, "Amidst an accelerating, reckless, winner-take-all race among states and their firms for ever more capable AGI and ASI forms of AI, there may still be time for humanity to properly come together to (A) ensure all AIs will be human-controllable and and largely controlled and shaped by humanity as a whole, (B) prevent catastrophic AI misuses and runway ASIs, and maximize the chances that, if ASI will emerge, it will result in a beneficial outcome for humanity, democratically imbued with the most widely-shared positive human values. We are leading an early open coalition of NGOs and states to advance a much more timely, effective, expert, and democratic treaty-making process for AI based on the open intergovernmental constituent assembly model - akin to that which led to the US federal constitution. This new constituent assembly welcomes all states and citizens' representative entities, irrespective of their political orientation. Over a time frame as short as realistically possible - yet with extreme care - the coalition aims to build a strong, expert-led, federal, and democratic governance of AI as it was attempted in 1946 for nuclear technologies via the Baruch Plan while avoiding its failure due to vetoes".

Expert’s calls for the Baruch Plan for AI

While calls by top AI scientists and CEOs about the need for a strong global and democratic governance of AI abound, several notable figures have referred to it directly and literally. Even more than Hiroshima, the fast emergence of a broad consensus among top scientists and political leaders on the immense challenge of state control of atomic energy coalesced in the Acheson-Lilienthal Report and led the US to propose the Baruch Plan to the UN. Such consensus is already emerging for AI.
In 2014, Nick Bostrom referred to the Baruch Plan in his foundational Superintelligence book as a (positive) future scenario for the governance of AI. In 2018, Ian Hogarth, Chair of the UK AI Safety Institute, called for the Baruch Plan as a solution to AI governance. In 2021, the Baruch Plan application to AI was explored in depth and endorsed as an inspiration in a paper titled "International Control of Powerful Technology: Lessons from the Baruch Plan for Nuclear Weapons" by Waqar Zaidi and Allan Dafoe, the president of the Centre for Governance of AI of Oxford and current Head of Long-Term AI Strategy and Governance at DeepMind. In March 2023, the Trustless Computing Association extensively referred to the Baruch Plan as a critical inspiration in a LinkedIn post announcing its Harnessing AI Risk Initiative. In May 2023, the Economist reported that Jack Clark, global head of policy and cofounder of Anthropic, suggested the Baruch Plan for AI. In December 2023, Jaan Tallinn suggested the Baruch Plan in an interview as a potential solution to the global governance of AI. On July 7th, 2024, Yoshua Bengio mentioned the Baruch Plan as an "important avenue to explore to avoid a globally catastrophic outcome." Also, in March 2023, Sam Altman suggested that the intergovernmental constituent assembly process that led to the US Constitutional Convention of 1787 should be the "platonic ideal" of the treaty-making process we need to build a proper global governance for AI.

Strategy

Our strategy is centered on expanding and deepening the Coalition by actively engaging with states, NGOs, industry experts, and companies by first expanding participation in its guidance via a Secretariat of mission-aligned NGOs and then, in parallel, an open coalition of states and AI labs. This will be achieved through a series of events and summits both in Geneva and other locations, beginning with the 1st Harnessing AI Risk Summit on November 20-21st, 2024. The first seven states and private AI labs to join the initiative will enjoy substantial yet temporary economic and political advantages and agree on an initial version of rules for the constituent assembly. Considering the vast disparity in power between states, particularly in AI and more broadly, and recognizing that three billion people are illiterate or lack internet access, we foresee the voting power in such assembly to be initially weighted by population and GDP.  

AI Superpowers' Veto?

While the participation of the US and China is crucial for achieving the initiative's AI safety goals, other states will have very strong incentives beforehand. By incorporating in the assembly's mandate the creation of a state-of-the-art public-private $15+ billion Global Public Benefit AI Lab - and a "mutually dependent" and eventually autonomous supply chain - participant states and AI labs will secure vast economic and political benefits. They will gain cutting-edge industrial AI capabilities, digital sovereignty, political leadership, and an enhanced negotiating power vis-a-vis other less inclusive global governance initiatives. Participation will remain open to all states at all stages of the process, including during the mandatory, periodic statutory reviews of the treaty's charter.

Reconciling Freedom, Democracy, and Safety in Global Governance of AI

We believe that rather than finding a balance or trade-off in a "zero-sum game" among those values and necessities, our initiative for a global constituent process for AI and the resulting global organization can maximize each concurrently in a sum that is larger than its parts. It can do so via a) careful engineering of the constituent processes centered on the principles of federalism and subsidiarity; b) sophisticated, time-proven constitution and constituent process science; b) applying battle-tested, democratic, and trustless computing paradigms for society-critical IT and socio-technical systems. Just as freedom, fairness, and safety were all concurrently increased when "Wild West" US states were turned in 1787 into liberal democratic states under the US federal constitution, the same can be achieved in the process of building federal global governance of AI in the "Wild West" that is the world today.

The Trustless Computing Association

The Trustless Computing Association (TCA), established in 2015 and headquartered in Geneva, is a non-profit organization that promotes the development of international standards and institutions for the secure and democratic governance of digital communications and Artificial Intelligence (AI). Our efforts span leading research, producing scholarly publications, organizing influential conference series, engaging with nation-state officials, and conducting focused campaigns.

Traction with NGOs and Experts: The Coalition 

We have built and are extending a vast network of over 22 world-class long-time R&D partners since 2015 and 28 distinguished advisors. A wide and expanding number of world-renowned experts, distinguished individuals, diplomats, and NGOs will participate as speakers of our 1st Harnessing AI Risk Summit to be held this Fall in Geneva participated as speakers of its hybrid Pre-Summit held along the G7 Summit in Italy last June 12th. In April, we launched a Coalition for the Harnessing AI Risk around an Open Call, joined by 7 NGOs and over 45 individuals. In July, such a coalition was merged into a new, more participatory Coalition for a Baruch Plan for AI launched with 5 leading NGOs and centered on the main goals and methods of the Initiative.

Traction with States and AI Firms

Over the last four months, we have held bilateral and multilateral meetings with several interested states, especially from Africa and Europe, including a number of ambassadors to the UN in Geneva. We  have received formal interest from the mission of the largest regional intergovernmental organization to the UN. Since December, we have had ongoing highly preliminary discussions with three of the top five AI labs regarding their participation in the Global Public Benefit AI Lab, as many of them have called for something similar. 

Goals and Impact

Within the next three years, through the Initiative and its Coalition, we aim to have contributed significantly to the creation of an open, efficient, expert, and democratic constituent process that has led to the establishment of an open intergovernmental organization for AI, capable of avoiding the immense risks, while realizing the extraordinary opportunities. At least dozens of diverse states will have joined the constituent process and the resulting organization. AI superpowers have joined, or at least are actively and collaboratively engaged. The initiative's success will serve as a model for the global governance of other dangerous technologies.

Funding

While up until last year our association has benefited from modest funding in various forms, including from its spin-off and from EU public agencies, our recent efforts have been dedicated to shaping the new initiative on a purely voluntary basis. Currently, our work is sustained by two committed full-time volunteer staff members, Rufo Guerreschi and Marta Jastrzębska, supported by contributions from 28 advisors and partners. To meet our strategic goals over the next three years, we require funding of $1.5 million, but we could effectively use up to $5 million to leverage and fully empower our Coalition and directly its members. Given our proven track record of efficiency and cost-effectiveness, securing $80,000 would allow us to cover essential operations, engage with key stakeholders and partners, secure the logistics for the Harnessing AI Risk Summit on November 28-29th, 2024, in Geneva, and accomplish significant goals for 2024.

TCA Track Record (2015-2023)

Between 2015 and the launch of the Harnessing AI Risk Initiative in March 2023, TCA successfully advanced the Trustless Computing Certification Body and SeevikNet Initiative to create an open intergovernmental organization aimed at radically increasing the actual and globally perceived trustworthiness of the most critical IT and socio-technical systems for digital diplomatic communications and for the control of critical infrastructure such as frontier AIs - based on the Trustless Computing Paradigms. We engaged extensively with over 15 states in bilateral and multilateral meetings and conferences we organized; we led 4 research initiatives with 32 industry experts and research R&D partners since 2015; we spun off TRUSTLESS.AI to build open technologies compliant to such certifications; and advanced it via 11 Free and Safe in Cyberspace conferences, held in three continents and three times in Geneva, and participated by over 100 distinguished speakers and advisors. The TCCB, under the name of IT Security Agency, is one of the three agencies of the IGO foreseen by the Initiative, to ensure trustworthiness of its "governance-support systems". Our work was published in Geneva Le Temps and The Yuan. 

Philanthropic Opportunities

We are seeking long-term philanthropic partners who are interested in accompanying us and the Coalition we are convening in this ambitious undertaking, providing financial support, contributing strategic support, or leveraging their networks.

Conclusion

Our Initiative and Coalition lead a strategic, timely, and expert-led effort to establish a democratic global governance model for AI and digital communications. Although ambitious, it directly confronts significant risks and aims to unlock substantial benefits for humanity, offering the potential for a profound positive impact for All.

Contact details:

  • Rufo Guerreschi, Founder and Executive Director
    Email: rufo@trustlesscomputing.org  — Tel: +393289376075

  • Marta Jastrzębska, Director of Fundraising and  Partnerships
    Email: marta@trustlesscomputing.org — Tel: +447502213612

  • Website: www.trustlesscomputing.org

More information

Call to Action

If you agree with our Initiative, please review opportunities to join the Coalition for a Baruch Plan for AI - the main vehicle to advance HARI since September 10th, 2024 - as an NGO, expert of volunteer, as a state, or as a donor.