Securing Our Digital Future: A CERN for Open Source large-scale AI Research and its Safety

Petīcija ir adresēta
EU, USA, UK, Canada, Australia
3 500
35% sasniedza 10 000 kolekcijas mērķim
  1. Sākās 2023.29.3
  2. Kolekcija vēl nav > 2 nedēļas
  3. Iesniegšana
  4. Dialogs ar saņēmēju
  5. Lēmums

Es piekrītu, ka mani dati tiks saglabāti . Jūs izlemjat, kurš var redzēt jūsu atbalstu. varu jebkurā laikā atsaukt šo piekrišanu .

Join us in our urgent mission to democratize AI research by establishing an international, publicly funded supercomputing facility equipped with 100,000 state-of-the-art AI accelerators to train open source foundation models. This monumental initiative will secure our technological independence, empower global innovation, and ensure safety, while safeguarding our democratic principles for generations to come.

In an era of unparalleled technological advancements, humanity stands on the precipice of a new epoch characterized by the profound influence of artificial intelligence (AI) and its foundational models, such as GPT-4. The potential applications of these technologies are vast, spanning scientific research, education, governance, and small and medium-sized enterprises. To harness their full potential as tools for societal betterment, it is vital to democratize research on and access to them, lest we face severe repercussions for our collective future.

Increasingly, we are witnessing the emergence of a system wherein educational institutions, government agencies, and entire nations become dependent on a select few large corporations that operate with little transparency or public accountability. To secure our society's technological independence, foster innovation, and safeguard the democratic principles that underpin our way of life, we must act now.

We call upon the global community, particularly the European Union, the United States, the United Kingdom, Canada, and Australia, to collaborate on a monumental initiative: the establishment of an international, publicly funded, open-source supercomputing research facility. This facility, analogous to the CERN project in scale and impact, should house a diverse array of machines equipped with at least 100,000 high-performance state-of-the-art accelerators (GPUs or ASICs), operated by experts from the machine learning and supercomputing research community and overseen by democratically elected institutions in the participating nations.

This ambitious endeavor will provide a platform for researchers and institutions worldwide to access and refine advanced AI models, such as GPT-4, harnessing their capabilities for the greater good. By making these models open source and incorporating multimodal data (audio, video, text, and program code), we can significantly enrich academic research, enhance transparency, and ensure data security. Furthermore, granting researchers access to the underlying training data will enable them to understand precisely what these models learn and how they function, an impossibility when restricted by APIs.

Additionally, the open-source nature of this project will promote safety and security research, allowing potential risks to be identified and addressed more rapidly and transparently by the academic community and open-source enthusiasts. This is a vital step in ensuring the safety and reliability of AI technologies as they become increasingly integrated into our lives.

The proposed facility should feature AI Safety research labs with well-defined security levels, akin to those used in biological research labs, where high-risk developments can be conducted by internationally renowned experts in the field, backed by regulations from democratic institutions. The results of such safety research should be transparent and available for the research community and society at large. These AI Safety research labs should be capable of designing timely countermeasures by studying developments that, according to broad scientific consensus, would predictably have a significant negative impact on our societies.

Economically, this initiative will bring substantial benefits to small and medium-sized companies worldwide.

By providing access to large foundation models, businesses can fine-tune these models for their specific use cases while retaining full control over the weights and data. This approach will also appeal to government institutions seeking transparency and control over AI applications in their operations.


The importance of this endeavor cannot be overstated. We must act swiftly to secure the independence of academia and government institutions from the technological monopoly of large corporations such as Microsoft, OpenAI, and Google. Technologies like GPT-4 are too powerful and significant to be exclusively controlled by a select few.

In a world where machine learning expertise and resources for AI development become increasingly concentrated in large corporations, it is imperative that smaller enterprises, academic institutions, municipal administrations, and social organizations, as well as nation-states, assert their autonomy and refrain from relying solely on the benevolence of these powerful entities that are often driven by short-term profit interests and act without properly taking democratic institutions into their decision-making loop. We must take immediate and decisive action to secure the technological independence of our society, nurturing innovation while ensuring the safety of these developments and protecting the democratic principles that form the foundation of our way of life.

The recent proposition of decelerating AI research as a means to ensure safety and progress presents a misguided approach that might be detrimental to both objectives. It could create a breeding ground for obscure and potentially malicious corporate or state actors to make advancements in the dark while simultaneously curtailing the public research community's ability to scrutinize the safety aspects of advanced AI systems thoroughly. Rather than impeding the momentum of AI development and shifting its development into underground areas, a more judicious and efficacious approach would be to foster a better-organized, transparent, safety-aware, and collaborative research environment. The establishment of transparent open-source AI safety labs tied to the international large-scale AI research facility as described above, which employ eligible AI safety experts, have corresponding publicly funded compute resources, and act according to regulations issued by democratic institutions, will cover the safety aspect without dampening progress. By embracing this cooperative framework, we can simultaneously ensure progress and the responsible development of AI technology, safeguarding the well-being of our society and the integrity of democratic values.

We urge you to join us in this crucial campaign. Sign this petition and make your voice heard. Our collective digital future, the autonomy of our academic research, and the equilibrium of our global economy depend on our ability to act quickly and decisively.

Together, we can build a future where advanced AI technologies are accessible to all, and where innovation and progress are not constrained by the boundaries of a few powerful corporations. Let us seize this opportunity and build a brighter future for generations to come.

Paldies par atbalstu , LAION e.V. no Hamburg
Jautājums iniciatoram



AI could be one of the biggest technological advancements of our time, allowing us to solve various other problems facing humanity. Its misuse may also be one of the biggest threats to humanity. AI will be far more crucial to the development of humankind in the next few decades than particle physics. As such, large-scale research should be an international effort, on the scale of CERN or bigger.

Continuing AI capailities advancments is catastrophic primarily not because of the biases the models might have or jobs people might lose but because, as half of ML researchers believe, there's at least a 10% chance of AI-induced existential catastrophe. We have to ensure the goals of first highly capable AI align with human values, but

Kāpēc cilvēki paraksta

pirms 19 h.

I'm interested in making a career in AI and large dataset related field. Monopolization of these techs behind corporate closed doors would negatively affect my path and the technological progress as the whole

pirms 2 dienām

As an AI researcher myself, I believe that this technology should be available to everyone, not just a few select organizations

pirms 2 dienām

I wholeheartedly believe that AI, AGI, LLM's should be OOS/FOSS. Available to all. For the benefit of the many.

pirms 5 dienām

AI can be of great benefit to humankind, but also a big threat. It is therefore of utmost importance to create an open source AI framework with strict ethic guidelines which are (by the nature of open source) transparent to everyone.

Rīki petīcijas izplatīšanai.

Vai jums ir sava vietne, emuārs vai viss tīmekļa portāls? Kļūsti par šīs petīcijas aizstāvi un pavairotāju. Mums ir reklāmkarogi, logrīki un API (interfeiss), ko iegult jūsu lapās.

Parakstiet logrīku savai vietnei

API (interfeiss)

Parakstu skaits uz openPetition un, ja piemērojams, uz ārējām lapām.
HTTP metode
Atgriešanās formāts

Palīdziet stiprināt pilsoņu līdzdalību. Mēs vēlamies padarīt jūsu bažas dzirdamas un palikt neatkarīgiem.

Veiciniet tūlīt