Faced with risks, experts in artificial intelligence, ask to stop experiments

For the development of Artificial Intelligence, they require collaboration between developers, legislators and regulatory authorities

Elon Musk, AI experts and technology industry leaders signed an open letter calling for a six-month pause on the development of AI systems that exceed OpenAI’s GPT-4, due to potential risks to the society and humanity in general.

In addition to Elon Musk, other titans of the world of technology and AI added their signatures to the letter. These include Stability AI CEO Emad Mostaque; DeepMind researchers and AI pioneers Stuart Russell and Yoshua Bengio. Apple co-founder Steve Wozniak also added his signature to the open letter. However, OpenAI CEO Sam Altman has not signed off on the open letter, according to a Future of Life spokesperson. The paper highlights the potential disruptions in politics and the economy caused by competing AI systems among humans. It also requires collaboration between developers, legislators, and regulatory authorities.

“OpenAI’s recent statement regarding artificial general intelligence states that ‘At some point, it may be important to get independent review before starting to train future systems, and for more advanced efforts to agree to limit the rate of growth of artificial general intelligence.’ computation used to create new models’. We agree. That point is now.

“Therefore, we call on all AI labs to immediately pause for at least 6 months the training of AI systems more powerful than GPT-4. This pause must be public and verifiable, and include all key stakeholders. If such a pause cannot be enacted quickly, governments should step in and institute a moratorium. “AI labs and independent experts should take advantage of this pause to jointly develop and implement a set of shared security protocols for advanced AI design and development that are rigorously audited and monitored by independent third-party experts. These protocols must ensure that the systems that adhere to them are secure beyond a reasonable doubt. This does not mean a pause in AI development in general, simply a step back from the perilous race to ever larger unpredictable black box models with emerging capabilities,” the letter read.