How friendly AI friend is

artbreeder image of Prototron AI
Leading figures in the field of artificial intelligence development, including Elon Musk and Apple co-founder Steve Wozniak, are proposing a pause in the development of AI technologies until reliable security protocols are developed and implemented. As a highly technological foundation, we believe that the most effective way to protect ourselves is to stay informed about what is happening.

The first thing we did after reading the open letter and some background information was to ask GPT 3.0 what it thinks about the threat to humanity from future versions of itself. Unfortunately, there is nothing new about what futurists were saying 20 years ago, what education specialists are saying, and what we all know: the threat of technology begins in the mind of the person trying to turn it into a tool to achieve their own, and sometimes, unfortunately, even military goals.

Yes, AI can be used for disinformation. Yes, today’s AI tools can violate your privacy – read the disclaimers “Developers can read your texts for AI development purposes”. The threat of more global security when AI can be used for cyber attacks. But what is currently being hyped in the media is the rise of inequality: if access to AI tools is limited to certain segments of society, this will further exacerbate the segregation of access to technology, reliable information, and as a result, to benefits. Hence the completely justified threat of job loss – automation, robotisation is already affecting the employment structure in industry (welders are replaced by welding robot operators, for example).

The released AI language model GPT 3.0 has drawn our attention to the fact that white-collar work can be done by machines much faster and sometimes even better, including written and oral translations, analytics and forecasting, content creation, information search and analysis, and optimization of many processes. Of course, it’s frightening to think about what will be left for us to do. Especially if AI becomes a black box for us, where it’s unclear what and under what conditions goes in.

The most contradictory aspect of AI technology is that its creators somehow choose the information that is fed to the machine, build the structure of evaluations, self-analysis, monitoring, select experts, on the basis of whose remarks further infrastructure is built. The popular language learning app Duolingvo already works with GPT 4.0 technology, and the government of Iceland has initiated a program to preserve the Icelandic language in cooperation with OpenAI based on GPT 4.0.

Returning to the open letter, it’s a perfect illustration of what happens to information as it moves from medium to medium. The letter calls for a halt to the use of more advanced technologies than GPT 4.0 until regulations are created for data selection, self-analysis, correction, and rules of AI usage. However, some specialists have already disputed the document, claiming they didn’t sign it, and some signature certificates are invalid. Signatures were collected with the argument that Elon Musk even signed it, so why don’t you? Mr. Musk, as we know, was one of the co-founders of OpenAI, refused external regulation for the AI used in Tesla cars, and sponsors the Institute, known for its black scenarios of the future.

It’s evident that the most effective way to protect ourselves is to stay informed about what’s happening.

Let me quote Yuval Noah Harari (who seemed to be among those who signed the open letter) on what he actually writes on his official channels: “The danger is that if we continue to invest too much in the development of AI and too little in the development of human consciousness, such a powerful technology can only serve to enhance human stupidity.”