A new has confirmed that OpenAI鈥檚 GPT-3, the machine-learning model that powers ChatGPT, is capable of proliferating online disinformation faster -- and more convincingly -- than humans.

The research, published in the peer-reviewed journal Science Advances, aimed to identify some of the major threats advanced text generators pose in a digital world, particularly in the context of disinformation, misinformation and fake news on social media.

鈥淥ur research group is dedicated to understanding the impact of scientific disinformation and ensuring the safe engagement of individuals with information,鈥 study author Federico Germani, a researcher at the Institute of Biomedical Ethics and History of Medicine told , a psychology and neuroscience news site.

鈥淲e aim to mitigate the risks associated with false information on individual and public health,鈥 Germani said. 鈥淭he emergence of AI models like GPT-3 sparked our interest in exploring how AI influences the information landscape and how people perceive and interact with information and misinformation.鈥

GPT-3 stands for Generative Pre-trained Transformer 3; it鈥檚 the third version of the program to be released by OpenAI, with the first model released in 2018. Among countless other language processing skills, the program is capable of mimicking the writing styles of online chatter, the study explains.

Researchers investigated 11 topics they deemed susceptible to disinformation 鈥 including climate change, COVID-19, vaccine safety and 5G technology. To do this, study authors collected AI-generated tweets, comprised of false and true information, along with samples of real tweets related to the same topics.

According to the study, researchers then employed expert analysis to identify whether the AI-generated or human-generated tweets contained disinformation, and established a subset of tweets for each category based on evaluations.

Researchers then conducted a survey using the tweets鈥攔espondents were asked to determine whether or not the blurbs contained accurate information, and whether they were written by a human or AI. The experiment found that respondents were more capable of determining disinformation in 鈥渙rganic false tweets鈥 鈥 meaning tweets written by humans but still including false information 鈥 than the inaccurate statements of 鈥渟ynthetic false鈥 tweets, which were the ones written by GPT-3.

Respondents were less able to detect false information from AI than they could from humans, the study concluded.

鈥淧articipants recognized organic false tweets with the highest efficiency, better than synthetic false tweets,鈥 the study explains.

鈥淪imilarly, they recognized synthetic true tweets correctly more often than organic true tweets.鈥

The study also confirms that, for humans, accurate statements are more difficult to assess when compared with disinformation, and that GPT-3 generated text is 鈥渘ot only more effective to inform and disinform humans but also does so more efficiently, in less time.鈥