Technology is also a double-edged sword. It can be used for upliftment and good, or, in the wrong hands, it can be used to spread dissension and division, such as ‘disinformation’ - false or misleading information (misinformation) spread deliberately to deceive.
Content is generated by the second and fed to us on digital platters to satisfy our rapacious appetites for being in the know. There is, consequently, an increasing need for measures to be put in place to help readers separate fact from fiction. This is especially pertinent when it comes to the fields of news, advertising, e-commerce and even politics.
Already, readers struggle to tell the difference between a genuine news item and an opinion piece, an authentic news site or a website made to look like one and that peddles half-truths or deliberate falsehoods. Add another layer to this, such as fabricated videos, for example, known as ‘deepfakes’ and it is no wonder there is growing concern over what could be termed next-generation information warfare.
Photo and image manipulation, as well as the generation of fake news, are no longer ‘news’. However, the deep learning capabilities of artificial intelligence (AI) and machine learning (ML) tools, is making it easier and faster, to create masterpieces of visual deception. This power can, of course, be abused, to create and spread large scale fear and confusion at a citizen and public level to the extent where it can affect national security. In today’s context, it is screen warfare that wins the hearts and minds of easily swayed people.
Whereas cybersecurity is the use of technologies, systems and processes mainly aimed at the protection of data, we need to go one step further by looking at the source material itself - before it is transmitted. We have to build robust forensic cybersecurity defences, to help ensure that digital content can be source verified.
By Craig Lebrau, Issued by Lebrau Press 24 Feb 2020
Cryptographic identifiers are also essential to assure the authenticity of news, marketing and advertising engines, to safeguard consumers from fake campaigns, but also increasingly, to help them distinguish whether external influence is getting in the way – such as foreign powers interfering in domestic issues for example.
Worryingly, studies are showing that many learners can’t distinguish between sponsored content and news stories on social media and e-commerce platforms. Even more concerning, is the growing need to authenticate some of the material used in the learning environment itself.
We, therefore, need to bolster media literacy in schools and tertiary institutions, with workable and scalable solutions that can authenticate materials and resources. Additionally, there is a growing need to develop the appropriate tools to train students themselves on how to evaluate trustworthy media on social media and e-commerce platforms.
Cybersecurity is not a nice to have or confined to businesses and governments, it has become an essential life tool for everyone, whether connected to the digital ecosystem or not. Faced with the post-Covid era of remote working, e-learning, e-commerce, slowing economic growth and even dissatisfied corporate boards, overwhelmed security systems, and the constant pressure and need to get ahead, we can expect information warfare to heat up, and therefore, the need for deep disinformation defences.
Of course, this is all semantics if the will for content producers to verify their sources before posting their content is not there.