Do companies need to address the threats posed by deepfakes?

As threat actors continue to make increasing use of deepfake technology, Europol warns that companies need to take more precautions

A global issue people are facing currently is disinformation, where false information is being spread with the intention to deceive. Today, threat actors are using disinformation campaigns and deepfake content to misinform the public about events, politics, and elections, contribute to fraud, and manipulate shareholders in a corporate context.

Many organisations have now begun to see deepfakes as an even bigger potential risk than identity theft (for which deepfakes can also be used), especially now that most interactions have moved online since the COVID-19 pandemic. 

The growing availability of disinformation and deepfakes will have a profound impact on the way people perceive authority and information media. With the increasing volume of deepfakes, trust in authorities and official facts is undermined.  

What is a deepfake? 

Deepfake technology uses artificial intelligence (AI) for audio and audio-visual content. Deepfake technology can produce content that convincingly shows people saying or doing things they never did, or create personas that never existed in the first place

Advances in machine learning and AI will continue enhancing the capabilities of the software used to create deepfakes.


Creating new security measure to contend with deepfakes 

Deepfake technology is set to be used extensively in organised crime over the coming years, according to new research by Europol.

Facing Reality? Law enforcement and the challenge of deepfakes, the first published analysis of the Europol Innovation Lab’s Observatory function, warned that law enforcement agencies will need to enhance the skills and technologies at officers’ disposal to keep pace with criminals use of deepfakes.

The analysis highlighted how deepfakes are being used in three key areas: disinformation, non-consensual pornography and document fraud. It predicts such attacks will become increasingly realistic and dangerous as the technology improves in the coming years.

In addition, the report observed that deepfakes could negatively impact the legal process, for example, by artificially manipulating or generating media to prove or disprove someone’s guilt. 

To effectively deal with these kinds of threats, Europol said law enforcement agencies must develop new skills and technologies. These include manual detection, which involves looking for inconsistencies, and automated detection techniques, including deepfake detection software using artificial intelligence that is being developed by organizations such as Facebook and security firm McAfee.

Policymakers also need to develop more legislation to set guidelines and enforce compliance around the use of deepfakes, the report added.

 

Share

Featured Articles

Google Securing WFH with Zscaler and Netskope Partnership

Google has added the expertise of Zscaler and Netskope into its Workspace Security Alliance to bolster the security issues created from work from home

Why Have Cybersecurity Budgets Soared for TMT Companies?

A recent report by Moody's shows Telecommunications, Media, and Technology companies have ballooned their cybersecurity budgets

Mandiant's Analysis Unveils Cause of Snowflake Data Theft

Mandiant identified three key issues things that the customers affected by the data breach shared

OpenText: AI Main Driver of Growth for MSPs & MSSPs

Technology & AI

Genetec’s Paul Dodds Talks Protecting IoT from Cyber Attacks

Network Security

Tech Mahindra and Cisco Partner on Next-Gen AI Firewall

Cyber Security