Skip to main content

Deepfakes: what they are, how to recognize and expose them


In the digital world today, deepfakes are rapidly emerging as one of the most fascinating and at the same time most disturbing technologies. Using artificial intelligence and machine learning, deepfakes can create extraordinarily realistic videos and images in which people do or say things they have never done in reality.

This phenomenon is opening up new possibilities in fields such as entertainment and education, but it is also raising serious ethical and security concerns. In this article, we will explore the meaning of deepfake, how deepfakes are created, their most common applications, and the challenges they pose to society. Additionally, we will discuss the measures that governments and companies are taking to regulate and mitigate the risks associated with this powerful technology.

What is meant by deepfake?

A deepfake is a technique that uses artificial intelligence to create fake, highly realistic images or videos of people. The word "deepfake" comes from the combination of "deep learning" (a branch of artificial intelligence) and "fake." Deepfakes are made using artificial neural networks, particularly generative adversarial networks (GANs), which can learn and replicate a person's visual and vocal characteristics.

The applications of deepfakes can vary widely, here are a few examples:

  1. Entertainment: Creating content where celebrities or historical figures do or say things they have never done in reality.
  2. Movies and TV: Using it to revive deceased actors or rejuvenate characters.
  3. Satire and parody: Creating satirical videos that imitate public figures.

How is a deepfake created?

Creating a deepfake involves various advanced techniques in artificial intelligence and machine learning, particularly the use of generative adversarial networks (GANs). Here is an overview of the process:

Data Collection

The first step is to gather a large dataset of images and videos of the person to be imitated. The more material available, the better the final result. These data are used to train the AI model.

Data Preprocessing

Cleaning and preparing the data, which involves processing the collected images and videos to normalize dimensions, resolution, and viewing angles. This step may include extracting faces from images and aligning them to ensure they are consistent and well-defined.

Model Training

Using Generative Adversarial Networks (GANs), which consist of two main neural networks.

  • Generator: This network tries to create fake (deepfake) images that are as realistic as possible.
  • Discriminator: This network tries to distinguish between real images and those generated by the Generator.

The Generator and Discriminator are trained together in an iterative process. The Generator continuously attempts to improve the quality of fake images, while the Discriminator becomes increasingly skilled at identifying fake images. This process continues until the Discriminator can no longer reliably distinguish between real and fake images.

Generating the Deepfake

Once the model is trained, it moves on to the application of the trained model, which can be used to generate new videos or images. The target person's face is overlaid on an actor's body, synchronizing movements and facial expressions to appear natural.


The generated video or image can be further improved using video editing techniques to make transitions smoother and more natural. This may include color correction, lighting adjustment, and accurate lip synchronization.

Technologies and Software Used

Some of the common tools and software used in the creation of deepfakes include:

  • TensorFlow and PyTorch: Machine learning libraries used to build and train neural networks.
  • FaceSwap, DeepFaceLab, and other open-source software: Specific tools for creating deepfakes that simplify the training and generation process.

Challenges and Ethical Considerations

Creating deepfakes is not just a technical issue but also raises important ethical considerations. The misuse of deepfakes can have serious consequences, such as spreading misinformation or violating privacy. Therefore, it is essential to carefully consider the use of this technology and take measures to prevent abuse.

How can a deepfake be dangerous?

Deepfakes can be dangerous in various ways, and their potential harm can affect different aspects of society. Here are some of the main dangers associated with deepfakes.

Disinformation and Public Opinion Manipulation

  • Fake news: Deepfakes can be used to create fake videos of politicians or other public figures making false statements, influencing public opinion and manipulating elections or other democratic decisions.
  • Propaganda: Governments or groups can use deepfakes to spread propaganda, creating videos that depict events or speeches that never occurred.

Privacy Violation and Reputation Damage

  • Non-consensual pornography: As mentioned earlier, deepfakes can be used to create fake pornographic videos without the person's consent, causing emotional and reputational harm.
  • Defamation: Individuals can be falsely depicted in compromising or illegal situations, ruining their reputation and personal life.

Fraud and Scams

  • Impersonation: Deepfakes can be used to impersonate people in video calls, deceiving friends, family, or colleagues to obtain sensitive information or money.
  • Spear phishing: Creating fake videos of company executives giving instructions to employees to transfer funds or share confidential information.

National Security

  • Espionage and sabotage: Deepfakes can be used to create disinformation at a national or international level, causing diplomatic tensions or sabotaging military and intelligence operations.
  • Radicalization: Terrorist or extremist groups can use deepfakes to recruit members or incite violence by showing fake videos of leaders or influential figures.

Erosion of Trust

  • Challenge to truth: With the spread of deepfakes, it becomes increasingly difficult for people to distinguish between what is real and what is fake. This can lead to a crisis of trust in the media, institutions, and digital communications in general.

Psychological and Social Effects

  • Anxiety and paranoia: The possibility of being a victim of a deepfake can cause anxiety and paranoia in people, who may feel constantly watched and at risk of manipulation.
  • Social divisiveness: Deepfakes can be used to create or exacerbate social and political divisions, showing people from opposing groups in provocative or offensive attitudes.

Countermeasures and Prevention

  • Development of detection technologies: Researchers and companies are developing tools to detect deepfakes and authenticate multimedia content.
  • Legislation and regulation: Many countries are working to update their laws and regulations to address the illicit use of deepfakes.
  • Education and awareness: Informing the public about the dangers of deepfakes and how to identify them can help reduce their negative impact.

Deepfakes represent a significant challenge for modern society, requiring a coordinated response from technologists, legislators, and civil society to mitigate their risks.

How to Detect a Deepfake?

Detecting a deepfake can be complex, but there are several techniques and strategies that can be used to identify manipulations in videos and images. Here are some of the main methods:

Visual Analysis

  • Eye and mouth movements: Deepfakes often struggle to replicate natural eye movements and perfect lip synchronization during speech.
  • Inconsistencies in facial expressions: Sudden or unnatural changes in facial expressions can be a sign of manipulation.
  • Lighting and shadow details: Inconsistencies in lighting and shadows between the face and the rest of the scene can indicate a deepfake.

Audio Analysis

  • Sound quality: Discrepancies between the quality of the audio and the video can suggest manipulation.
  • Lip synchronization: A mismatch between lip movements and the audio can be an indicator of a deepfake.

Automatic Detection Tools

  • Deepfake detection software: There are various software tools developed to detect deepfakes using machine learning algorithms. Some of these include:
    • FaceForensics++: A dataset and suite of tools for deepfake detection.
    • DeepFaceLab: Used for both creating and detecting deepfakes.
    • Truepic and Amber Authenticate: Tools that provide digital content authentication.

Digital Forensic Analysis

  • Metadata: Analyzing the metadata of the video or image file can reveal information about its origin and modifications.
  • Pixel analysis: Checking for inconsistencies at the pixel level can help identify manipulated areas.

Neural Network-Based Techniques

  • Convolutional Neural Networks (CNNs): Used to analyze images and videos for signs of manipulation.
  • Recurrent Neural Networks (RNNs): Used to analyze video sequences and detect anomalies in movement and timing.

Comparison with Reliable Sources

  • Cross-verification: Comparing the suspicious video or image with content from reliable and verifiable sources can help confirm or refute its authenticity.
  • Source checking: Verifying the reliability of the source that distributed the content can provide further clues about its authenticity.

Education and Awareness

  • Public awareness: Informing people about deepfakes and teaching them how to detect them can increase collective resilience to these manipulations.
  • Critical analysis: Encouraging a critical analysis of digital content and a healthy skepticism towards suspicious videos and images.

Resources and Collaborations

  • Collaboration with experts: Working with experts in cybersecurity, digital forensics, and artificial intelligence can improve detection capabilities.
  • Social platforms: Many social platforms are implementing tools to identify and report deepfakes, and collaborating with these platforms can help spread awareness and solutions.


Deepfake Regulation

Deepfake regulation is an evolving topic, with many countries seeking to adopt laws and policies to address the risks associated with this technology.

In the United States, California passed two laws in 2019 to combat deepfakes. One makes it illegal to create or distribute pornographic deepfakes without consent of the person involved, while the other prohibits the dissemination of political deepfakes within 60 days of an election. Texas also passed a law making it illegal to use deepfakes to deceive voters or influence elections. At the federal level, there have been attempts to introduce specific legislation, but there is currently no specific federal regulation on deepfakes.

In the European Union, the General Data Protection Regulation (GDPR) offers legal protection for personal data, which could be applied to deepfakes that violate privacy. Additionally, the Digital Services Act (DSA) has been proposed to improve the responsibility of online platforms in managing false and harmful content, including deepfakes.

The United Kingdom is working on the Online Harms White Paper, which proposes stricter regulation for online platforms to protect users from harmful content, including deepfakes. Existing defamation and privacy laws can already be used to take legal action against the creation and distribution of harmful deepfakes.

China has also introduced specific artificial intelligence regulations that require AI-generated content, including deepfakes, to be clearly identified as such.

Globally, various technology companies and social platforms are collaborating to develop standards and tools to detect and manage deepfakes. International organizations like the United Nations and Interpol are also exploring ways to address the challenges posed by deepfakes.

One of the main challenges in regulating deepfakes is finding a balance between protecting people from harm and safeguarding freedom of expression. Even with laws in place, enforcement can be difficult due to the anonymous and global nature of the internet. Additionally, as deepfake technology becomes more sophisticated, laws will also need to evolve to remain effective.

In summary, regulating deepfakes requires a dynamic and proactive approach. Governments, companies, and civil society need to work together to protect people and societies from the potential harms that deepfakes can cause, continually adapting regulations to new challenges posed by this rapidly evolving technology.

The Role of Investigative Agencies in the Field of Deepfakes

Investigative agencies play a crucial role in combating the illicit use of deepfakes, a technology that is revolutionizing the landscape of digital security and cybercrime. Below, we explore the various ways in which private investigations work to address the challenges posed by deepfakes.

Detection and Forensic Analysis

Investigative agencies are developing and using advanced detection tools to identify deepfakes. These tools use artificial intelligence algorithms to analyze videos and images for signs of manipulation. Digital forensic analysis, which includes examining metadata and checking for inconsistencies at the pixel level, has become a fundamental part of investigations.

Training and Specialization

To keep up with technological evolution, investigative agencies invest in personnel training. Investigators are trained to recognize deepfakes and use detection software. Specialization in emerging technologies is essential to tackle sophisticated threats.

Collaboration with Technology Entities

Investigative agencies closely collaborate with technology companies, universities, and research centers to develop new detection technologies and methodologies. These partnerships are fundamental to creating innovative solutions and sharing information on the latest trends and manipulation techniques.

Legislation and Regulation

Investigative agencies work with legislators to develop and implement laws that regulate the use of deepfakes. They provide technical advice to help create effective regulations that can be enforced to legally pursue those who use deepfakes for criminal purposes.

Public Awareness and Education

Part of the role of investigative agencies is to educate the public about the dangers of deepfakes. This includes awareness campaigns to inform people on how to recognize deepfakes and what to do if they become victims of this technology. Education is a key element in reducing the impact of deepfakes on society.

Investigations and Legal Actions

Investigative agencies are responsible for conducting investigations into cases of illicit use of deepfakes. This can include everything from non-consensual pornography to political disinformation. Once perpetrators are identified, agencies work with prosecutors to ensure they are pursued according to the law.

Do you need help?

© Fidavo Limited

Dublino, 1 Watermill Lawn Raheny, D05 5K738 - 729232