Skip to content

D3V1L1805/FauxShield

Repository files navigation

Deepfake Detection: A Multimodal Approach

Overview This project focuses on detecting deepfakes using a comprehensive approach that combines video, audio, and image data. While many existing solutions focus on a single method, our model integrates all three to enhance detection accuracy and robustness. The performance metrics achieved are impressive, with video detection performing well, image detection achieving 90.85% accuracy, and audio detection reaching 99.05% accuracy.

Features Multimodal Detection: Integrates video, audio, and image data for deepfake detection.

High Accuracy:

Video detection: High performance

Image detection: 90.85% accuracy

Audio detection: 99.05% accuracy

Comprehensive Analysis: Utilizes multiple data types to improve the reliability of deepfake detection.

Getting Started

Prerequisites To run this project, you will need:

Python 3.7

TensorFlow or PyTorch

OpenCV

Librosa (for audio processing)

MTCNN (for face detection)

Other necessary libraries specified in requirements.txt

Navigate to the project directory:

bash Copy code cd deepfake-detection-multimodal Install the required libraries: bash Copy code pip install -r requirements.txt Usage Data Preparation: Ensure you have your video, audio, and image data organized. Follow the instructions in data_preparation.md for guidance on how to prepare your datasets.

Training: To train the model, run: bash Copy code python train.py --data_path /path/to/your/data Evaluation: To evaluate the model on your test data, use: bash Copy code python evaluate.py --model_path /path/to/your/model --data_path /path/to/your/test/data Prediction: For predicting whether a given input is a deepfake, run: bash Copy code python predict.py --input_path /path/to/your/input --model_path /path/to/your/model Results Video Detection: Our video detection model has shown high performance, making it reliable for identifying deepfakes in video content. Image Detection: Achieved an accuracy of 90.85%, proving its effectiveness in detecting fake images. Audio Detection: Reached an impressive accuracy of 99.05%, highlighting its strength in identifying manipulated audio. Model Architecture The project utilizes a combination of convolutional neural networks (CNNs) for image and video data and recurrent neural networks (RNNs) for audio data. The models are fine-tuned using transfer learning to leverage pre-trained weights and improve accuracy.

Contribution

We welcome contributions to enhance the project. Please fork the repository and create a pull request with your changes. Ensure that your contributions are well-documented and include relevant test cases.

License

This project is licensed under the MIT License. See the LICENSE file for details.

Acknowledgments

We would like to thank the open-source community and the creators of the pre-trained models used in this project.

Contact

For any questions or feedback, please contact Debojyoti Bhuinya, Subhamay Ganguly, Akash Das, Gargi Chakraborty.