AVR Logo

AI Voice Recognition Final Project

About Us

Final project - Our final project is an AI system designed to detect whether a voice file is genuine or generated using deepfake technology (Ai Voice Recognize - AVR). Its purpose is to provide a safety layer to protect people from scams and to verify using voice authentication. We utilize Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) networks in its development, along with the use of Mel Spectrogram for audio signal processing.

Guy Ben Ari - Graduated Software Engineering student from Afeka College of Engineering. Passionate about technology and innovation, I contributed to this project with a focus on production development.

Ynon Friedman - Graduated Software Engineering student from Afeka College of Engineering. With a keen interest in artificial intelligence and machine learning, I was primarily involved in implementing and optimizing the AI models for voice recognition.

Revital Marom Elgrabli - Our guide in the process, Revital is an experienced mentor who has provided support and guidance throughout the development of this project.

Afeka College of Engineering - Afeka is a leading academic institution located in Tel Aviv, Israel, specializing in engineering and technology. Afeka supplied to us a virtual machine to develop our training pipeline, with the use of CUDA.

FAQ

What is Artificial Intelligence (AI)?

Artificial Intelligence, or AI, refers to the simulation of human intelligence processes by machines, particularly computer systems. While popular culture often depicts AI as androids like C3PO or advanced spaceship computers, modern AI is primarily based on statistical algorithms that utilize mathematical models to analyze data and make predictions. AI technology has various applications, including natural language processing (such as chatbots), image generation (like DALL-E), and voice and video synthesis.

What is Deepfake?

Deepfake refers to the use of artificial intelligence techniques, particularly deep learning algorithms, to create realistic fake content, typically videos or images. In deepfake videos, a person's likeness and voice can be manipulated to make it appear as though they are saying or doing something they never did in reality. This technology raises concerns about its potential misuse for spreading misinformation, defamation, and other malicious purposes.

What is a Spoof?

A spoof is a type of fraudulent activity where someone impersonates another individual or entity, usually using their voice or image. In the context of voice recognition, a spoof attack involves obtaining enough samples of a person's voice and then using AI to analyze and create a profile for that voice. This allows the attacker to generate synthetic audio that mimics the target's voice, enabling them to manipulate voice-based authentication systems or create fake audio recordings. Addressing the threat of spoof attacks is a crucial issue in voice recognition technology, and our project aims to tackle this challenge by developing robust AI algorithms for voice authentication and verification.

How does Voice Recognition Work?

Voice recognition, also known as speech recognition, is a technology that enables computers to interpret and understand human speech. It involves converting spoken words into text or commands that a computer can process. Voice recognition systems use a combination of techniques, including pattern recognition, natural language processing, and machine learning, to analyze audio input and identify the words spoken by the user. These systems can then perform various tasks based on the recognized speech, such as transcribing spoken words into text, controlling devices through voice commands, or providing interactive responses.

What are the Applications of Voice Recognition?

Voice recognition technology has numerous applications across various industries and sectors. Some common examples include:

How Secure is Voice Recognition?

Voice recognition technology offers a convenient and user-friendly way to interact with devices and systems, but its security depends on various factors. While voice biometrics can provide robust authentication and verification mechanisms, they are not immune to spoofing attacks or other forms of manipulation. Developers must implement additional security measures, such as multi-factor authentication, continuous authentication, and anti-spoofing techniques, to enhance the security of voice recognition systems and protect against fraudulent activities.

Downloads

Provide links to download the project from GitHub.

Download from GitHub

Contact Us

If you have any questions, feedback, or inquiries, please feel free to reach out to us using the contact information below:

You can also fill out the form below, and we'll get back to you as soon as possible:





Human VS Machine

Here are some audio samples. Can you tell which are fake?

Voice 1

Voice 2

Voice 3

Voice 4

Voice 5

Voice 6