Mittal et al., 2025 - Google Patents
PITCH: AI-assisted Tagging of Deepfake Audio Calls using Challenge-ResponseMittal et al., 2025
View PDF- Document ID
- 11667357583729333245
- Author
- Mittal G
- Jakobsson A
- Marshall K
- Hegde C
- Memon N
- Publication year
- Publication venue
- Proceedings of the 20th ACM Asia Conference on Computer and Communications Security
External Links
Snippet
The rise of AI voice-cloning technology, particularly audio Real-time Deepfakes (RTDFs), has intensified social engineering attacks by enabling real-time voice impersonation that bypasses conventional enrollment-based authentication. This technology represents an …
- 241000282414 Homo sapiens 0 abstract description 82
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/316—User authentication by observing the pattern of computer usage, e.g. typical user behaviour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/554—Detecting local intrusion or implementing counter-measures involving event detection and direct action
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00221—Acquiring or recognising human faces, facial parts, facial sketches, facial expressions
- G06K9/00288—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/70—Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer
- G06F21/82—Protecting input, output or interconnection devices
- G06F21/83—Protecting input, output or interconnection devices input devices, e.g. keyboards, mice or controllers thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00221—Acquiring or recognising human faces, facial parts, facial sketches, facial expressions
- G06K9/00268—Feature extraction; Face representation
- G06K9/00281—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/20—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/70—Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer
- G06F21/71—Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information
- G06F21/77—Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information in smart cards
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wenger et al. | " Hello, It's Me": Deep Learning-based Speech Synthesis Attacks in the Real World | |
US10276152B2 (en) | System and method for discriminating between speakers for authentication | |
Müller et al. | Human perception of audio deepfakes | |
Yu et al. | Antifake: Using adversarial audio to prevent unauthorized speech synthesis | |
Firc et al. | The dawn of a text-dependent society: deepfakes as a threat to speech verification systems | |
Han et al. | Uncovering human traits in determining real and spoofed audio: Insights from blind and sighted individuals | |
Munir et al. | Deepfake defense: Constructing and evaluating a specialized Urdu deepfake audio dataset | |
Shirvanian et al. | Voicefox: Leveraging inbuilt transcription to enhance the security of machine-human speaker verification against voice synthesis attacks | |
Watt et al. | Forensic phonetics and automatic speaker recognition: The complementarity of human-and machine-based forensic speaker comparison | |
Kang et al. | Anonymization of Voices in Spaces for Civic Dialogue: Measuring Impact on Empathy, Trust, and Feeling Heard | |
Wang et al. | From one stolen utterance: Assessing the risks of voice cloning in the aigc era | |
Shirvanian et al. | Short voice imitation man-in-the-middle attacks on Crypto Phones: Defeating humans and machines | |
Sharevski et al. | Blind and Low-Vision Individuals' Detection of Audio Deepfakes | |
Mallinson et al. | A place for (socio) linguistics in audio deepfake detection and discernment: Opportunities for convergence and interdisciplinary collaboration | |
CN115050390B (en) | Voice privacy protection method and device, electronic equipment and storage medium | |
Mittal et al. | PITCH: AI-assisted Tagging of Deepfake Audio Calls using Challenge-Response | |
Li et al. | AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models | |
Turner | Security and privacy in speaker recognition systems | |
Mittal | Media Integrity in Real-Time Communication Streams | |
Harrington et al. | Variability in the performance of automatic speaker recognition systems across modelling approaches | |
Ahmed | Trustworthy User-Machine Interactions | |
Zhang | Understanding and Securing Voice Assistant Applications | |
Arzbecker | Evaluating Levenshtein distance: Assessing perception of accented speech through low-and high-pass filtering | |
Steffens et al. | How does a credible voice sound? | |
Grachek | Understanding Acceptability Judgements: Grammatical Knowledge vs. Lexical Search |