Patents Assigned to PINDROP SECURITY, INC.
-
Publication number: 20250124945Abstract: Embodiments described herein provide for a machine-learning architecture for modeling quality measures for enrollment signals. Modeling these enrollment signals enables the machine-learning architecture to identify deviations from expected or ideal enrollment signal in future test phase calls. These differences can be used to generate quality measures for the various audio descriptors or characteristics of audio signals. The quality measures can then be fused at the score-level with the speaker recognition's embedding comparisons for verifying the speaker. Fusing the quality measures with the similarity scoring essentially calibrates the speaker recognition's outputs based on the realities of what is actually expected for the enrolled caller and what was actually observed for the current inbound caller.Type: ApplicationFiled: December 20, 2024Publication date: April 17, 2025Applicant: PINDROP SECURITY, INC.Inventors: Hrishikesh RAO, Kedar PHATAK, Elie KHOURY
-
Publication number: 20250029614Abstract: Disclosed are systems and methods including software processes executed by a server for obtaining, by a computer, an audio signal including synthetic speech, extracting, by the computer, metadata from a watermark of the audio signal by applying a set of keys associated with a plurality of text-to-speech (TTS) services to the audio signal, the metadata indicating an origin of the synthetic speech in the audio signal, and generating, by the computer, based on the extracted metadata, a notification indicating that the audio signal includes the synthetic speech.Type: ApplicationFiled: July 18, 2024Publication date: January 23, 2025Applicant: PINDROP SECURITY, INC.Inventors: David LOONEY, Nikolay GAUBITCH, Elie KHOURY
-
Publication number: 20240363099Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: November 9, 2023Publication date: October 31, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair Altaf, Sai Pradeep Peri, Lakshay Phatela, Payas Gupta, Yitao Sun, Svetlana Afanaseva, Kailash Patil, Elie Khoury, Bradley Magnetta, Vijay Balasubramaniyan, Tianxiang Chen
-
Publication number: 20240355323Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: November 9, 2023Publication date: October 24, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair Altaf, Sai Pradeep PERI, Lakshay PHATELA, Payas GUPTA, Yitao SUN, Svetlane AFANASEVA, Kailash PATIL, Elie KHOURY, Bradley MAGNETTA, Vijay BALASUBRAMANIYAN, Tianxiang CHEN
-
Publication number: 20240355336Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: February 12, 2024Publication date: October 24, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair ALTAF, Sai Pradeep PERI, Lakshay PHATELA, Payas GUPTA, Yitao SUN, Svetlana AFANASEVA, Kailash PATIL, Elie KHOURY, Bradley MAGNETTA, Vijay BALASUBRAMANIYAN, Tianxiang CHEN
-
Publication number: 20240355337Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: November 9, 2023Publication date: October 24, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair Altaf, Sai Pradeep Peri, Lakshay Phatela, Payas Gupta, Yitao Sun, Svetlana Afanaseva, Kailash Patil, Elie Khoury, Bradley Magnetta, Vijay Balasubramaniyan, Tianxiang Chen
-
Publication number: 20240355334Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: November 9, 2023Publication date: October 24, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair Altaf, Sai Pradeep PERI, Lakshay PHATELA, Payas GUPTA, Yitao SUN, Svetlana AFANASEVA, Kailash PATIL, Elie KHOURY, Bradley MAGNETTA, Vijay BALASUBRAMANIYAN, Tianxiang CHEN
-
Publication number: 20240355319Abstract: Disclosed are systems and methods including software processes executed by a server that detect audio-based synthetic speech (“deepfakes”) in a call conversation. The server applies an NLP engine to transcribe call audio and analyze the text for anomalous patterns to detect synthetic speech. Additionally or alternatively, the server executes a voice “liveness” detection system for detecting machine speech, such as synthetic speech or replayed speech. The system performs phrase repetition detection, background change detection, and passive voice liveness detection in call audio signals to detect liveness of a speech utterance. An automated model update module allows the liveness detection model to adapt to new types of presentation attacks, based on the human provided feedback.Type: ApplicationFiled: November 9, 2023Publication date: October 24, 2024Applicant: PINDROP SECURITY, INC.Inventors: Umair Altaf, Sai Pradeep Peri, Lakshay Phatela, Payas Gupta, Yitao Sun, Svetlana Afanaseva, Kailash Patil, Elie Khoury, Bradley Magnetta, Vijay Balasubramaniyan, Tianxiang Chen
-
Publication number: 20240311474Abstract: Embodiments include a computing device that executes software routines and/or one or more machine-learning architectures including obtaining training audio signals having corresponding training impulse responses associated with reverberation degradation, training a machine-learning model of a presentation attack detection engine to generate one or more acoustic parameters by executing the presentation attack detection engine using the training impulse responses of the training audio signals and a loss function, obtaining an audio signal having an acoustic impulse response associated with reverberation degradation caused by one or more rooms, generating the one or more acoustic parameters for the audio signal by executing the machine-learning model using the audio signal as input, and generating an attack score for the audio signal based upon the one or more parameters generated by the machine-learning model.Type: ApplicationFiled: March 7, 2024Publication date: September 19, 2024Applicant: PINDROP SECURITY, INC.Inventors: Nikolay Gaubitch, David Looney
-
Publication number: 20240233709Abstract: Embodiments described herein provide for audio processing operations that evaluate characteristics of audio signals that are independent of the speaker's voice. A neural network architecture trains and applies discriminatory neural networks tasked with modeling and classifying speaker-independent characteristics. The task-specific models generate or extract feature vectors from input audio data based on the trained embedding extraction models. The embeddings from the task-specific models are concatenated to form a deep-phoneprint vector for the input audio signal. The DP vector is a low dimensional representation of the each of the speaker-independent characteristics of the audio signal and applied in various downstream operations.Type: ApplicationFiled: February 23, 2024Publication date: July 11, 2024Applicant: PINDROP SECURITY, INC.Inventors: Kedar PHATAK, Elie KHOURY
-
Publication number: 20240214490Abstract: Embodiments described herein provide for a fraud detection engine for detecting various types of fraud at a call center and a fraud importance engine for tailoring the fraud detection operations to relative importance of fraud events. Fraud importance engine determines which fraud events are comparative more important than others. The fraud detection engine comprises machine-learning models that consume contact data and fraud importance information for various anti-fraud processes. The fraud importance engine calculates importance scores for fraud events based on user-customized attributes, such as fraud-type or fraud activity. The fraud importance scores are used in various processes, such as model training, model selection, and selecting weights or hyper-parameters for the ML models, among others. The fraud detection engine uses the importance scores to prioritize fraud alerts for review.Type: ApplicationFiled: February 5, 2024Publication date: June 27, 2024Applicant: PINDROP SECURITY, INC.Inventors: Kedar PHATAK, Jayaram RAGHURAM
-
Publication number: 20240212688Abstract: Utterances of at least two speakers in a speech signal may be distinguished and the associated speaker identified by use of diarization together with automatic speech recognition of identifying words and phrases commonly in the speech signal. The diarization process clusters turns of the conversation while recognized special form phrases and entity names identify the speakers. A trained probabilistic model deduces which entity name(s) correspond to the clusters.Type: ApplicationFiled: February 8, 2024Publication date: June 27, 2024Applicant: PINDROP SECURITY, INC.Inventors: Ellie KHOURY, Matthew GARLAND
-
Publication number: 20240169040Abstract: Embodiments include a computing device that executes software routines and/or one or more machine-learning architectures including a neural network-based embedding extraction system that to produce an embedding vector representing a user's behavior's keypresses, where the system extracts the behaviorprint embedding vector using the keypress features that the system references later for authenticating users. Embodiments may extract and evaluate keypress features, such as keypress sequences, keypress pressure or volume, and temporal keypress features, such as the duration of keypresses and the interval between keypresses, among others. Some embodiments employ a deep neural network architecture that generates a behaviorprint embedding vector representation of the keypress duration and interval features that is used for enrollment and at inference time to authenticate users.Type: ApplicationFiled: November 20, 2023Publication date: May 23, 2024Applicant: PINDROP SECURITY, INC.Inventors: Hrishikesh RAO, Ricky CASAL, Elie KHOURY, Eric LORIMER, John CORNWELL, Kailash PATIL
-
Publication number: 20240153510Abstract: Embodiments described herein provide for systems and methods for implementing a neural network architecture for spoof detection in audio signals. The neural network architecture contains a layers defining embedding extractors that extract embeddings from input audio signals. Spoofprint embeddings are generated for particular system enrollees to detect attempts to spoof the enrollee's voice. Optionally, voiceprint embeddings are generated for the system enrollees to recognize the enrollee's voice. The voiceprints are extracted using features related to the enrollee's voice. The spoofprints are extracted using features related to features of how the enrollee speaks and other artifacts. The spoofprints facilitate detection of efforts to fool voice biometrics using synthesized speech (e.g., deepfakes) that spoof and emulate the enrollee's voice.Type: ApplicationFiled: December 22, 2023Publication date: May 9, 2024Applicant: PINDROP SECURITY, INC.Inventors: Tianxiang CHEN, Elie KHOURY
-
Publication number: 20240062753Abstract: Embodiments described herein provide for a computer that detects one or more keywords of interest using acoustic features, to detect or query commonalities across multiple fraud calls. Embodiments described herein may implement unsupervised keyword spotting (UKWS) or unsupervised word discovery (UWD) in order to identify commonalities across a set of calls, where both UKWS and UWD employ Gaussian Mixture Models (GMM) and one or more dynamic time-warping algorithms. A user may indicate a training exemplar or occurrence of call-specific information, referred to herein as “a named entity,” such as a person's name, an account number, account balance, or order number. The computer may perform a redaction process that computationally nullifies the import of the named entity in the modeling processes described herein.Type: ApplicationFiled: October 31, 2023Publication date: February 22, 2024Applicant: PINDROP SECURITY, INC.Inventor: Hrishikesh Rao
-
Patent number: 11810559Abstract: Embodiments described herein provide for a computer that detects one or more keywords of interest using acoustic features, to detect or query commonalities across multiple fraud calls. Embodiments described herein may implement unsupervised keyword spotting (UKWS) or unsupervised word discovery (UWD) in order to identify commonalities across a set of calls, where both UKWS and UWD employ Gaussian Mixture Models (GMM) and one or more dynamic time-warping algorithms. A user may indicate a training exemplar or occurrence of call-specific information, referred to herein as “a named entity,” such as a person's name, an account number, account balance, or order number. The computer may perform a redaction process that computationally nullifies the import of the named entity in the modeling processes described herein.Type: GrantFiled: June 6, 2022Date of Patent: November 7, 2023Assignee: PINDROP SECURITY, INC.Inventor: Hrishikesh Rao
-
Patent number: 11783839Abstract: Embodiments described herein provide for a voice biometrics system execute machine-learning architectures capable of passive, active, continuous, or static operations, or a combination thereof. Systems passively and/or continuously, in some cases in addition to actively and/or statically, enrolling speakers. The system may dynamically generate and update profiles corresponding to end-users who contact a call center. The system may determine a level of enrollment for the enrollee profiles that limits the types of functions that the user may access. The system may update the profiles as new contact events are received or based on certain temporal triggering conditions.Type: GrantFiled: September 30, 2021Date of Patent: October 10, 2023Assignee: PINDROP SECURITY, INC.Inventors: Payas Gupta, Terry Nelms, II
-
Patent number: 11756564Abstract: A computer may segment a noisy audio signal into audio frames and execute a deep neural network (DNN) to estimate an instantaneous function of clean speech spectrum and noisy audio spectrum in the audio frame. This instantaneous function may correspond to a ratio of an a-priori signal to noise ratio (SNR) and an a-posteriori SNR of the audio frame. The computer may add estimated instantaneous function to the original noisy audio frame to output an enhanced speech audio frame.Type: GrantFiled: June 14, 2019Date of Patent: September 12, 2023Assignee: PINDROP SECURITY, INC.Inventors: Ganesh Sivaraman, Elie Khoury
-
Publication number: 20230283711Abstract: A method of obtaining and automatically providing secure authentication information includes registering a client device over a data line, storing information and a changeable value for authentication in subsequent telephone-only transactions. In the subsequent transactions, a telephone call placed from the client device to an interactive voice response server is intercepted and modified to include dialing of a delay and at least a passcode, the passcode being based on the unique information and the changeable value, where the changeable value is updated for every call session. The interactive voice response server forwards the passcode and a client device identifier to an authentication function, which compares the received passcode to plural passcodes generated based on information and iterations of a value stored in correspondence with the client device identifier. Authentication is confirmed when a generated passcode matches the passcode from the client device.Type: ApplicationFiled: May 15, 2023Publication date: September 7, 2023Applicant: PINDROP SECURITY, INC.Inventor: Payas GUPTA
-
Publication number: 20230284016Abstract: Embodiments described herein provide for evaluating call metadata and certificates of inbound calls for authentication. The computer identifies a service provider indicated by the SPID and/or the ANI (or other identifier) of the metadata and identifies a service provider indicated by the SPID and/or ANI (or other identifier) of the certificate, then compares identities of the service providers and/or compares the data values associated with the service providers (e.g., SPIDs, ANIs). Based on this comparison, the computer determines whether the service provider that signed the certificate is first-party signer (e.g., carrier) for the ANI or a third-party signer that is signing certificates as the first-party signer for the ANI.Type: ApplicationFiled: March 3, 2023Publication date: September 7, 2023Applicant: PINDROP SECURITY, INC.Inventors: MohammedAli Merchant, Yitao Sun