Patents by Inventor Cagdas Bilen
Cagdas Bilen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20230317102Abstract: A method of detecting occurrences of a sound event in an audio signal comprising a sequence of frames of audio data, each frame corresponding to a respective time in the audio signal, the method comprising: for each frame in the sequence: determining, using the audio data of the frame, a first conditional probability value that a transition occurred from a sound event not having started to the sound event being ongoing, and a second conditional probability value that a transition occurred from a sound event being ongoing to the sound event having ended; and determining a marginal probability value that a sound event was ongoing at the time corresponding to the frame, the marginal probability value being determined using the first and second conditional probability values for the frame and a previously determined marginal probability value that a sound event was ongoing at a time corresponding to a frame preceding the frame in the sequence.Type: ApplicationFiled: April 5, 2022Publication date: October 5, 2023Inventors: Cagdas Bilen, Giacomo Ferroni, Juan Azcaretta Ortiz, Francesco Tuveri, Sacha Krstulovic
-
Patent number: 11587556Abstract: A method for recognising at least one of a non-verbal sound event and a scene in an audio signal comprising a sequence of frames of audio data, the method comprising: for each frame of the sequence: processing the frame of audio data to extract multiple acoustic features for the frame of audio data; and classifying the acoustic features to classify the frame by determining, for each of a set of sound classes, a score that the frame represents the sound class; processing the sound class scores for multiple frames of the sequence of frames to generate, for each frame, a sound class decision for each frame; and processing the sound class decisions for the sequence of frames to recognise the at least one of a non-verbal sound event and a scene.Type: GrantFiled: October 7, 2019Date of Patent: February 21, 2023Assignee: META PLATFORMS TECHNOLOGIES, LLCInventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Juan Azcarreta Ortiz, Giacomo Ferroni, Arnoldas Jasonas, Francesco Tuveri
-
Patent number: 11380349Abstract: Verification of presence of a detected target is carried out following an initial presence determination on the basis of detected non-verbal sound.Type: GrantFiled: September 24, 2019Date of Patent: July 5, 2022Assignee: AUDIO ANALYTIC LTDInventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Patent number: 11250848Abstract: Control of navigation of a content item is carried out by detection of non-verbal sound events. On the basis of receipt of one or more non-verbal sound event reports, a navigation tool is provided with a corresponding sequence of navigation commands. The correspondence between navigation command sequences and non-verbal sound events is established through analysis or markup of the content item.Type: GrantFiled: September 24, 2019Date of Patent: February 15, 2022Assignee: AUDIO ANALYTIC LTDInventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Patent number: 11133020Abstract: A device or system is provided which is configured to detect one or more sound events and/or scenes associated with a predetermined context, and to provide an assistive output on fulfilment of that context.Type: GrantFiled: October 7, 2019Date of Patent: September 28, 2021Assignee: AUDIO ANALYTIC LTDInventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Amoldas Jasonas, Joe Patrick Lynas
-
Publication number: 20210104255Abstract: A device or system is provided which is configured to detect one or more sound events and/or scenes associa ted with a predetermined context, and to provide an assistive output on fulfilment of that context.Type: ApplicationFiled: October 7, 2019Publication date: April 8, 2021Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Publication number: 20210104230Abstract: A method for recognising at least one of a non-verbal sound event and a scene in an audio signal comprising a sequence of frames of audio data, the method comprising: for each frame of the sequence: processing the frame of audio data to extract multiple acoustic features for the frame of audio data; and classifying the acoustic features to classify the frame by determining, for each of a set of sound classes, a score that the frame represents the sound class; processing the sound class scores for multiple frames of the sequence of frames to generate, for each frame, a sound class decision for each frame; and processing the sound class decisions for the sequence of frames to recognise the at least one of a non-verbal sound event and a scene.Type: ApplicationFiled: October 7, 2019Publication date: April 8, 2021Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Juan Azcarreta Ortiz, Giacomo Ferroni, Amoldas Jasonas, Francesco Tuveri
-
Publication number: 20210097727Abstract: Sound detection and identification leads to responsiveness within an augmented reality environment. Information about an identified sound can be converted into a command for implementation by an augmented reality system, for display of a desired on-screen augmented reality effect.Type: ApplicationFiled: September 27, 2019Publication date: April 1, 2021Inventors: Chris James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Publication number: 20210090591Abstract: Verification of presence of a detected target is carried out following an initial presence determination on the basis of detected non-verbal sound.Type: ApplicationFiled: September 24, 2019Publication date: March 25, 2021Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Publication number: 20210090560Abstract: Control of navigation of a content item is carried out by detection of non-verbal sound events. On the basis of receipt of one or more non-verbal sound event reports, a navigation tool is provided with a corresponding sequence of navigation commands. The correspondence between navigation command sequences and non-verbal sound events is established through analysis or markup of the content item.Type: ApplicationFiled: September 24, 2019Publication date: March 25, 2021Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
-
Patent number: 10878840Abstract: A method for recognising at least one of a non-verbal sound event and a scene in an audio signal comprising a sequence of frames of audio data, the method comprising: for each frame of the sequence: receiving at least one sound class score, wherein each sound class score is representative of a degree of affiliation of the frame with a sound class of a plurality of sound classes; for a sound class score of the at least one sound class scores: determining a confidence that the sound class score is representative of a degree of affiliation of the frame with the sound class by processing a value for a property associated with the frame, wherein the value is processed using a learned model for the property; adjusting the sound class score for the frame based at least on the determined confidence.Type: GrantFiled: October 15, 2019Date of Patent: December 29, 2020Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Juan Azcarreta Ortiz, Giacomo Ferroni, Arnoldas Jasonas, Francesco Tuveri
-
Patent number: 10783434Abstract: A method of training a non-verbal sound class detection machine learning system, the non-verbal sound class detection machine learning system comprising a machine learning model configured to: receive data for each frame of a sequence of frames of audio data obtained from an audio signal; for each frame of the sequence of frames: process the data for multiple frames; and output data for at least one sound class score representative of a degree of affiliation of the frame with at least one sound class of a plurality of sound classes, wherein the plurality of sound classes comprises: one or more target sound classes; and a non-target sound class representative of an absence of each of the one or more target sound classes; wherein the method comprises: training the machine learning model using a loss function.Type: GrantFiled: October 7, 2019Date of Patent: September 22, 2020Assignee: AUDIO ANALYTIC LTDInventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Juan Azcarreta Ortiz, Giacomo Ferroni, Arnoldas Jasonas, Francesco Tuveri
-
Publication number: 20180341805Abstract: In a particular implementation, a codebook C can be used for quantizing a feature vector of a database image into a quantization index, and then a different codebook (B) can be used to approximate the feature vector based on the quantization index. The codebooks B and C can have different sizes. Before performing image search, a lookup table can be built offline to include distances between the feature vector for a query image and codevectors in codebook B to speed up the image search. Using triplet constraints wherein a first image and a second image are indicated as a matching pair and the first image and a third image as non-matching, the codebooks B and C can be trained for the task of image search. The present principles can be applied to regular vector quantization, product quantization, and residual quantization.Type: ApplicationFiled: November 4, 2016Publication date: November 29, 2018Inventors: Himalaya JAIN, Cagdas BILEN, Joaquin ZEPEDA SALVATIERRA, Patrick PEREZ
-
Publication number: 20180211672Abstract: A method for performing audio inpainting, wherein missing portions in an input audio signal are recovered and a recovered audio signal is obtained, comprises computing a Short-Time Fourier Transform (STFT) on portions of the input audio signal, computing conditional expectations of the source power spectra of the input audio signal, wherein estimated source power spectra P(f, n, j) are obtained and wherein the variance tensor V and complex Short-Time Fourier Transform (STFT) coefficients of the input audio signals are used, iteratively re-calculating the variance tensor V from the estimated power spectra P(f, n, j) and re-calculating updated estimated power spectra P(f, n, j), computing an array of STFT coefficients ? from the resulting variance tensor V according to ?(f, n, j)=E{S(f, n, j)|x, Is, IL, V}, and converting the array of STFT coefficients ? to the time domain, wherein coefficients {tilde over (s)}1, {tilde over (s)}2, . . . , {tilde over (s)}j of the recovered audio signal are obtained.Type: ApplicationFiled: April 6, 2016Publication date: July 26, 2018Applicant: Dolby International ABInventors: Cagdas BILEN, Alexey OZEROV, Patrick PEREZ
-
Publication number: 20180082693Abstract: A method for encoding multiple audio signals comprises random sampling and quantizing each of the multiple audio signals, and encoding the sampled and quantized multiple audio signals as side information that can be used for decoding and separating the multiple audio signals from a mixture of said multiple audio signals. A method for decoding a mixture of multiple audio signals comprises decoding and demultiplexing side information, the side information comprising quantized samples of each of the multiple audio signals, receiving or retrieving from any data source a mixture of said multiple audio signals, and generating multiple estimated audio signals that approximate said multiple audio signals, wherein said quantized samples of each of the multiple audio signals are used.Type: ApplicationFiled: March 10, 2016Publication date: March 22, 2018Inventors: Cagdas BILEN, Alexey OZEROV, Patrick PEREZ
-
Publication number: 20120106625Abstract: Adapting characteristics of a video stream. A video stream is received, at a video adaptation device, which employs a first reference frame. A determination is made, at a video adaptation device, to splice the video stream to employ a second reference frame in place of the first reference frame. The video stream is spliced, at a video adaptation device, to create a spliced video stream that employs the second reference frame.Type: ApplicationFiled: January 21, 2011Publication date: May 3, 2012Inventors: Wai-Tian TAN, Andrew J. Patti, Mitchell Trott, Cagdas Bilen