Patents by Inventor Justin Salamon
Justin Salamon has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240161735Abstract: Embodiments are disclosed for performing a filler word detection process on input audio by a media editing system using trained neural networks. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving an input including an audio sequence, analyzing the audio sequence to determine filler word candidates, classifying, by a filler word classification model, each filler word candidate of the filler word candidates into one of a set of categories, and generating an output audio sequence, the output audio sequence including an identification of a subset of the filler word candidates in a filler words category of the set of categories as identified filler words.Type: ApplicationFiled: November 15, 2022Publication date: May 16, 2024Applicant: Adobe Inc.Inventors: Justin SALAMON, Juan-Pablo CACERES CHOMALI, Ge ZHU, Nicholas J. BRYAN
-
Patent number: 11887371Abstract: Embodiments are directed to a thumbnail segmentation that defines the locations on a video timeline where thumbnails are displayed. Candidate thumbnail locations are determined from boundaries of feature ranges of the video indicating when instances of detected features are present in the video. In some embodiments, candidate thumbnail separations are penalized for being separated by less than a minimum duration corresponding to a minimum pixel separation (e.g., the width of a thumbnail) between consecutive thumbnail locations on a video timeline. The thumbnail segmentation is computed by solving a shortest path problem through a graph that models different thumbnail locations and separations. As such, a video timeline is displayed with thumbnails at locations on the timeline defined by the thumbnail segmentation, with each thumbnail depicting a portion of the video associated with the thumbnail location.Type: GrantFiled: May 26, 2021Date of Patent: January 30, 2024Assignee: Adobe Inc.Inventors: Seth Walker, Hijung Shin, Cristin Ailidh Fraser, Aseem Agarwala, Lubomira Dontcheva, Joel Richard Brandt, Jovan Popović, Joy Oakyung Kim, Justin Salamon, Jui-hsien Wang, Timothy Jeewun Ganter, Xue Bai, Dingzeyu Li
-
Patent number: 11887629Abstract: Embodiments are directed to interactive tiles that represent video segments of a segmentation of a video. In some embodiments, each interactive tile represents a different video segment from a particular video segmentation (e.g., a default video segmentation). Each interactive tile includes a thumbnail (e.g., the first frame of the video segment represented by the tile), some transcript from the beginning of the video segment, a visualization of detected faces in the video segment, and one or more faceted timelines that visualize a category of detected features (e.g., a visualization of detected visual scenes, audio classifications, visual artifacts). In some embodiments, interacting with a particular interactive tile navigates to a corresponding portion of the video, adds a corresponding video segment to a selection, and/or scrubs through tile thumbnails.Type: GrantFiled: May 26, 2021Date of Patent: January 30, 2024Assignee: Adobe Inc.Inventors: Seth Walker, Hijung Shin, Cristin Ailidh Fraser, Aseem Agarwala, Lubomira Dontcheva, Joel Richard Brandt, Jovan Popović, Joy Oakyung Kim, Justin Salamon, Jui-hsien Wang, Timothy Jeewun Ganter, Xue Bai, Dingzeyu Li
-
Publication number: 20230368503Abstract: Embodiments are disclosed for correlating video sequences and audio sequences by a media recommendation system using a trained encoder network.Type: ApplicationFiled: May 11, 2022Publication date: November 16, 2023Applicant: Adobe Inc.Inventors: Justin SALAMON, Bryan RUSSELL, Didac SURIS COLL-VINENT
-
Publication number: 20230129350Abstract: Embodiments are disclosed for performing a section-based, within-song music similarity search by an audio recommendation system. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving an input including an audio sequence and a request to determine similar audio sequences to the audio sequence from a pre-processed audio catalog, analyzing the audio sequence to generate an audio embedding for the audio sequence, querying a pre-processed audio catalog to retrieve audio embeddings for catalog audio sequences at different time resolutions, generating a set of candidate audio sequences from the pre-processed audio catalog based on the audio embedding for the audio sequence, and providing the set of candidate audio sequences.Type: ApplicationFiled: May 11, 2022Publication date: April 27, 2023Applicant: Adobe Inc.Inventors: Nicholas J. BRYAN, Justin SALAMON
-
Publication number: 20230115212Abstract: Embodiments are disclosed for generating an audio segmentation of an audio sequence using deep embeddings. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving an input including an audio sequence and extracting features for each frame of the audio sequence, where each frame is associated with a beat of the audio sequence. The method may further comprise clustering frames of the audio sequence into one or more clusters based on the extracted features and generating segments of the audio sequence based on the clustered frames, where each segment includes frames of the audio sequence from a same cluster. The method may further comprise constructing a multi-level audio segmentation of the audio sequence and performing a segment fusioning process that merges shorter segments with neighboring segments based on cluster assignments.Type: ApplicationFiled: May 11, 2022Publication date: April 13, 2023Applicant: Adobe Inc.Inventors: Justin SALAMON, Oriol NIETO-CABALLERO, Nicholas J. BRYAN
-
Patent number: 11562668Abstract: Various embodiments of a modular signage system having a modular post that enables placement and display of a plurality of signs. The modular post includes a tab chamber that engages one or more signs and is configured to engage a stake for embedding within a ground surface.Type: GrantFiled: March 31, 2021Date of Patent: January 24, 2023Assignee: Becon Post LLCInventors: Randall Toltzman, Chris Buttenob, William Bingman, Justin Salamon, Scott Jarson, Debbie Jarson, Scott Smith
-
Patent number: 11501102Abstract: Certain embodiments involve techniques for automatically identifying sounds in an audio recording that match a selected sound. An audio search and editing system receives the audio recording and preprocesses the audio recording into audio portions. The audio portions are provided as a query to the neural network that includes a trained embedding model used to analyze the audio portions in view of the selected sound to estimate feature vectors. The audio search and editing system compares the feature vectors for the audio portions against the feature vector for the selected sound and the feature vector for the negative samples to generate an audio score that is a numerical representation of the level of similarity between the audio portion and the selected sound and uses the audio scores to classify the audio portions into a first class of matching sounds and a second class of non-matching sounds.Type: GrantFiled: November 21, 2019Date of Patent: November 15, 2022Assignee: Adobe Inc.Inventors: Justin Salamon, Yu Wang, Nicholas J. Bryan
-
Patent number: 11308329Abstract: A computer system is trained to understand audio-visual spatial correspondence using audio-visual clips having multi-channel audio. The computer system includes an audio subnetwork, video subnetwork, and pretext subnetwork. The audio subnetwork receives the two channels of audio from the audio-visual clips, and the video subnetwork receives the video frames from the audio-visual clips. In a subset of the audio-visual clips the audio-visual spatial relationship is misaligned, causing the audio-visual spatial cues for the audio and video to be incorrect. The audio subnetwork outputs an audio feature vector for each audio-visual clip, and the video subnetwork outputs a video feature vector for each audio-visual clip. The audio and video feature vectors for each audio-visual clip are merged and provided to the pretext subnetwork, which is configured to classify the merged vector as either having a misaligned audio-visual spatial relationship or not.Type: GrantFiled: May 7, 2020Date of Patent: April 19, 2022Assignee: Adobe Inc.Inventors: Justin Salamon, Bryan Russell, Karren Yang
-
Publication number: 20220076706Abstract: Embodiments are directed to interactive tiles that represent video segments of a segmentation of a video. In some embodiments, each interactive tile represents a different video segment from a particular video segmentation (e.g., a default video segmentation). Each interactive tile includes a thumbnail (e.g., the first frame of the video segment represented by the tile), some transcript from the beginning of the video segment, a visualization of detected faces in the video segment, and one or more faceted timelines that visualize a category of detected features (e.g., a visualization of detected visual scenes, audio classifications, visual artifacts). In some embodiments, interacting with a particular interactive tile navigates to a corresponding portion of the video, adds a corresponding video segment to a selection, and/or scrubs through tile thumbnails.Type: ApplicationFiled: May 26, 2021Publication date: March 10, 2022Inventors: Seth Walker, Hijung Shin, Cristin Ailidh Fraser, Aseem Agarwala, Lubomira Dontcheva, Joel Richard Brandt, Jovan Popovic, Joy Oakyung Kim, Justin Salamon, Jui-hsien Wang, Timothy Jeewun Ganter, Xue Bai, Dingzeyu Li
-
Publication number: 20220076707Abstract: Embodiments are directed to a snap point segmentation that defines the locations of selection snap points for a selection of video segments. Candidate snap points are determined from boundaries of feature ranges of the video indicating when instances of detected features are present in the video. In some embodiments, candidate snap point separations are penalized for being separated by less than a minimum duration corresponding to a minimum pixel separation between consecutive snap points on a video timeline. The snap point segmentation is computed by solving a shortest path problem through a graph that models different snap point locations and separations. When a user clicks or taps on the video timeline and drags, a selection snaps to the snap points defined by the snap point segmentation. In some embodiments, the snap points are displayed during a drag operation and disappear when the drag operation is released.Type: ApplicationFiled: May 26, 2021Publication date: March 10, 2022Inventors: Seth Walker, Hijung Shin, Cristin Ailidh Fraser, Aseem Agarwala, Lubomira Dontcheva, Joel Richard Brandt, Jovan Popovic, Joy Oakyung Kim, Justin Salamon, Jui-hsien Wang, Timothy Jeewun Ganter, Xue Bai, Dingzeyu Li
-
Publication number: 20220076026Abstract: Embodiments are directed to a thumbnail segmentation that defines the locations on a video timeline where thumbnails are displayed. Candidate thumbnail locations are determined from boundaries of feature ranges of the video indicating when instances of detected features are present in the video. In some embodiments, candidate thumbnail separations are penalized for being separated by less than a minimum duration corresponding to a minimum pixel separation (e.g., the width of a thumbnail) between consecutive thumbnail locations on a video timeline. The thumbnail segmentation is computed by solving a shortest path problem through a graph that models different thumbnail locations and separations. As such, a video timeline is displayed with thumbnails at locations on the timeline defined by the thumbnail segmentation, with each thumbnail depicting a portion of the video associated with the thumbnail location.Type: ApplicationFiled: May 26, 2021Publication date: March 10, 2022Inventors: Seth Walker, Hijung Shin, Cristin Ailidh Fraser, Aseem Agarwala, Lubomira Dontcheva, Joel Richard Brandt, Jovan Popovic, Joy Oakyung Kim, Justin Salamon, Jui-hsien Wang, Timothy Jeewun Ganter, Xue Bai, Dingzeyu Li
-
Publication number: 20210350135Abstract: A computer system is trained to understand audio-visual spatial correspondence using audio-visual clips having multi-channel audio. The computer system includes an audio subnetwork, video subnetwork, and pretext subnetwork. The audio subnetwork receives the two channels of audio from the audio-visual clips, and the video subnetwork receives the video frames from the audio-visual clips. In a subset of the audio-visual clips the audio-visual spatial relationship is misaligned, causing the audio-visual spatial cues for the audio and video to be incorrect. The audio subnetwork outputs an audio feature vector for each audio-visual clip, and the video subnetwork outputs a video feature vector for each audio-visual clip. The audio and video feature vectors for each audio-visual clip are merged and provided to the pretext subnetwork, which is configured to classify the merged vector as either having a misaligned audio-visual spatial relationship or not.Type: ApplicationFiled: May 7, 2020Publication date: November 11, 2021Applicant: Adobe Inc.Inventors: Justin Salamon, Bryan Russell, Karren Yang
-
Publication number: 20210304640Abstract: Various embodiments of a modular signage system having a modular post that enables placement and display of a plurality of signs. The modular post includes a tab chamber that engages one or more signs and is configured to engage a stake for embedding within a ground surface.Type: ApplicationFiled: March 31, 2021Publication date: September 30, 2021Inventors: Randall Toltzman, Chris Buttenob, William Bingman, Justin Salamon, Scott Jarson, Debbie Jarson, Scott Smith
-
Publication number: 20210158086Abstract: Certain embodiments involve techniques for automatically identifying sounds in an audio recording that match a selected sound. An audio search and editing system receives the audio recording and preprocesses the audio recording into audio portions. The audio portions are provided as a query to the neural network that includes a trained embedding model used to analyze the audio portions in view of the selected sound to estimate feature vectors. The audio search and editing system compares the feature vectors for the audio portions against the feature vector for the selected sound and the feature vector for the negative samples to generate an audio score that is a numerical representation of the level of similarity between the audio portion and the selected sound and uses the audio scores to classify the audio portions into a first class of matching sounds and a second class of non-matching sounds.Type: ApplicationFiled: November 21, 2019Publication date: May 27, 2021Inventors: Justin Salamon, Yu Wang, Nicholas J. Bryan
-
Publication number: 20200233397Abstract: A system for monitoring a condition of a machine includes an acoustic detector configured to capture an audio signal of the machine. A controller is communicatively coupled to the audio detector and configured to transmit the audio signal to a remote computing unit. The remote computing unit configured to generate a condition status signal based on at least one of an unsupervised machine learning process or a supervised machine learning process. The controller is configured to receive the condition status signal from the remote computing unit and communicate a condition status based on the received condition status signal.Type: ApplicationFiled: January 23, 2020Publication date: July 23, 2020Inventors: Juan Pablo Bello, Charlie Mydlarz, Justin Salamon
-
Patent number: D931833Type: GrantFiled: November 3, 2019Date of Patent: September 28, 2021Assignee: Rockford CorporationInventors: Jason Braaten, Justin Salamon
-
Patent number: D967267Type: GrantFiled: May 11, 2020Date of Patent: October 18, 2022Assignee: Becon Post LLCInventors: Randall Toltzman, Chris Buttenob, William Bingman, Justin Salamon, Scott Jarson, Debbie Jarson, Scott Smith