Patents by Inventor Malcolm Slaney

Malcolm Slaney has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230376527
    Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.
    Type: Application
    Filed: July 31, 2023
    Publication date: November 23, 2023
    Inventors: Malcolm SLANEY, Kilian WEINBERGER
  • Patent number: 11823439
    Abstract: Generally, the present disclosure is directed to systems and methods that train machine-learned models (e.g., artificial neural networks) to perform perceptual or cognitive task(s) based on biometric data (e.g., brain wave recordings) collected from living organism(s) while the living organism(s) are performing the perceptual or cognitive task(s). In particular, aspects of the present disclosure are directed to a new supervision paradigm, by which machine-learned feature extraction models are trained using example stimuli paired with companion biometric data such as neural activity recordings (e g electroencephalogram data, electrocorticography data, functional near-infrared spectroscopy, and/or magnetoencephalography data) collected from a living organism (e.g., human being) while the organism perceived those examples (e.g., viewing the image, listening to the speech, etc.).
    Type: Grant
    Filed: January 16, 2020
    Date of Patent: November 21, 2023
    Assignee: GOOGLE LLC
    Inventors: Aren Jansen, Malcolm Slaney
  • Patent number: 11748401
    Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.
    Type: Grant
    Filed: February 21, 2019
    Date of Patent: September 5, 2023
    Assignee: YAHOO ASSETS LLC
    Inventors: Malcolm Slaney, Kilian Weinberger
  • Patent number: 11510013
    Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.
    Type: Grant
    Filed: March 16, 2021
    Date of Patent: November 22, 2022
    Assignee: Iyo Inc.
    Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
  • Publication number: 20220130134
    Abstract: Generally, the present disclosure is directed to systems and methods that train machine-learned models (e.g., artificial neural networks) to perform perceptual or cognitive task(s) based on biometric data (e.g., brain wave recordings) collected from living organism(s) while the living organism(s) are performing the perceptual or cognitive task(s). In particular, aspects of the present disclosure are directed to a new supervision paradigm, by which machine-learned feature extraction models are trained using example stimuli paired with companion biometric data such as neural activity recordings (e g electroencephalogram data, electrocorticography data, functional near-infrared spectroscopy, and/or magnetoencephalography data) collected from a living organism (e.g., human being) while the organism perceived those examples (e.g., viewing the image, listening to the speech, etc.).
    Type: Application
    Filed: January 16, 2020
    Publication date: April 28, 2022
    Inventors: Aren Jansen, Malcolm Slaney
  • Patent number: 11256739
    Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicated matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.
    Type: Grant
    Filed: May 17, 2017
    Date of Patent: February 22, 2022
    Assignee: YAHOO ASSETS LLC
    Inventors: Arun Ramanujapuram, Malcolm Slaney
  • Publication number: 20210211810
    Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.
    Type: Application
    Filed: March 16, 2021
    Publication date: July 8, 2021
    Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
  • Patent number: 10959026
    Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.
    Type: Grant
    Filed: July 25, 2019
    Date of Patent: March 23, 2021
    Assignee: X Development LLC
    Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
  • Publication number: 20210029472
    Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.
    Type: Application
    Filed: July 25, 2019
    Publication date: January 28, 2021
    Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
  • Patent number: 10901500
    Abstract: Improving accuracy in understanding and/or resolving references to visual elements in a visual context associated with a computerized conversational system is described. Techniques described herein leverage gaze input with gestures and/or speech input to improve spoken language understanding in computerized conversational systems. Leveraging gaze input and speech input improves spoken language understanding in conversational systems by improving the accuracy by which the system can resolve references—or interpret a user's intent—with respect to visual elements in a visual context. In at least one example, the techniques herein describe tracking gaze to generate gaze input, recognizing speech input, and extracting gaze features and lexical features from the user input. Based at least in part on the gaze features and lexical features, user utterances directed to visual elements in a visual context can be resolved.
    Type: Grant
    Filed: April 30, 2019
    Date of Patent: January 26, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anna Prokofieva, Fethiye Asli Celikyilmaz, Dilek Z Hakkani-Tur, Larry Heck, Malcolm Slaney
  • Publication number: 20190179850
    Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.
    Type: Application
    Filed: February 21, 2019
    Publication date: June 13, 2019
    Inventors: Malcolm SLANEY, Kilian WEINBERGER
  • Patent number: 10317992
    Abstract: Improving accuracy in understanding and/or resolving references to visual elements in a visual context associated with a computerized conversational system is described. Techniques described herein leverage gaze input with gestures and/or speech input to improve spoken language understanding in computerized conversational systems. Leveraging gaze input and speech input improves spoken language understanding in conversational systems by improving the accuracy by which the system can resolve references—or interpret a user's intent—with respect to visual elements in a visual context. In at least one example, the techniques herein describe tracking gaze to generate gaze input, recognizing speech input, and extracting gaze features and lexical features from the user input. Based at least in part on the gaze features and lexical features, user utterances directed to visual elements in a visual context can be resolved.
    Type: Grant
    Filed: September 25, 2014
    Date of Patent: June 11, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anna Prokofieva, Fethiye Asli Celikyilmaz, Dilek Z Hakkani-Tur, Larry Heck, Malcolm Slaney
  • Patent number: 10216761
    Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.
    Type: Grant
    Filed: March 4, 2008
    Date of Patent: February 26, 2019
    Assignee: OATH INC.
    Inventors: Malcolm Slaney, Kilian Weinberger
  • Patent number: 10152517
    Abstract: The systems and methods described create a mathematical representation of each of the media objects for which user ratings are known. The mathematical representations take into account the subjective rating value assigned by a user to the respective media object and the user that assigned the rating value. The media object with the mathematical representation closest to that of the seed media object is then selected as the most similar media object to the seed media object. In an embodiment, the mathematical representation is a vector representation in which each user is a different dimension and each user's rating value is the magnitude of the vector in that dimension. Similarity between two songs is determined by identifying the closest vectors to that of the seed song. Closeness may be determined by subtracting or by calculating the dot product of each of the vectors with that of the seed media object.
    Type: Grant
    Filed: February 21, 2013
    Date of Patent: December 11, 2018
    Assignee: Excalibur IP, LLC
    Inventors: Malcolm Slaney, William White
  • Patent number: 9830351
    Abstract: Systems and methods for generating and playing a sequence of media objects based on a mood gradient are also disclosed. A mood gradient is a sequence of items, in which each item is media object having known characteristics or a representative set of characteristics of a media object, that is created or used by a user for a specific purpose. Given a mood gradient, one or more new media objects are selected for each item in the mood gradient based on the characteristics associated with that item. In this way, a sequence of new media objects is created but the sequence exhibits a similar variation in media object characteristics. The mood gradient may be presented to a user or created via a display illustrating a three-dimensional space in which each dimension corresponds to a different characteristic. The mood gradient may be represented as a path through the three-dimensional space and icons representing media objects are located within the three-dimensional space based on their characteristics.
    Type: Grant
    Filed: November 6, 2013
    Date of Patent: November 28, 2017
    Assignee: Yahoo! Inc.
    Inventors: William White, Malcolm Slaney
  • Publication number: 20170255650
    Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicated matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.
    Type: Application
    Filed: May 17, 2017
    Publication date: September 7, 2017
    Inventors: Arun Ramanujapuram, Malcolm Slaney
  • Patent number: 9665596
    Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicate matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.
    Type: Grant
    Filed: October 4, 2016
    Date of Patent: May 30, 2017
    Assignee: YAHOO! INC.
    Inventors: Arun Ramanujapuram, Malcolm Slaney
  • Patent number: 9639780
    Abstract: A system and method for improved classification. A first classifier is trained using a first process running on at least one computing device using a first set of training images relating to a class of images. A set of additional images are selected using the first classifier from a source of additional images accessible to the computing device. The first set of training images and the set of additional images are merged using the computing device to create a second set of training images. A second classifier is trained using a second process running on the computing device using the second set of training images. A set of unclassified images are classified using the second classifier thereby creating a set of classified images. The first classifier and the second classifier employ different classification methods.
    Type: Grant
    Filed: December 22, 2008
    Date of Patent: May 2, 2017
    Assignee: Excalibur IP, LLC
    Inventors: Marc Aurelio Ranzato, Kilian Quirin Weinberger, Eva Hoerster, Malcolm Slaney
  • Patent number: 9583105
    Abstract: Technologies described herein relate to modifying visual content for presentment on a display to facilitate improving performance of an automatic speech recognition (ASR) system. The visual content is modified to move elements further away from one another, wherein the moved elements give rise to ambiguity from the perspective of the ASR system. The visual content is modified to take into consideration accuracy of gaze tracking. When a user views an element in the modified visual content, the ASR system is customized as a function of the element being viewed by the user.
    Type: Grant
    Filed: June 6, 2014
    Date of Patent: February 28, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Andreas Stolcke, Geoffrey Zweig, Malcolm Slaney
  • Publication number: 20170024414
    Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicate matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.
    Type: Application
    Filed: October 4, 2016
    Publication date: January 26, 2017
    Inventors: Arun Ramanujapuram, Malcolm Slaney