Patents by Inventor Malcolm Slaney
Malcolm Slaney has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20230376527Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.Type: ApplicationFiled: July 31, 2023Publication date: November 23, 2023Inventors: Malcolm SLANEY, Kilian WEINBERGER
-
Patent number: 11823439Abstract: Generally, the present disclosure is directed to systems and methods that train machine-learned models (e.g., artificial neural networks) to perform perceptual or cognitive task(s) based on biometric data (e.g., brain wave recordings) collected from living organism(s) while the living organism(s) are performing the perceptual or cognitive task(s). In particular, aspects of the present disclosure are directed to a new supervision paradigm, by which machine-learned feature extraction models are trained using example stimuli paired with companion biometric data such as neural activity recordings (e g electroencephalogram data, electrocorticography data, functional near-infrared spectroscopy, and/or magnetoencephalography data) collected from a living organism (e.g., human being) while the organism perceived those examples (e.g., viewing the image, listening to the speech, etc.).Type: GrantFiled: January 16, 2020Date of Patent: November 21, 2023Assignee: GOOGLE LLCInventors: Aren Jansen, Malcolm Slaney
-
Patent number: 11748401Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.Type: GrantFiled: February 21, 2019Date of Patent: September 5, 2023Assignee: YAHOO ASSETS LLCInventors: Malcolm Slaney, Kilian Weinberger
-
Patent number: 11510013Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.Type: GrantFiled: March 16, 2021Date of Patent: November 22, 2022Assignee: Iyo Inc.Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
-
Publication number: 20220130134Abstract: Generally, the present disclosure is directed to systems and methods that train machine-learned models (e.g., artificial neural networks) to perform perceptual or cognitive task(s) based on biometric data (e.g., brain wave recordings) collected from living organism(s) while the living organism(s) are performing the perceptual or cognitive task(s). In particular, aspects of the present disclosure are directed to a new supervision paradigm, by which machine-learned feature extraction models are trained using example stimuli paired with companion biometric data such as neural activity recordings (e g electroencephalogram data, electrocorticography data, functional near-infrared spectroscopy, and/or magnetoencephalography data) collected from a living organism (e.g., human being) while the organism perceived those examples (e.g., viewing the image, listening to the speech, etc.).Type: ApplicationFiled: January 16, 2020Publication date: April 28, 2022Inventors: Aren Jansen, Malcolm Slaney
-
Patent number: 11256739Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicated matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.Type: GrantFiled: May 17, 2017Date of Patent: February 22, 2022Assignee: YAHOO ASSETS LLCInventors: Arun Ramanujapuram, Malcolm Slaney
-
Publication number: 20210211810Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.Type: ApplicationFiled: March 16, 2021Publication date: July 8, 2021Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
-
Patent number: 10959026Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.Type: GrantFiled: July 25, 2019Date of Patent: March 23, 2021Assignee: X Development LLCInventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
-
Publication number: 20210029472Abstract: In some embodiments, an ear-mounted sound reproduction system is provided. The system includes an ear-mountable housing that sits within the pinna of the ear and occludes the ear canal. In some embodiments, the ear-mountable housing includes a plurality of external-facing microphones. Because the external-facing microphones may be situated within the pinna of the ear but outside of the ear canal, the microphones will experience some, but not all, of the three-dimensional acoustic effects of the pinna. In some embodiments, sound is reproduced by an internal-facing driver element of the housing using a plurality of filters applied to the signals received by the plurality of external-facing microphones to preserve three-dimensional localization cues that would be present at the eardrum in the absence of the housing, such that the housing is essentially transparent to the user. In some embodiments, techniques are provided for deriving the plurality of filters.Type: ApplicationFiled: July 25, 2019Publication date: January 28, 2021Inventors: Malcolm Slaney, Ricardo Garcia, William Woods, Jason Rugolo
-
Patent number: 10901500Abstract: Improving accuracy in understanding and/or resolving references to visual elements in a visual context associated with a computerized conversational system is described. Techniques described herein leverage gaze input with gestures and/or speech input to improve spoken language understanding in computerized conversational systems. Leveraging gaze input and speech input improves spoken language understanding in conversational systems by improving the accuracy by which the system can resolve references—or interpret a user's intent—with respect to visual elements in a visual context. In at least one example, the techniques herein describe tracking gaze to generate gaze input, recognizing speech input, and extracting gaze features and lexical features from the user input. Based at least in part on the gaze features and lexical features, user utterances directed to visual elements in a visual context can be resolved.Type: GrantFiled: April 30, 2019Date of Patent: January 26, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Anna Prokofieva, Fethiye Asli Celikyilmaz, Dilek Z Hakkani-Tur, Larry Heck, Malcolm Slaney
-
Publication number: 20190179850Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.Type: ApplicationFiled: February 21, 2019Publication date: June 13, 2019Inventors: Malcolm SLANEY, Kilian WEINBERGER
-
Patent number: 10317992Abstract: Improving accuracy in understanding and/or resolving references to visual elements in a visual context associated with a computerized conversational system is described. Techniques described herein leverage gaze input with gestures and/or speech input to improve spoken language understanding in computerized conversational systems. Leveraging gaze input and speech input improves spoken language understanding in conversational systems by improving the accuracy by which the system can resolve references—or interpret a user's intent—with respect to visual elements in a visual context. In at least one example, the techniques herein describe tracking gaze to generate gaze input, recognizing speech input, and extracting gaze features and lexical features from the user input. Based at least in part on the gaze features and lexical features, user utterances directed to visual elements in a visual context can be resolved.Type: GrantFiled: September 25, 2014Date of Patent: June 11, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Anna Prokofieva, Fethiye Asli Celikyilmaz, Dilek Z Hakkani-Tur, Larry Heck, Malcolm Slaney
-
Patent number: 10216761Abstract: A method of generating congruous metadata is provided. The method includes receiving a similarity measure between at least two multimedia objects. Each multimedia object has associated metadata. If the at least two multimedia objects are similar based on the similarity measure and a similarity threshold, the associated metadata of each of the multimedia objects are compared. Then, based on the comparison of the associated metadata of each of the at least two multimedia objects, the method further includes generating congruous metadata. Metadata may be tags, for example.Type: GrantFiled: March 4, 2008Date of Patent: February 26, 2019Assignee: OATH INC.Inventors: Malcolm Slaney, Kilian Weinberger
-
Patent number: 10152517Abstract: The systems and methods described create a mathematical representation of each of the media objects for which user ratings are known. The mathematical representations take into account the subjective rating value assigned by a user to the respective media object and the user that assigned the rating value. The media object with the mathematical representation closest to that of the seed media object is then selected as the most similar media object to the seed media object. In an embodiment, the mathematical representation is a vector representation in which each user is a different dimension and each user's rating value is the magnitude of the vector in that dimension. Similarity between two songs is determined by identifying the closest vectors to that of the seed song. Closeness may be determined by subtracting or by calculating the dot product of each of the vectors with that of the seed media object.Type: GrantFiled: February 21, 2013Date of Patent: December 11, 2018Assignee: Excalibur IP, LLCInventors: Malcolm Slaney, William White
-
Patent number: 9830351Abstract: Systems and methods for generating and playing a sequence of media objects based on a mood gradient are also disclosed. A mood gradient is a sequence of items, in which each item is media object having known characteristics or a representative set of characteristics of a media object, that is created or used by a user for a specific purpose. Given a mood gradient, one or more new media objects are selected for each item in the mood gradient based on the characteristics associated with that item. In this way, a sequence of new media objects is created but the sequence exhibits a similar variation in media object characteristics. The mood gradient may be presented to a user or created via a display illustrating a three-dimensional space in which each dimension corresponds to a different characteristic. The mood gradient may be represented as a path through the three-dimensional space and icons representing media objects are located within the three-dimensional space based on their characteristics.Type: GrantFiled: November 6, 2013Date of Patent: November 28, 2017Assignee: Yahoo! Inc.Inventors: William White, Malcolm Slaney
-
Publication number: 20170255650Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicated matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.Type: ApplicationFiled: May 17, 2017Publication date: September 7, 2017Inventors: Arun Ramanujapuram, Malcolm Slaney
-
Patent number: 9665596Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicate matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.Type: GrantFiled: October 4, 2016Date of Patent: May 30, 2017Assignee: YAHOO! INC.Inventors: Arun Ramanujapuram, Malcolm Slaney
-
Patent number: 9639780Abstract: A system and method for improved classification. A first classifier is trained using a first process running on at least one computing device using a first set of training images relating to a class of images. A set of additional images are selected using the first classifier from a source of additional images accessible to the computing device. The first set of training images and the set of additional images are merged using the computing device to create a second set of training images. A second classifier is trained using a second process running on the computing device using the second set of training images. A set of unclassified images are classified using the second classifier thereby creating a set of classified images. The first classifier and the second classifier employ different classification methods.Type: GrantFiled: December 22, 2008Date of Patent: May 2, 2017Assignee: Excalibur IP, LLCInventors: Marc Aurelio Ranzato, Kilian Quirin Weinberger, Eva Hoerster, Malcolm Slaney
-
Patent number: 9583105Abstract: Technologies described herein relate to modifying visual content for presentment on a display to facilitate improving performance of an automatic speech recognition (ASR) system. The visual content is modified to move elements further away from one another, wherein the moved elements give rise to ambiguity from the perspective of the ASR system. The visual content is modified to take into consideration accuracy of gaze tracking. When a user views an element in the modified visual content, the ASR system is customized as a function of the element being viewed by the user.Type: GrantFiled: June 6, 2014Date of Patent: February 28, 2017Assignee: Microsoft Technology Licensing, LLCInventors: Andreas Stolcke, Geoffrey Zweig, Malcolm Slaney
-
Publication number: 20170024414Abstract: Embodiments of the invention are directed to using image data and contextual data to determine information about a scene, based on one or more previously obtained images. Contextual data, such location of image capture, can be used to determine previously obtained images related to the contextual data and other location-related information, such as billboard locations. With even low resolution devices, such as cell phone, image attributes, such as a histogram or optically recognized characters, can be compared between the previously obtained images and the newly captured image. Attributes matching within a predefined threshold indicate matching images. Information on the content of matching previously obtained images can be provided back to a user who captured the new image. User profile data can refine the content information. The content information can also be used as search terms for additional searching or other processing.Type: ApplicationFiled: October 4, 2016Publication date: January 26, 2017Inventors: Arun Ramanujapuram, Malcolm Slaney