Patents by Inventor Eyal Krupka

Eyal Krupka has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230402038
    Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.
    Type: Application
    Filed: May 15, 2023
    Publication date: December 14, 2023
    Inventors: Adi DIAMANT, Xuedong HUANG, Karen MASTER BEN-DOR, Eyal KRUPKA, Raz HALALY, Yoni SMOLIN, Ilya GURVICH, Aviv HURVITZ, Lijuan QIN, Wei XIONG, Shixiong ZHANG, Lingfeng WU, Xiong XIAO, Ido LEICHTER, Moshe DAVID, Amit Kumar AGARWAL
  • Patent number: 11688399
    Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.
    Type: Grant
    Filed: December 8, 2020
    Date of Patent: June 27, 2023
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Adi Diamant, Karen Master Ben-Dor, Eyal Krupka, Raz Halaly, Yoni Smolin, Ilya Gurvich, Aviv Hurvitz, Lijuan Qin, Wei Xiong, Shixiong Zhang, Lingfeng Wu, Xiong Xiao, Ido Leichter, Moshe David, Xuedong Huang, Amit Kumar Agarwal
  • Publication number: 20230154238
    Abstract: Computer implemented method for detecting a hand gesture of a user, comprising: (a) Receiving sequential logic models each representing a hand gesture. The sequential logic model maps pre-defined hand poses and motions each represented by a hand features record defined by discrete hand values each indicating a state of respective hand feature. (b) Receiving a runtime sequence of runtime hand datasets each defined by discrete hand values scores indicating current state hand features of a user's moving hand which are inferred by analyzing timed images depicting the moving hand. (c) Submitting the runtime hand datasets and the pre-defined hand features records in SSVM functions to generate estimation terms for the runtime hand datasets with respect to the hand features records. (d) Estimating which of the hand gestures best matches the runtime sequence depicted in the timed images by optimizing score functions using the estimation terms for the runtime hand datasets.
    Type: Application
    Filed: January 18, 2023
    Publication date: May 18, 2023
    Inventors: Daniel FREEDMAN, Kfir KARMON, Eyal KRUPKA, Yagil ENGEL, Yevgeny SHAPIRO
  • Publication number: 20220374636
    Abstract: In a system including a processor and memory, the memory includes instructions that, when executed by the processor, cause the processor to control the system to perform receiving a video stream capturing objects; identifying, based on the received video stream, object areas corresponding to the objects, respectively; tracking the object areas in the received video stream; generating, based on the tracking of the object areas, visual data sets at a plurality of times, wherein each visual data set is generated at a different time and includes visual data representing each object area; determining a priority of each visual data in each visual data set; selecting, based on the determined priority of each visual data, a group of the visual data to be transmitted to a remote system; and transmitting, to the remote system, the selected group of the visual data.
    Type: Application
    Filed: May 24, 2021
    Publication date: November 24, 2022
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Moshe DAVID, Aviv HURVITZ, Eyal KRUPKA, Qingfen LIN, Arash GHANAIE-SICHANIE
  • Publication number: 20220343689
    Abstract: Computer implemented method for detecting a hand gesture of a user, comprising: (a) Receiving sequential logic models each representing a hand gesture. The sequential logic model maps pre-defined hand poses and motions each represented by a hand features record defined by discrete hand values each indicating a state of respective hand feature. (b) Receiving a runtime sequence of runtime hand datasets each defined by discrete hand values scores indicating current state hand features of a user's moving hand which are inferred by analyzing timed images depicting the moving hand. (c) Submitting the runtime hand datasets and the pre-defined hand features records in SSVM functions to generate estimation terms for the runtime hand datasets with respect to the hand features records. (d) Estimating which of the hand gestures best matches the runtime sequence depicted in the timed images by optimizing score functions using the estimation terms for the runtime hand datasets.
    Type: Application
    Filed: July 5, 2022
    Publication date: October 27, 2022
    Inventors: Daniel FREEDMAN, Kfir KARMON, Eyal KRUPKA, Yagil ENGEL, Yevgeny SHAPIRO
  • Publication number: 20220329960
    Abstract: The disclosed technology is generally directed to audio capture. In one example of the technology, recorded sounds are received such that the sounds recorded were emitted from multiple locations in an environment and such that the sounds recorded are sounds that can be converted to room impulse responses. The room impulse responses are generated from the recorded sounds. Location information that is associated with the multiple locations is received. At least the room impulses responses and the location information are used to generate at least one environment-specific model. Audio captured in the environment is received. An output is generated by processing the captured audio with the at least one environment-specific model such that the output includes at least one adjustment of the captured audio based on at least one acoustical property of the environment.
    Type: Application
    Filed: April 13, 2021
    Publication date: October 13, 2022
    Inventors: Stav YAGEV, Sharon KOUBI, Aviv HURVITZ, Igor ABRAMOVSKI, Eyal KRUPKA
  • Patent number: 11410464
    Abstract: Computer implemented method for detecting a hand gesture of a user, comprising: (a) Receiving sequential logic models each representing a hand gesture. The sequential logic model maps pre-defined hand poses and motions each represented by a hand features record defined by discrete hand values each indicating a state of respective hand feature. (b) Receiving a runtime sequence of runtime hand datasets each defined by discrete hand values scores indicating current state hand features of a user's moving hand which are inferred by analyzing timed images depicting the moving hand. (c) Submitting the runtime hand datasets and the pre-defined hand features records in SSVM functions to generate estimation terms for the runtime hand datasets with respect to the hand features records. (d) Estimating which of the hand gestures best matches the runtime sequence depicted in the timed images by optimizing score functions using the estimation terms for the runtime hand datasets.
    Type: Grant
    Filed: February 19, 2020
    Date of Patent: August 9, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Daniel Freedman, Kfir Karmon, Eyal Krupka, Yagil Engel, Yevgeny Shapiro
  • Patent number: 11182643
    Abstract: A user's collection of images may be analyzed to identify people's faces within the images, then create clusters of similar faces, where each of the clusters may represent a person. The clusters may be ranked in order of size to determine a relative importance of the associated person to the user. The ranking may be used in many social networking applications to filter and present content that may be of interest to the user. In one use scenario, the clusters may be used to identify images from a second user's image collection, where the identified images may be pertinent or interesting to the first user. The ranking may also be a function of user interactions with the images, as well as other input not related to the images. The ranking may be incrementally updated when new images are added to the user's collection.
    Type: Grant
    Filed: April 1, 2019
    Date of Patent: November 23, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Eyal Krupka, Igor Abramovski, Igor Kviatkovsky
  • Patent number: 11152006
    Abstract: Examples are disclosed that relate to voice identification enrollment. One example provides a method of voice identification enrollment comprising, during a meeting in which two or more human speakers speak at different times, determining whether one or more conditions of a protocol for sampling meeting audio used to establish human speaker voiceprints are satisfied, and in response to determining that the one or more conditions are satisfied, selecting a sample of meeting audio according to the protocol, the sample representing an utterance made by one of the human speakers. The method further comprises establishing, based at least on the sample, a voiceprint of the human speaker.
    Type: Grant
    Filed: June 27, 2018
    Date of Patent: October 19, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Eyal Krupka, Shixiong Zhang, Xiong Xiao
  • Publication number: 20210210097
    Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.
    Type: Application
    Filed: December 8, 2020
    Publication date: July 8, 2021
    Inventors: Adi DIAMANT, Karen MASTER BEN-DOR, Eyal KRUPKA, Raz HALALY, Yoni SMOLIN, Ilya GURVICH, Aviv HURVITZ, Lijuan QIN, Wei XIONG, Shixiong ZHANG, Lingfeng WU, Xiong XIAO, Ido LEICHTER, Moshe DAVID, Xuedong HUANG, Amit Kumar AGARWAL
  • Patent number: 10867610
    Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: December 15, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Adi Diamant, Karen Master Ben-Dor, Eyal Krupka, Raz Halaly, Yoni Smolin, Ilya Gurvich, Aviv Hurvitz, Lijuan Qin, Wei Xiong, Shixiong Zhang, Lingfeng Wu, Xiong Xiao, Ido Leichter, Moshe David, Xuedong Huang, Amit Kumar Agarwal
  • Patent number: 10847162
    Abstract: Multi-modal speech localization is achieved using image data captured by one or more cameras, and audio data captured by a microphone array. Audio data captured by each microphone of the array is transformed to obtain a frequency domain representation that is discretized in a plurality of frequency intervals. Image data captured by each camera is used to determine a positioning of each human face. Input data is provided to a previously-trained, audio source localization classifier, including: the frequency domain representation of the audio data captured by each microphone, and the positioning of each human face captured by each camera in which the positioning of each human face represents a candidate audio source. An identified audio source is indicated by the classifier based on the input data that is estimated to be the human face from which the audio data originated.
    Type: Grant
    Filed: June 27, 2018
    Date of Patent: November 24, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Eyal Krupka, Xiong Xiao
  • Publication number: 20200184204
    Abstract: Computer implemented method for detecting a hand gesture of a user, comprising: (a) Receiving sequential logic models each representing a hand gesture. The sequential logic model maps pre-defined hand poses and motions each represented by a hand features record defined by discrete hand values each indicating a state of respective hand feature. (b) Receiving a runtime sequence of runtime hand datasets each defined by discrete hand values scores indicating current state hand features of a user's moving hand which are inferred by analyzing timed images depicting the moving hand. (c) Submitting the runtime hand datasets and the pre-defined hand features records in SSVM functions to generate estimation terms for the runtime hand datasets with respect to the hand features records. (d) Estimating which of the hand gestures best matches the runtime sequence depicted in the timed images by optimizing score functions using the estimation terms for the runtime hand datasets.
    Type: Application
    Filed: February 19, 2020
    Publication date: June 11, 2020
    Inventors: Daniel FREEDMAN, Kfir KARMON, Eyal KRUPKA, Yagil ENGEL, Yevgeny SHAPIRO
  • Patent number: 10621991
    Abstract: A speaker recognition system includes a previously-trained joint neural network. An enrollment machine of the speaker recognition system is configured to operate the previously-trained joint neural network to enroll a new speaker based on audiovisual data featuring the newly enrolled speaker. A recognition machine of the speaker recognition system is configured to operate the previously-trained joint neural network to recognize a previously-enrolled speaker based on audiovisual data featuring the previously-enrolled speaker.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: April 14, 2020
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Shixiong Zhang, Eyal Krupka
  • Patent number: 10599919
    Abstract: Computer implemented method for detecting a hand gesture of a user, comprising: (a) Receiving sequential logic models each representing a hand gesture. The sequential logic model maps pre-defined hand poses and motions each represented by a hand features record defined by discrete hand values each indicating a state of respective hand feature. (b) Receiving a runtime sequence of runtime hand datasets each defined by discrete hand values scores indicating current state hand features of a user's moving hand which are inferred by analyzing timed images depicting the moving hand. (c) Submitting the runtime hand datasets and the pre-defined hand features records in SSVM functions to generate estimation terms for the runtime hand datasets with respect to the hand features records. (d) Estimating which of the hand gestures best matches the runtime sequence depicted in the timed images by optimizing score functions using the estimation terms for the runtime hand datasets.
    Type: Grant
    Filed: December 31, 2015
    Date of Patent: March 24, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Daniel Freedman, Kfir Karmon, Eyal Krupka, Yagil Engel, Yevgeny Shapiro
  • Patent number: 10488939
    Abstract: A gesture recognition method comprises receiving at a processor from a sensor a sequence of captured signal frames for extracting hand pose information for a hand and using at least one trained predictor executed on the processor to extract hand pose information from the received signal frames. For at least one defined gesture, defined as a time sequence comprising hand poses, with each of the hand poses defined as a conjunction or disjunction of qualitative propositions relating to interest points on the hand, truth values are computed for the qualitative propositions using the hand pose information extracted from the received signal frames, and execution of the gesture is tracked, by using the truth values to determine which of the hand poses in the time sequence have already been executed and which of the hand poses in the time sequence is expected next.
    Type: Grant
    Filed: August 7, 2017
    Date of Patent: November 26, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kfir Karmon, Aharon Bar-Hillel, Eyal Krupka, Noam Bloom, Ilya Gurvich, Aviv Hurvitz, Ido Leichter, Yoni Smolin, Yuval Tzairi, Alon Vinnikov
  • Publication number: 20190347516
    Abstract: A user's collection of images may be analyzed to identify people's faces within the images, then create clusters of similar faces, where each of the clusters may represent a person. The clusters may be ranked in order of size to determine a relative importance of the associated person to the user. The ranking may be used in many social networking applications to filter and present content that may be of interest to the user. In one use scenario, the clusters may be used to identify images from a second user's image collection, where the identified images may be pertinent or interesting to the first user. The ranking may also be a function of user interactions with the images, as well as other input not related to the images. The ranking may be incrementally updated when new images are added to the user's collection.
    Type: Application
    Filed: April 1, 2019
    Publication date: November 14, 2019
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Eyal Krupka, Igor Abramovski, Igor Kviatkovsky
  • Publication number: 20190341053
    Abstract: A computerized conference assistant includes a camera and a microphone. A face location machine of the computerized conference assistant finds a physical location of a human, based on a position of a candidate face in digital video captured by the camera. A beamforming machine of the computerized conference assistant outputs a beamformed signal isolating sounds originating from the physical location of the human. A diarization machine of the computerized conference assistant attributes information encoded in the beamformed signal to the human.
    Type: Application
    Filed: June 26, 2018
    Publication date: November 7, 2019
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Shixiong ZHANG, Lingfeng WU, Eyal KRUPKA, Xiong XIAO, Yifan GONG
  • Publication number: 20190341055
    Abstract: Examples are disclosed that relate to voice identification enrollment. One example provides a method of voice identification enrollment comprising, during a meeting in which two or more human speakers speak at different times, determining whether one or more conditions of a protocol for sampling meeting audio used to establish human speaker voiceprints are satisfied, and in response to determining that the one or more conditions are satisfied, selecting a sample of meeting audio according to the protocol, the sample representing an utterance made by one of the human speakers. The method further comprises establishing, based at least on the sample, a voiceprint of the human speaker.
    Type: Application
    Filed: June 27, 2018
    Publication date: November 7, 2019
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Eyal KRUPKA, Shixiong ZHANG, Xiong XIAO
  • Publication number: 20190341058
    Abstract: A speaker recognition system includes a previously-trained joint neural network. An enrollment machine of the speaker recognition system is configured to operate the previously-trained joint neural network to enroll a new speaker based on audiovisual data featuring the newly enrolled speaker. A recognition machine of the speaker recognition system is configured to operate the previously-trained joint neural network to recognize a previously-enrolled speaker based on audiovisual data featuring the previously-enrolled speaker.
    Type: Application
    Filed: June 28, 2018
    Publication date: November 7, 2019
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Shixiong ZHANG, Eyal KRUPKA