Patents by Inventor Ruxin Chen

Ruxin Chen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11386657
    Abstract: Methods and systems for performing sequence level prediction of a video scene are described. Video information in a video scene is represented as a sequence of features depicted each frame. One or more scene affective labels are provided at the end of the sequence. Each label pertains to the entire sequence of frames of data. An action is taken with an agent controlled by a machine learning algorithm for a current frame of the sequence at a current time step. An output of the action represents affective label prediction for the frame at the current time step. A pool of actions taken up until the current time step including the action taken with the agent is transformed into a predicted affective history for a subsequent time step. A reward is generated on predicted actions up to the current time step by comparing the predicted actions against corresponding annotated scene affective labels.
    Type: Grant
    Filed: January 4, 2021
    Date of Patent: July 12, 2022
    Assignee: SONY INTERACTIVE ENTERTAINMENT INC.
    Inventors: Ruxin Chen, Naveen Kumar, Haoqi Li
  • Publication number: 20220203248
    Abstract: Methods and systems are provided for providing real world assistance by a robot utility and interface device (RUID) are provided. A method provides for identifying a position of a user in a physical environment and a surface within the physical environment for projecting an interactive interface. The method also provides for moving to a location within the physical environment based on the position of the user and the surface for projecting the interactive interface. Moreover, the method provides for capturing a plurality of images of the interactive interface while the interactive interface is being interacted with by the use and for determining a selection of an input option made by the user.
    Type: Application
    Filed: January 11, 2022
    Publication date: June 30, 2022
    Inventors: Javier Fernandez Rico, Erik Beran, Michael Taylor, Ruxin Chen
  • Patent number: 11281709
    Abstract: For image captioning such as for computer game images or other images, bottom-up attention is combined with top-down attention to provide a multi-level residual attention-based image captioning model. A residual attention mechanism is first applied in the Faster R-CNN network to learn better feature representations for each region by taking spatial information into consideration. In the image captioning network, taking the extracted regional features as input, a second residual attention network is implemented to fuse the regional features attentionally for subsequent caption generation.
    Type: Grant
    Filed: July 28, 2020
    Date of Patent: March 22, 2022
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Jian Zheng, Ruxin Chen
  • Patent number: 11219837
    Abstract: Methods and systems are provided for providing real world assistance by a robot utility and interface device (RUID) are provided. A method provides for identifying a position of a user in a physical environment and a surface within the physical environment for projecting an interactive interface. The method also provides for moving to a location within the physical environment based on the position of the user and the surface for projecting the interactive interface. Moreover, the method provides for capturing a plurality of images of the interactive interface while the interactive interface is being interacted with by the use and for determining a selection of an input option made by the user.
    Type: Grant
    Filed: September 29, 2017
    Date of Patent: January 11, 2022
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Javier Fernandez Rico, Erik Beran, Michael Taylor, Ruxin Chen
  • Patent number: 11079999
    Abstract: Method for providing image of HMD user to a non-HMD user includes, receiving a first image of a user including the user's facial features captured by an external camera when the user is not wearing a head mounted display (HMD). A second image capturing a portion of the facial features of the user when the user is wearing the HMD is received. An image overlay data is generated by mapping contours of facial features captured in the second image with contours of corresponding facial features captured in the first image. The image overlay data is forwarded to the HMD for rendering on a second display screen that is mounted on a front face of the HMD.
    Type: Grant
    Filed: April 23, 2019
    Date of Patent: August 3, 2021
    Assignee: Sony Interactive Entertainment Inc.
    Inventor: Ruxin Chen
  • Patent number: 11064185
    Abstract: The present disclosure discloses a system, a method and a device for generating depth image. The system includes an illumination source, an optical system, a control device, and at least one set of a dynamic aperture and an image sensor, wherein the dynamic aperture is configured to dynamically change a light transmittance, an exposure start time, and an exposure end time under a control of the control device. The control device is configured to acquire a first photo and a second photo, and generate a depth image of the target scene according to the first photo, the first shooting configuration information, the second photo, and the second shooting configuration information.
    Type: Grant
    Filed: September 2, 2020
    Date of Patent: July 13, 2021
    Assignee: Rayz Technologies Co. Ltd.
    Inventors: Detao Du, Ruxin Chen
  • Publication number: 20210124930
    Abstract: Methods and systems for performing sequence level prediction of a video scene are described. Video information in a video scene is represented as a sequence of features depicted each frame. One or more scene affective labels are provided at the end of the sequence. Each label pertains to the entire sequence of frames of data. An action is taken with an agent controlled by a machine learning algorithm for a current frame of the sequence at a current time step. An output of the action represents affective label prediction for the frame at the current time step. A pool of actions taken up until the current time step including the action taken with the agent is transformed into a predicted affective history for a subsequent time step. A reward is generated on predicted actions up to the current time step by comparing the predicted actions against corresponding annotated scene affective labels.
    Type: Application
    Filed: January 4, 2021
    Publication date: April 29, 2021
    Inventors: Ruxin Chen, Naveen Kumar, Haoqi Li
  • Publication number: 20210074009
    Abstract: The present application discloses a 3D imaging method, device and depth camera, wherein, the 3D imaging method includes: acquiring depth information of points in a to-be-captured scene corresponding to at least one pixel; generating a 3D image of the to-be-captured scene based on the acquired depth information; and determining the depth information of the each of the points in the to-be-captured scene corresponding to the at least one pixel based on a result of the distributing. This implementation utilizes multiple charge collection areas to collect the charges obtained by photoelectric conversion of the light that reaches the pixels, thereby achieving the capture of the depth information of points in the to-be-captured scene corresponding to each pixel.
    Type: Application
    Filed: September 2, 2020
    Publication date: March 11, 2021
    Inventors: Detao Du, Ruxin Chen
  • Publication number: 20210067762
    Abstract: The present disclosure discloses a system, a method and a device for generating depth image. The system includes an illumination source, an optical system, a control device, and at least one set of a dynamic aperture and an image sensor, wherein the dynamic aperture is configured to dynamically change a light transmittance, an exposure start time, and an exposure end time under a control of the control device. The control device is configured to acquire a first photo and a second photo, and generate a depth image of the target scene according to the first photo, the first shooting configuration information, the second photo, and the second shooting configuration information.
    Type: Application
    Filed: September 2, 2020
    Publication date: March 4, 2021
    Inventors: Detao Du, Ruxin Chen
  • Publication number: 20210058535
    Abstract: Provided is a system with image sensors combination, including: a plurality of image sensors located in a same semiconductor chip package; and a controller for controlling the plurality of image sensors to operate in coordination, wherein each image sensor includes: a plurality of photosensitive pixel cells to be simultaneously exposed to generate analog photosensitive signals; an analog-to-digital converters for converting the analog photosensitive signals generated by the plurality of photosensitive pixel cells into image sensor signals; and a timing circuit for controlling the generation, readout, and transmission of electrical signals.
    Type: Application
    Filed: August 24, 2020
    Publication date: February 25, 2021
    Applicant: RAYZ TECHNOLOGIES CO., LTD.
    Inventors: Ruxin Chen, Detao Du
  • Publication number: 20210014401
    Abstract: The present disclosure provides a three-dimensional distance measuring method and device. The three-dimensional distance measuring method includes: emitting a first light pulse to illuminate a scene, and acquiring a first image of the scene within a first exposure time; emitting a second light pulse to illuminate the scene, and acquiring a second image of the scene within a second exposure time; acquiring scene distance information of the scene based on the first image of the scene and the second image of the scene, wherein the ratio of the first pulse envelope of the first light pulse to the second pulse envelope of the second light pulse is a monotonic function of time.
    Type: Application
    Filed: July 10, 2020
    Publication date: January 14, 2021
    Applicant: RAYZ TECHNOLOGIES CO., LTD.
    Inventors: Detao Du, Ruxin Chen
  • Publication number: 20210012117
    Abstract: Provided is a method for controlling a vehicle, including: receiving a plurality of visual sensor signals containing three-dimensional visual information outside and inside the vehicle; receiving a plurality of audio sensor signals containing three-dimensional audio information outside and inside the mobile vehicle; determining three-dimensional spatial positions of at least two objects and object cognitive information of the at least two objects according to at least a portion of the plurality of visual sensor signals and at least a portion of the plurality of audio sensor signals, wherein the at least two objects include at least one object outside the vehicle and at least one object inside the vehicle; obtaining a circumstance information of the vehicle according to the object cognitive information; and determining operation instructions for the vehicle according to the circumstance information.
    Type: Application
    Filed: July 8, 2020
    Publication date: January 14, 2021
    Applicant: RAYZ TECHNOLOGIES CO., LTD.
    Inventors: Ruxin Chen, Detao Du
  • Patent number: 10885341
    Abstract: Methods and systems for performing sequence level prediction of a video scene are described. Video information in a video scene is represented as a sequence of features depicted each frame. An environment state for each time step t corresponding to each frame is represented by the video information for time step t and predicted affective information from a previous time step t?1. An action A(t) as taken with an agent controlled by a machine learning algorithm for the frame at step t, wherein an output of the action A(t) represents affective label prediction for the frame at the time step t. A pool of predicted actions is transformed to a predicted affective history at a next time step t+1. The predictive affective history is included as part of the environment state for the next time step t+1. A reward R is generated on predicted actions up to the current time step t, by comparing them against corresponding annotated movie scene affective labels.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: January 5, 2021
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Ruxin Chen, Naveen Kumar, Haoqi Li
  • Patent number: 10850838
    Abstract: The present disclosure is related to unmanned aerial vehicles or drones that have a capability of quickly swapping batteries. This may be accomplished even as the drone continues to fly. A drone consistent with the present disclosure may drop one battery and pickup another using an attachment mechanism. Attachment mechanisms of the present disclosure may include electro-magnets, mechanical actuators, pins, or hooks. Systems consistent with the present disclosure may also include locations where replacement batteries may be provided to aircraft via actuation devices coupled to a physical location.
    Type: Grant
    Filed: December 29, 2016
    Date of Patent: December 1, 2020
    Assignee: SONY INTERACTIVE ENTERTAINMENT INC.
    Inventors: Dennis Dale Castleman, Ruxin Chen, Frank Zhao, Glenn Black
  • Publication number: 20200372058
    Abstract: For image captioning such as for computer game images or other images, bottom-up attention is combined with top-down attention to provide a multi-level residual attention-based image captioning model. A residual attention mechanism is first applied in the Faster R-CNN network to learn better feature representations for each region by taking spatial information into consideration. In the image captioning network, taking the extracted regional features as input, a second residual attention network is implemented to fuse the regional features attentionally for subsequent caption generation.
    Type: Application
    Filed: July 28, 2020
    Publication date: November 26, 2020
    Inventors: Jian Zheng, Ruxin Chen
  • Patent number: 10726062
    Abstract: For image captioning such as for computer game images or other images, bottom-up attention is combined with top-down attention to provide a multi-level residual attention-based image captioning model. A residual attention mechanism is first applied in the Faster R-CNN network to learn better feature representations for each region by taking spatial information into consideration. In the image captioning network, taking the extracted regional features as input, a second residual attention network is implemented to fuse the regional features attentionally for subsequent caption generation.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: July 28, 2020
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Jian Zheng, Ruxin Chen
  • Patent number: 10714076
    Abstract: A method for improved initialization of speech recognition system comprises mapping a trained hidden markov model based recognition node network (HMM) to a Connectionist Temporal Classification (CTC) based node label scheme. The central state of each frame in the HMM are mapped to CTC-labeled output nodes and the non-central states of each frame are mapped to CTC-blank nodes to generate a CTC-labeled HMM and each central state represents a phoneme from human speech detected and extracted by a computing device. Next the CTC-labeled HMM is trained using a cost function, wherein the cost function is not part of a CTC cost function. Finally the CTC-labeled HMM is trained using a CTC cost function to produce a CTC node network. The CTC node network may be iteratively trained by repeating the initialization steps.
    Type: Grant
    Filed: July 10, 2017
    Date of Patent: July 14, 2020
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Xavier Menendez-Pidal, Ruxin Chen
  • Publication number: 20200175053
    Abstract: For image captioning such as for computer game images or other images, bottom-up attention is combined with top-down attention to provide a multi-level residual attention-based image captioning model. A residual attention mechanism is first applied in the Faster R-CNN network to learn better feature representations for each region by taking spatial information into consideration. In the image captioning network, taking the extracted regional features as input, a second residual attention network is implemented to fuse the regional features attentionally for subsequent caption generation.
    Type: Application
    Filed: November 30, 2018
    Publication date: June 4, 2020
    Inventors: Jian Zheng, Ruxin Chen
  • Patent number: 10657701
    Abstract: Systems and methods for processing operations for head mounted display (HMD) users to join virtual reality (VR) scenes are provided. A computer-implemented method includes providing a first perspective of a VR scene to a first HMD of a first user and receiving an indication that a second user is requesting to join the VR scene provided to the first HMD. The method further includes obtaining real-world position and orientation data of the second HMD relative to the first HMD and then providing, based on said data, a second perspective of the VR scene. The method also provides that the first and second perspectives are each controlled by respective position and orientation changes while viewing the VR scene.
    Type: Grant
    Filed: January 11, 2017
    Date of Patent: May 19, 2020
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Steven Osman, Javier Fernandez Rico, Ruxin Chen
  • Publication number: 20200134444
    Abstract: A domain adaptation module is used to optimize a first domain derived from a second domain using respective outputs from respective parallel hidden layers of the domains.
    Type: Application
    Filed: October 31, 2018
    Publication date: April 30, 2020
    Inventors: Ruxin Chen, Min-Hung Chen, Jaekwon Yoo, Xiaoyu Liu