Patents by Inventor Rohit Kumar Pandey
Rohit Kumar Pandey has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240290025Abstract: A method comprises receiving a first sequence of images of a portion of a user, the first sequence of images being monocular images; generating an avatar based on the first sequence of images, the avatar being based on a model including a feature vector associated with a vertex; receiving a second sequence of images of the portion of the user; and based on the second sequence of images, modifying the avatar with a displacement of the vertex to represent a gesture of the avatar.Type: ApplicationFiled: February 27, 2024Publication date: August 29, 2024Inventors: Yinda Zhang, Sean Ryan Francesco Fanello, Ziqian Bai, Feitong Tan, Zeng Huang, Kripasindhu Sarkar, Danhang Tang, Di Qiu, Abhimitra Meka, Ruofei Du, Mingsong Dou, Sergio Orts Escolano, Rohit Kumar Pandey, Thabo Beeler
-
Patent number: 12066282Abstract: A lighting stage includes a plurality of lights that project alternating spherical color gradient illumination patterns onto an object or human performer at a predetermined frequency. The lighting stage also includes a plurality of cameras that capture images of an object or human performer corresponding to the alternating spherical color gradient illumination patterns. The lighting stage also includes a plurality of depth sensors that capture depth maps of the object or human performer at the predetermined frequency. The lighting stage also includes (or is associated with) one or more processors that implement a machine learning algorithm to produce a three-dimensional (3D) model of the object or human performer. The 3D model includes relighting parameters used to relight the 3D model under different lighting conditions.Type: GrantFiled: November 11, 2020Date of Patent: August 20, 2024Assignee: GOOGLE LLCInventors: Sean Ryan Francesco Fanello, Kaiwen Guo, Peter Christopher Lincoln, Philip Lindsley Davidson, Jessica L. Busch, Xueming Yu, Geoffrey Harvey, Sergio Orts Escolano, Rohit Kumar Pandey, Jason Dourgarian, Danhang Tang, Adarsh Prakash Murthy Kowdle, Emily B. Cooper, Mingsong Dou, Graham Fyffe, Christoph Rhemann, Jonathan James Taylor, Shahram Izadi, Paul Ernest Debevec
-
Patent number: 12026833Abstract: Systems and methods are described for utilizing an image processing system with at least one processing device to perform operations including receiving a plurality of input images of a user, generating a three-dimensional mesh proxy based on a first set of features extracted from the plurality of input images and a second set of features extracted from the plurality of input images. The method may further include generating a neural texture based on a three-dimensional mesh proxy and the plurality of input images, generating a representation of the user including at least a neural texture, and sampling at least one portion of the neural texture from the three-dimensional mesh proxy. In response to providing the at least one sampled portion to a neural renderer, the method may include receiving, from the neural renderer, a synthesized image of the user that is previously not captured by the image processing system.Type: GrantFiled: October 28, 2020Date of Patent: July 2, 2024Assignee: Google LLCInventors: Ricardo Martin Brualla, Moustafa Meshry, Daniel Goldman, Rohit Kumar Pandey, Sofien Bouaziz, Ke Li
-
Publication number: 20240212106Abstract: Apparatus and methods related to applying lighting models to images are provided. An example method includes receiving, via a computing device, an image comprising a subject. The method further includes relighting, via a neural network, a foreground of the image to maintain a consistent lighting of the foreground with a target illumination. The relighting is based on a per-pixel light representation indicative of a surface geometry of the foreground. The light representation includes a specular component, and a diffuse component, of surface reflection. The method additionally includes predicting, via the neural network, an output image comprising the subject in the relit foreground. One or more neural networks can be trained to perform one or more of the aforementioned aspects.Type: ApplicationFiled: April 28, 2021Publication date: June 27, 2024Inventors: Chloe LeGendre, Paul Debevec, Sean Ryan Francesco Fanello, Rohit Kumar Pandey, Sergio Orts Escolano, Christian Haene, Sofien Bouaziz
-
Publication number: 20240212325Abstract: Systems and methods for training models to predict dense correspondences across images such as human images. A model may be trained using synthetic training data created from one or more 3D computer models of a subject. In addition, one or more geodesic distances derived from the surfaces of one or more of the 3D models may be used to generate one or more loss values, which may in turn be used in modifying the model's parameters during training.Type: ApplicationFiled: March 6, 2024Publication date: June 27, 2024Inventors: Yinda Zhang, Feitong Tan, Danhang Tang, Mingsong Dou, Kaiwen Guo, Sean Ryan Francesco Fanello, Sofien Bouaziz, Cem Keskin, Ruofei Du, Rohit Kumar Pandey, Deqing Sun
-
Patent number: 11954899Abstract: Systems and methods for training models to predict dense correspondences across images such as human images. A model may be trained using synthetic training data created from one or more 3D computer models of a subject. In addition, one or more geodesic distances derived from the surfaces of one or more of the 3D models may be used to generate one or more loss values, which may in turn be used in modifying the model's parameters during training.Type: GrantFiled: March 11, 2021Date of Patent: April 9, 2024Assignee: GOOGLE LLCInventors: Yinda Zhang, Feitong Tan, Danhang Tang, Mingsong Dou, Kaiwen Guo, Sean Ryan Francesco Fanello, Sofien Bouaziz, Cem Keskin, Ruofei Du, Rohit Kumar Pandey, Deqing Sun
-
Publication number: 20240046618Abstract: Systems and methods for training models to predict dense correspondences across images such as human images. A model may be trained using synthetic training data created from one or more 3D computer models of a subject. In addition, one or more geodesic distances derived from the surfaces of one or more of the 3D models may be used to generate one or more loss values, which may in turn be used in modifying the model's parameters during training.Type: ApplicationFiled: March 11, 2021Publication date: February 8, 2024Inventors: Yinda Zhang, Feitong Tan, Danhang Tang, Mingsong Dou, Kaiwen Guo, Sean Ryan Francesco Fanello, Sofien Bouaziz, Cem Keskin, Ruofei Du, Rohit Kumar Pandey, Deqing Sun
-
Publication number: 20240020915Abstract: Techniques include introducing a neural generator configured to produce novel faces that can be rendered at free camera viewpoints (e.g., at any angle with respect to the camera) and relit under an arbitrary high dynamic range (HDR) light map. A neural implicit intrinsic field takes a randomly sampled latent vector as input and produces as output per-point albedo, volume density, and reflectance properties for any queried 3D location. These outputs are aggregated via a volumetric rendering to produce low resolution albedo, diffuse shading, specular shading, and neural feature maps. The low resolution maps are then upsampled to produce high resolution maps and input into a neural renderer to produce relit images.Type: ApplicationFiled: July 17, 2023Publication date: January 18, 2024Inventors: Yinda Zhang, Feitong Tan, Sean Ryan Francesco Fanello, Abhimitra Meka, Sergio Orts Escolano, Danhang Tang, Rohit Kumar Pandey, Jonathan James Taylor
-
Patent number: 11868523Abstract: Techniques of tracking a user's gaze includes identifying a region of a display at which a gaze of a user is directed, the region including a plurality of pixels. By determining a region rather than a point, when the regions correspond to elements of a user interface, the improved technique enables a system to activate the element to which a determined region is selected. In some implementations, the system makes the determination using a classification engine including a convolutional neural network; such an engine takes as input images of the user's eye and outputs a list of probabilities that the gaze is directed to each of the regions.Type: GrantFiled: July 1, 2021Date of Patent: January 9, 2024Assignee: GOOGLE LLCInventors: Ivana Tosic Rodgers, Sean Ryan Francesco Fanello, Sofien Bouaziz, Rohit Kumar Pandey, Eric Aboussouan, Adarsh Prakash Murthy Kowdle
-
Publication number: 20230419600Abstract: Example embodiments relate to techniques for volumetric performance capture with neural rendering. A technique may involve initially obtaining images that depict a subject from multiple viewpoints and under various lighting conditions using a light stage and depth data corresponding to the subject using infrared cameras. A neural network may extract features of the subject from the images based on the depth data and map the features into a texture space (e.g., the UV texture space). A neural renderer can be used to generate an output image depicting the subject from a target view such that illumination of the subject in the output image aligns with the target view. The neural render may resample the features of the subject from the texture space to an image space to generate the output image.Type: ApplicationFiled: November 5, 2020Publication date: December 28, 2023Inventors: Sean Ryan Francesco FANELLO, Abhi MEKA, Rohit Kumar PANDEY, Christian HAENE, Sergio Orts ESCOLANO, Christoph RHEMANN, Paul DEBEVEC, Sofien BOUAZIZ, Thabo BEELER, Ryan OVERBECK, Peter BARNUM, Daniel ERICKSON, Philip DAVIDSON, Yinda ZHANG, Jonathan TAYLOR, Chloe LeGENDRE, Shahram IZADI
-
Publication number: 20230360182Abstract: Apparatus and methods related to applying lighting models to images of objects are provided. An example method includes applying a geometry model to an input image to determine a surface orientation map indicative of a distribution of lighting on an object based on a surface geometry. The method further includes applying an environmental light estimation model to the input image to determine a direction of synthetic lighting to be applied to the input image. The method also includes applying, based on the surface orientation map and the direction of synthetic lighting, a light energy model to determine a quotient image indicative of an amount of light energy to be applied to each pixel of the input image. The method additionally includes enhancing, based on the quotient image, a portion of the input image. One or more neural networks can be trained to perform one or more of the aforementioned aspects.Type: ApplicationFiled: May 17, 2021Publication date: November 9, 2023Inventors: Sean Ryan Francesco Fanello, Yun-Ta Tsai, Rohit Kumar Pandey, Paul Debevec, Michael Milne, Chloe LeGendre, Jonathan Tilton Barron, Christoph Rhemann, Sofien Bouaziz, Navin Padman Sarma
-
Patent number: 11710287Abstract: Systems and methods are described for generating a plurality of three-dimensional (3D) proxy geometries of an object, generating, based on the plurality of 3D proxy geometries, a plurality of neural textures of the object, the neural textures defining a plurality of different shapes and appearances representing the object, providing the plurality of neural textures to a neural renderer, receiving, from the neural renderer and based on the plurality of neural textures, a color image and an alpha mask representing an opacity of at least a portion of the object, and generating a composite image based on the pose, the color image, and the alpha mask.Type: GrantFiled: August 4, 2020Date of Patent: July 25, 2023Assignee: GOOGLE LLCInventors: Ricardo Martin Brualla, Daniel Goldman, Sofien Bouaziz, Rohit Kumar Pandey, Matthew Brown
-
Publication number: 20230004216Abstract: Techniques of tracking a user's gaze includes identifying a region of a display at which a gaze of a user is directed, the region including a plurality of pixels. By determining a region rather than a point, when the regions correspond to elements of a user interface, the improved technique enables a system to activate the element to which a determined region is selected. In some implementations, the system makes the determination using a classification engine including a convolutional neural network; such an engine takes as input images of the user's eye and outputs a list of probabilities that the gaze is directed to each of the regions.Type: ApplicationFiled: July 1, 2021Publication date: January 5, 2023Inventors: Ivana Tosic Rodgers, Sean Ryan Francesco Fanello, Sofien Bouaziz, Rohit Kumar Pandey, Eric Aboussouan, Adarsh Prakash Murthy Kowdle
-
Publication number: 20220130111Abstract: Systems and methods are described for utilizing an image processing system with at least one processing device to perform operations including receiving a plurality of input images of a user, generating a three-dimensional mesh proxy based on a first set of features extracted from the plurality of input images and a second set of features extracted from the plurality of input images. The method may further include generating a neural texture based on a three-dimensional mesh proxy and the plurality of input images, generating a representation of the user including at least a neural texture, and sampling at least one portion of the neural texture from the three-dimensional mesh proxy. In response to providing the at least one sampled portion to a neural renderer, the method may include receiving, from the neural renderer, a synthesized image of the user that is previously not captured by the image processing system.Type: ApplicationFiled: October 28, 2020Publication date: April 28, 2022Inventors: Ricardo Martin Brualla, Moustafa Meshry, Daniel Goldman, Rohit Kumar Pandey, Sofien Bouaziz, Ke Li
-
Patent number: 11269172Abstract: Embodiments of present disclosure discloses system and method for reconstruction of FOV. Initially, presence of one of single object and distinct objects in FOV of image of sample comprising one or more objects is determined based on sharpness of one or more objects. A single optimal representation of FOV may be generated when presence of single object is determined. At least one of single optimal representation and a depth-based enhanced representation of FOV may be generated when presence of distinct objects is determined. For generating depth-based enhanced representation, one or more first optimal images associated with each of distinct objects in FOV may be retrieved. An optimal representation of each of distinct objects is generated based on corresponding one or more first optimal images. Further, optimal representation of each of distinct objects is placed at corresponding optimal depth associated with respective distinct object to generate depth-based enhanced representation.Type: GrantFiled: February 8, 2018Date of Patent: March 8, 2022Assignee: SIGTUPLE TECHNOLOGIES PRIVATE LIMITEDInventors: Harshit Pande, Abdul Aziz, Bharath Cheluvaraju, Tathagato Rai Dastidar, Apurv Anand, Rohit Kumar Pandey
-
Publication number: 20220065620Abstract: A lighting stage includes a plurality of lights that project alternating spherical color gradient illumination patterns onto an object or human performer at a predetermined frequency. The lighting stage also includes a plurality of cameras that capture images of an object or human performer corresponding to the alternating spherical color gradient illumination patterns. The lighting stage also includes a plurality of depth sensors that capture depth maps of the object or human performer at the predetermined frequency. The lighting stage also includes (or is associated with) one or more processors that implement a machine learning algorithm to produce a three-dimensional (3D) model of the object or human performer. The 3D model includes relighting parameters used to relight the 3D model under different lighting conditions.Type: ApplicationFiled: November 11, 2020Publication date: March 3, 2022Inventors: Sean Ryan Francesco Fanello, Kaiwen Guo, Peter Christopher Lincoln, Philip Lindsley Davidson, Jessica L. Busch, Xueming Yu, Geoffrey Harvey, Sergio Orts Escolano, Rohit Kumar Pandey, Jason Dourgarian, Danhang Tang, Adarsh Prakash Murthy Kowdle, Emily B. Cooper, Mingsong Dou, Graham Fyffe, Christoph Rhemann, Jonathan James Taylor, Shahram Izadi, Paul Ernest Debevec
-
Publication number: 20220051485Abstract: Systems and methods are described for generating a plurality of three-dimensional (3D) proxy geometries of an object, generating, based on the plurality of 3D proxy geometries, a plurality of neural textures of the object, the neural textures defining a plurality of different shapes and appearances representing the object, providing the plurality of neural textures to a neural renderer, receiving, from the neural renderer and based on the plurality of neural textures, a color image and an alpha mask representing an opacity of at least a portion of the object, and generating a composite image based on the pose, the color image, and the alpha mask.Type: ApplicationFiled: August 4, 2020Publication date: February 17, 2022Inventors: Ricardo Martin Brualla, Daniel Goldman, Sofien Bouaziz, Rohit Kumar Pandey, Matthew Brown
-
Patent number: 11189027Abstract: Disclosed subject matter relates to Peripheral Blood Smear (PBS) that determines an area to be scanned in PBS for analysis. A PBS analysing system captures a focused image at each of plurality of positions in the PBS and determines Quality Indicators (QIs) in focused image. Further, a region is identified in PBS where QIs of focused image satisfy predefined QI threshold limits, as a monolayer region of PBS and determines an initiation point in monolayer region based on cell count value and co-ordinates of each of the plurality of positions located in the monolayer region. Finally, the area to be scanned in monolayer region is determined based on the initiation point and a predefined scan pattern. Determining the area to be scanned yields accurate and faster results.Type: GrantFiled: May 15, 2018Date of Patent: November 30, 2021Assignee: Sigtuple Technologies Private LimitedInventors: Shreepad Potadar, Dheeraj Mundhra, Abhishek Shukla, Raghu G, Amrutha Muralidharan, Deepak Kapoor, Vijay Muralidharan, Nivedita Muthusubramanian, Bharath Cheluvaraju, Apurv Anand, Tathagato Rai Dastidar, Rohit Kumar Pandey
-
Publication number: 20210165204Abstract: Embodiments of present disclosure discloses system and method for reconstruction of FOV. Initially, presence of one of single object and distinct objects in FOV of image of sample comprising one or more objects is determined based on sharpness of one or more objects. A single optimal representation of FOV may be generated when presence of single object is determined. At least one of single optimal representation and a depth-based enhanced representation of FOV may be generated when presence of distinct objects is determined. For generating depth-based enhanced representation, one or more first optimal images associated with each of distinct objects in FOV may be retrieved. An optimal representation of each of distinct objects is generated based on corresponding one or more first optimal images. Further, optimal representation of each of distinct objects is placed at corresponding optimal depth associated with respective distinct object to generate depth-based enhanced representation.Type: ApplicationFiled: February 8, 2018Publication date: June 3, 2021Applicant: SIGTUPLE TECHNOLOGIES PRIVATE LIMITEDInventors: Harshit PANDE, Abdul AZIZ, Bharath CHELUVARAJU, Tathagato Rai DASTIDAR, Apurv ANAND, Rohit Kumar PANDEY
-
Patent number: 10997457Abstract: Methods, systems, and media for relighting images using predicted deep reflectance fields are provided.Type: GrantFiled: October 16, 2019Date of Patent: May 4, 2021Assignee: Google LLCInventors: Christoph Rhemann, Abhimitra Meka, Matthew Whalen, Jessica Lynn Busch, Sofien Bouaziz, Geoffrey Douglas Harvey, Andrea Tagliasacchi, Jonathan Taylor, Paul Debevec, Peter Joseph Denny, Sean Ryan Francesco Fanello, Graham Fyffe, Jason Angelo Dourgarian, Xueming Yu, Adarsh Prakash Murthy Kowdle, Julien Pascal Christophe Valentin, Peter Christopher Lincoln, Rohit Kumar Pandey, Christian Häne, Shahram Izadi