Patents by Inventor Vijay S. YELLAPRAGADA

Vijay S. YELLAPRAGADA has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11140290
    Abstract: Aspects of the present disclosure provide methods and apparatuses for processing a digital image of a document, for example, to determine whether the document is a long document. An exemplary method generally includes obtaining a plurality of digital images of the document, segmenting at least a first digital image of the plurality of images into pixels associated with a foreground of the first digital image and pixels associated with a background of the first digital image, detecting a plurality of contours in the segmented first digital image, deciding, for each detected contour of the plurality of contours, whether that contour is an open contour or a closed contour, and determining that one or more sides of the document is out of bounds based, at least in part, on the decisions.
    Type: Grant
    Filed: April 16, 2020
    Date of Patent: October 5, 2021
    Assignee: INTUIT, INC.
    Inventors: Vijay S. Yellapragada, Peijun Chiang, Daniel Lee, Jason Hall, Shailesh Soliwal
  • Publication number: 20200244831
    Abstract: Aspects of the present disclosure provide methods and apparatuses for processing a digital image of a document, for example, to determine whether the document is a long document. An exemplary method generally includes obtaining a plurality of digital images of the document, segmenting at least a first digital image of the plurality of images into pixels associated with a foreground of the first digital image and pixels associated with a background of the first digital image, detecting a plurality of contours in the segmented first digital image, deciding, for each detected contour of the plurality of contours, whether that contour is an open contour or a closed contour, and determining that one or more sides of the document is out of bounds based, at least in part, on the decisions.
    Type: Application
    Filed: April 16, 2020
    Publication date: July 30, 2020
    Inventors: Vijay S. YELLAPRAGADA, Daniel LEE, Jason HALL, Shailesh SOLIWAL
  • Patent number: 10659643
    Abstract: Aspects of the present disclosure provide methods and apparatuses for processing a digital image of a document, for example, to determine whether the document is a long document. An exemplary method generally includes obtaining a plurality of digital images of the document, segmenting at least a first digital image of the plurality of images into pixels associated with a foreground of the first digital image and pixels associated with a background of the first digital image, detecting a plurality of contours in the segmented first digital image, deciding, for each detected contour of the plurality of contours, whether that contour is an open contour or a closed contour, and determining that one or more sides of the document is out-of-bounds based, at least in part, on the decisions.
    Type: Grant
    Filed: November 15, 2018
    Date of Patent: May 19, 2020
    Assignee: INTUIT, INC.
    Inventors: Vijay S. Yellapragada, Peijun Chiang, Daniel Lee, Jason Hall, Shailesh Soliwal
  • Patent number: 10558856
    Abstract: The present disclosure relates to optical character recognition using captured video. According to one embodiment, using a first image in stream of images depicting a document, the device extracts text data in a portion of the document depicted in the first image and determines a first confidence level regarding an accuracy of the extracted text data. If the first confidence level satisfies a threshold value, the device saves the extracted text data as recognized content of the source document. Otherwise, the device extracts the text data from the portion of the document as depicted in one or more second images in the stream and determines a second confidence level for the text data extracted from each second image until identifying one of the second images where the second confidence level associated with the text data extracted from the identified second image satisfies the threshold value.
    Type: Grant
    Filed: September 4, 2018
    Date of Patent: February 11, 2020
    Assignee: INTUIT INC.
    Inventors: Vijay S. Yellapragada, Peijun Chiang, Sreeneel K. Maddika
  • Patent number: 10339373
    Abstract: Techniques are disclosed for performing optical character recognition (OCR) by identifying a template based on a hash of a document. One embodiment includes a method for identifying a template associated with an image. The method includes receiving a digital image, a portion of the image depicting a first document, and extracting the portion of the image. The method further includes scaling the portion of the image and generating a first hash from the scaled image. The method further includes comparing the first hash to a set of hashes, each corresponding to a template. The method further includes selecting a first template as corresponding to the first document based on comparing the first hash to the set of hashes and extracting one or more sections of the portion of the image based on the selected first template. The method further includes performing OCR on the extracted one or more sections.
    Type: Grant
    Filed: August 24, 2018
    Date of Patent: July 2, 2019
    Assignee: INTUIT INC.
    Inventors: Vijay S. Yellapragada, Peijun Chiang, Sreeneel K. Maddika
  • Patent number: 10289905
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Grant
    Filed: August 24, 2018
    Date of Patent: May 14, 2019
    Assignee: Intuit Inc.
    Inventors: Eugene Krivopaltsev, Sreeneel K. Maddika, Vijay S. Yellapragada
  • Patent number: 10282604
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Grant
    Filed: August 23, 2018
    Date of Patent: May 7, 2019
    Assignee: Intuit, Inc.
    Inventors: Eugene Krivopaltsev, Sreeneel K. Maddika, Vijay S. Yellapragada
  • Publication number: 20190089856
    Abstract: Aspects of the present disclosure provide methods and apparatuses for processing a digital image of a document, for example, to determine whether the document is a long document. An exemplary method generally includes obtaining a plurality of digital images of the document, segmenting at least a first digital image of the plurality of images into pixels associated with a foreground of the first digital image and pixels associated with a background of the first digital image, detecting a plurality of contours in the segmented first digital image, deciding, for each detected contour of the plurality of contours, whether that contour is an open contour or a closed contour, and determining that one or more sides of the document is out-of-bounds based, at least in part, on the decisions.
    Type: Application
    Filed: November 15, 2018
    Publication date: March 21, 2019
    Inventors: Vijay S. YELLAPRAGADA, Peijun CHIANG, Daniel LEE, Jason HALL, Shailesh SOLIWAL
  • Publication number: 20180365487
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Application
    Filed: August 23, 2018
    Publication date: December 20, 2018
    Inventors: Eugene KRIVOPALTSEV, Sreeneel K. MADDIKA, Vijay S. YELLAPRAGADA
  • Publication number: 20180365488
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Application
    Filed: August 24, 2018
    Publication date: December 20, 2018
    Inventors: Eugene KRIVOPALTSEV, Sreeneel K. MADDIKA, Vijay S. YELLAPRAGADA
  • Patent number: 10108879
    Abstract: The present disclosure includes techniques for selecting a candidate presentation style for individual documents for inclusion in an aggregate training data set for a document type that may be used to train an OCR processing engine prior to identifying text in an image of a document of the document type. In one embodiment, text input corresponding to a text sample in a document is received, and an image of the text sample in the document is received. For each of a plurality of candidate presentation styles, an OCR processing engine is trained using a training data set corresponding to the given candidate presentation style, and the OCR processing engine is used, as trained, to identify text in the received image. The OCR processing results for each candidate presentation style are compared to the received text input. A candidate presentation style for the document is selected based on the comparisons.
    Type: Grant
    Filed: September 21, 2016
    Date of Patent: October 23, 2018
    Assignee: Intuit inc.
    Inventors: Eugene Krivopaltsev, Sreeneel K. Maddika, Vijay S. Yellapragada
  • Patent number: 10095920
    Abstract: Techniques are disclosed for performing optical character recognition (OCR) by identifying a template based on a hash of a document. One embodiment includes a method for identifying a template associated with an image. The method includes receiving a digital image, a portion of the image depicting a first document, and extracting the portion of the image. The method further includes scaling the portion of the image and generating a first hash from the scaled image. The method further includes comparing the first hash to a set of hashes, each corresponding to a template. The method further includes selecting a first template as corresponding to the first document based on comparing the first hash to the set of hashes and extracting one or more sections of the portion of the image based on the selected first template. The method further includes performing OCR on the extracted one or more sections.
    Type: Grant
    Filed: July 28, 2016
    Date of Patent: October 9, 2018
    Assignee: INTUIT INC
    Inventors: Vijay S. Yellapragada, Peijun Chiang, Sreeneel Maddika
  • Patent number: 10089523
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Grant
    Filed: October 5, 2016
    Date of Patent: October 2, 2018
    Assignee: INTUIT INC.
    Inventors: Eugene Krivopaltsev, Sreeneel K. Maddika, Vijay S. Yellapragada
  • Publication number: 20180096200
    Abstract: Systems of the present disclosure generate accurate training data for optical character recognition (OCR). Systems disclosed herein generates images of a text passage as displayed piecemeal in a user interface (UI) element rendered in a selected font type and size, determine accurate dimensions and locations of bounding boxes for each character pictured in the images, stitch together a training image by concatenating the images, and associate the training image, the bounding box dimensions and locations, and the text passage together in a collection of training data. The collection of training data also includes a computer-readable master copy of the text passage with newline characters inserted therein.
    Type: Application
    Filed: October 5, 2016
    Publication date: April 5, 2018
    Inventors: Eugene KRIVOPALTSEV, Sreeneel K. MADDIKA, Vijay S. YELLAPRAGADA
  • Publication number: 20180082146
    Abstract: The present disclosure includes techniques for selecting a candidate presentation style for individual documents for inclusion in an aggregate training data set for a document type that may be used to train an OCR processing engine prior to identifying text in an image of a document of the document type. In one embodiment, text input corresponding to a text sample in a document is received, and an image of the text sample in the document is received. For each of a plurality of candidate presentation styles, an OCR processing engine is trained using a training data set corresponding to the given candidate presentation style, and the OCR processing engine is used, as trained, to identify text in the received image. The OCR processing results for each candidate presentation style are compared to the received text input. A candidate presentation style for the document is selected based on the comparisons.
    Type: Application
    Filed: September 21, 2016
    Publication date: March 22, 2018
    Inventors: Eugene KRIVOPALTSEV, Sreeneel K. MADDIKA, Vijay S. YELLAPRAGADA
  • Publication number: 20180032804
    Abstract: Techniques are disclosed for performing optical character recognition (OCR) by identifying a template based on a hash of a document. One embodiment includes a method for identifying a template associated with an image. The method includes receiving a digital image, a portion of the image depicting a first document, and extracting the portion of the image. The method further includes scaling the portion of the image and generating a first hash from the scaled image. The method further includes comparing the first hash to a set of hashes, each corresponding to a template. The method further includes selecting a first template as corresponding to the first document based on comparing the first hash to the set of hashes and extracting one or more sections of the portion of the image based on the selected first template. The method further includes performing OCR on the extracted one or more sections.
    Type: Application
    Filed: July 28, 2016
    Publication date: February 1, 2018
    Inventors: Vijay S. YELLAPRAGADA, Peijun CHIANG, Sreeneel MADDIKA