Context Analysis Or Word Recognition (e.g., Character String) Patents (Class 382/229)
  • Patent number: 8988543
    Abstract: The present invention relates to a camera based method for text input and detection of a keyword or of a text-part within page or a screen comprising the steps of: directing a camera module on the printed page and capturing an image thereof; digital image filtering of the captured image; detection of word blocks contained in the image, each word block containing most likely a recognizable word; performing OCR within each word block; determination of A-blocks among the word blocks according to a keyword probability determination rule, wherein each of the A-blocks contains most likely the keyword; assignment of an attribute to each A-block; indication of the A-blocks in the display by a frame or the like for a further selection of the keyword; further selection of the A-block containing the keyword based on the displayed attribute of the keyword; forwarding the text content as text input to an application.
    Type: Grant
    Filed: April 28, 2011
    Date of Patent: March 24, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Cüneyt Göktekin, Oliver Tenchio
  • Patent number: 8989431
    Abstract: A Mixed Media Reality (MMR) system and associated techniques are disclosed. The MMR system provides mechanisms for forming a mixed media document that includes media of at least two types (e.g., printed paper as a first medium and digital content and/or web link as a second medium). The present invention provides a system, method, and computer program product for creating an ad hoc association between users. Captured digital images of paper documents or portions associated with a user are received. Document recognition features are extracted from the captured digital image. An ad hoc association is created between the user and another user associated with an existing document recognition feature similar to the extracted document recognition features.
    Type: Grant
    Filed: March 31, 2008
    Date of Patent: March 24, 2015
    Assignee: Ricoh Co., Ltd.
    Inventors: Berna Erol, Jonathan J. Hull, Hidenobu Kishi, Qifa Ke, Jorge Moraleda
  • Patent number: 8983211
    Abstract: A method, a system, and a computer program product for processing the output of an OCR are disclosed. The system receives a first character sequence from the OCR. A first set of characters from the first character sequence are converted to a corresponding second set of characters to generate a second character sequence based on a look-up table and language scores.
    Type: Grant
    Filed: May 14, 2012
    Date of Patent: March 17, 2015
    Assignee: Xerox Corporation
    Inventors: Sriram Venkatapathy, Nicola Cancedda
  • Publication number: 20150063714
    Abstract: A system for processing a text capture operation is described. The system receives text captured from a rendered document in the text capture operation. The system also receives supplemental information distinct from the captured text. The system determines an action to perform in response to the text capture operation based upon both the captured text and the supplemental information.
    Type: Application
    Filed: September 5, 2014
    Publication date: March 5, 2015
    Inventors: Martin T. King, Dale L. Grover, Clifford A. Kushler, James Q. Stafford-Fraser
  • Patent number: 8971670
    Abstract: A system includes preparing respective proof reading tools for performing carpet proof reading and side-by-side proof reading of text data, recording a log of time to perform proof reading operations by using the first and second proof reading tools. The method further includes estimating, based on times stored in a log, times to perform proof reading of a character using 1) the first proof reading tool followed by using the second proof reading tool, and 2) the second proof reading tool. The method further includes determining for each character value, based on the estimated times, to use the first proof reading tool along with using the second proof reading tool and determining, or to use the second proof reading tool without using the first proof reading tool.
    Type: Grant
    Filed: November 6, 2012
    Date of Patent: March 3, 2015
    Assignee: International Business Machines Corporation
    Inventors: Takashi Itoh, Toshinari Itoki, Takayuki Osogami
  • Publication number: 20150043832
    Abstract: An information processing apparatus includes a storage unit, an interpretation unit, and a correction unit. The storage unit stores plural correction instructions. The interpretation unit interprets a correction instruction stored in the storage unit. The correction unit corrects a recognized character string in accordance with the correction instruction interpreted by the interpretation unit. The interpretation unit determines the type of the correction instruction, and extracts a first character string including one or more characters serving as a target of the correction instruction and a second character string obtained by performing conversion of a part of or whole the first character string, in accordance with the type of the correction instruction. The correction unit, in a case where the first character string exists in the recognized character string, converts a part of or whole the first character string within the recognized character string into the second character string.
    Type: Application
    Filed: February 25, 2014
    Publication date: February 12, 2015
    Applicant: FUJI XEROX CO., LTD.
    Inventors: Satoshi KUBOTA, Shunichi KIMURA
  • Patent number: 8953910
    Abstract: A method includes preparing respective proof reading tools for performing carpet proof reading and side-by-side proof reading of text data, recording a log of time to perform proof reading operations by using the first and second proof reading tools. The method further includes estimating, based on times stored in a log, times to perform proof reading of a character using 1) the first proof reading tool followed by using the second proof reading tool, and 2) the second proof reading tool. The method further includes determining for each character value, based on the estimated times, to use the first proof reading tool along with using the second proof reading tool or to use the second proof reading tool without using the first proof reading tool.
    Type: Grant
    Filed: September 28, 2012
    Date of Patent: February 10, 2015
    Assignee: International Business Machines Corporation
    Inventors: Takashi Itoh, Toshinari Itoki, Takayuki Osogami
  • Patent number: 8949267
    Abstract: An appropriate search is carried out even with images including a complicated layout structure, decorated characters, and so on.
    Type: Grant
    Filed: February 28, 2011
    Date of Patent: February 3, 2015
    Assignee: Rakuten, Inc.
    Inventor: Soh Masuko
  • Patent number: 8942489
    Abstract: A vector graphics classification engine and associated method for classifying vector graphics in a fixed format document is described herein and illustrated in the accompanying figures. The vector graphics classification engine defines a pipeline for categorizing vector graphics parsed from the fixed format document as font, text, paragraph, table, and page effects, such as shading, borders, underlines, and strikethroughs. Vector graphics that are not otherwise classified are designated as basic graphics. By sequencing the detection operations in a selected order, misclassification is minimized or eliminated.
    Type: Grant
    Filed: January 23, 2012
    Date of Patent: January 27, 2015
    Assignee: Microsoft Corporation
    Inventors: Milan Sesum, Milos Raskovic, Drazen Zaric, Milos Lazarevic, Aljosa Obuljen
  • Patent number: 8938383
    Abstract: Apparatus and method for allowing a test script to be played back correctly in a locale of different test language. The invention uses a synonymy dictionary storing the different appearances of the property value of a property in an object of a software product to be tested in different test locales; and compares the property value of the property in the object of the software product to be tested to the corresponding property value pre-recorded in a test script to detect whether they match each other.
    Type: Grant
    Filed: August 25, 2006
    Date of Patent: January 20, 2015
    Assignee: International Business Machines Corporation
    Inventors: Min Ding, Chi Qi, Xiao Bin Yang
  • Patent number: 8934723
    Abstract: Embodiments are provided for organization and presentation of content. In some embodiments, a plurality of images and a plurality of similarity rules for image categorization are received. For each image in the plurality of images, that image and each remaining image from the plurality is compared by: applying each similarity rule to the image and a remaining image from the plurality to obtain a numeric result, and recording the numeric result for the pair of images in a numeric representation, the numeric representation embodying similarities. The numeric representation is used as a reference for clustering the plurality of images into clusters of similar images, and each image is stored with a marker denoting a cluster to which it has been assigned.
    Type: Grant
    Filed: December 17, 2013
    Date of Patent: January 13, 2015
    Assignee: Dropbox, Inc.
    Inventors: Michael Dwan, Jinpeng Ren
  • Patent number: 8929461
    Abstract: Machine-readable media, methods, apparatus and system for caption detection are described. In some embodiments, a plurality of text boxes may be detected from a plurality of frames. A first percentage of the plurality of text boxes whose locations on the plurality of frames fall into a location range may be obtained. A second percentage of the plurality of text boxes whose sizes fall into a size range may be obtained. Then, it may be determined if the first percentage and the location range are acceptable and if the second percentage and the size range are acceptable.
    Type: Grant
    Filed: April 17, 2007
    Date of Patent: January 6, 2015
    Assignee: Intel Corporation
    Inventors: Wei Hu, Rui Ma
  • Patent number: 8918734
    Abstract: A reduced virtual keyboard system for text input on electronic devices is disclosed. Text input is performed by creating a tracing trajectory. Dynamic prediction solutions are created during the tracing process, thus avoiding the need for a user to complete the entire word trajectory. The system also allows a mixture of tapping actions and sliding motions for the same word. The system may comprise a Long Words Dictionary database having first letters corresponding to predetermined keys of the keyboard. Alternatively, the system uses a Dictionary and a database management tool to find long words.
    Type: Grant
    Filed: July 24, 2011
    Date of Patent: December 23, 2014
    Assignee: Nuance Communications, Inc.
    Inventor: Daniel Suraqui
  • Patent number: 8914278
    Abstract: A computer-assisted language correction system including spelling correction functionality, misused word correction functionality, grammar correction functionality and vocabulary enhancement functionality utilizing contextual feature-sequence functionality employing an internet corpus.
    Type: Grant
    Filed: July 31, 2008
    Date of Patent: December 16, 2014
    Assignee: Ginger Software, Inc.
    Inventors: Yael Karov Zangvil, Avner Zangvil
  • Patent number: 8913063
    Abstract: Techniques are described for detecting script-controlled avatars in a virtual environment engaging in certain pattern formation behaviors. In particular, a virtual environment server may select a group of avatars and determine a pattern formed by the selected group. The virtual environment system may further search for a match of the pattern from a library of prohibited words or symbols. In the event the pattern formed by the selected avatars matches one of the prohibited patterns, some form of remedial action may be performed to disrupt the pattern.
    Type: Grant
    Filed: August 27, 2010
    Date of Patent: December 16, 2014
    Assignee: Disney Enterprises, Inc.
    Inventor: Kevin Yockey
  • Publication number: 20140355896
    Abstract: According to an embodiment, an image processing apparatus selects as an output image a candidate character component, from which a non-character component is removed, in a gradation having the largest number of pixels when there is a significant difference between the number of character pixels in the gradation having the largest number of character pixels and the number of character pixels in a gradation having the second largest number of character pixels, and selects as an output image a candidate character component, from which the non-character component is removed, in a gradation having the smallest number of edge pixels when there is no significant difference between the number of character pixels in the gradation having the largest number of character pixels and the number of character pixels in the gradation having the second largest number of character pixels.
    Type: Application
    Filed: August 18, 2014
    Publication date: December 4, 2014
    Inventor: Kunio Osada
  • Patent number: 8897579
    Abstract: A computer-implemented method of managing information is disclosed. The method can include receiving a message from a mobile device configured to connect to a mobile device network (the message including a digital image taken by the mobile device and including information corresponding to words), determining the words from the digital image information using optical character recognition, indexing the digital image based on the words, and storing the digital image for later retrieval of the digital image based on one or more received search terms.
    Type: Grant
    Filed: October 9, 2013
    Date of Patent: November 25, 2014
    Assignee: Google Inc.
    Inventors: Krishnendu Chaudhury, Ashutosh Garg, Prasenjit Phukan, Arvind Saraf
  • Patent number: 8879853
    Abstract: Provided is a system for enumerating local alignments, comprising a local alignment enumeration module for enumerating local alignments, in a case where the cell of a second matrix corresponding to a cell at a transition source indicates that the cell belongs to a local alignment, and in a case where the maximum score calculated is larger than the predetermined value, registers, in the cell of the second matrix corresponding to the cell to be calculated, an identifier of a local alignment registered in the cell of the second matrix corresponding to the cell at the transition source, and further, in a case where the maximum score calculated is larger than a maximum score of the cells belonging to the same local alignment, stores the cell to be calculated as an end point of the local alignment.
    Type: Grant
    Filed: August 31, 2011
    Date of Patent: November 4, 2014
    Assignee: Hitachi, Ltd.
    Inventor: Makoto Iwayama
  • Patent number: 8872979
    Abstract: Techniques are presented for analyzing audio-video segments, usually from multiple sources. A combined similarity measure is determined from text similarities and video similarities. The text and video similarities measure similarity between audio-video scenes for text and video, respectively. The combined similarity measure is then used to determine similar scenes in the audio-video segments. When the audio-video segments are from multiple audio-video sources, the similar scenes are common scenes in the audio-video segments. Similarities may be converted to or measured by distance. Distance matrices may be determined by using the similarity matrices. The text and video distance matrices are normalized before the combined similarity matrix is determined. Clustering is performed using distance values determined from the combined similarity matrix.
    Type: Grant
    Filed: May 21, 2002
    Date of Patent: October 28, 2014
    Assignee: Avaya Inc.
    Inventors: Amit Bagga, Jianying Hu, Jialin Zhong
  • Publication number: 20140307973
    Abstract: Text recognition techniques are described. In one or more implementations, image data is received via a network at a service provider. One or more image deblurring or curve correction techniques are applied to the image data, text is recognized from the deblurred image data using one or more optical character recognition techniques, and the recognized text is exposed for access via the network.
    Type: Application
    Filed: April 10, 2013
    Publication date: October 16, 2014
    Applicant: Adobe Systems Incorporated
    Inventor: Barry Young
  • Patent number: 8861858
    Abstract: Methods and devices for providing companion services to video are described. In one example embodiment, the method includes: identifying text contained within a video; determining, by performing pattern matching, if the identified text in the video contains actionable text; and if the identified text in the video contains actionable text, providing access to one or more features based on the actionable text.
    Type: Grant
    Filed: June 1, 2012
    Date of Patent: October 14, 2014
    Assignee: Blackberry Limited
    Inventor: Neil Patrick Adams
  • Patent number: 8855424
    Abstract: A word recognition method in which as a result of a recognition process performed on an image of a character string, one or more character candidates are obtained for each of characters forming the character string, according to which a word corresponding to the character string is recognized using a word database having registered therein a plurality of words includes setting a predetermined number of words included in the word database, as initial word candidates, performing a process in which the characters forming the recognition target character string are set as processing targets, one character by one character, and every time a processing target character is set, word candidates present at a time of the setting are narrowed down to words in which character candidates obtained for the processing target character are arranged at a same location as a location where the processing target character is arranged in the recognition target character string, and identifying, when a narrowing-down process perfor
    Type: Grant
    Filed: October 29, 2010
    Date of Patent: October 7, 2014
    Assignee: OMRON Corporation
    Inventor: Tomoyoshi Aizawa
  • Patent number: 8839118
    Abstract: A method may include receiving a plurality of images of a face of a user, wherein each image of the face of the user is associated with one of a plurality of different expressions. The method may also include receiving a first video, wherein the first video comprises images of an actor including images of a face of the actor, wherein each image of the face of the actor is associated with one of the plurality of different expressions. Further, the method may include replacing the images of the face of the actor with corresponding images of the face of the user in a same expression to create a second video and transmitting the second video to the user.
    Type: Grant
    Filed: June 30, 2010
    Date of Patent: September 16, 2014
    Assignee: Verizon Patent and Licensing Inc.
    Inventors: Sameer Vasant Gavade, Venkata S. Adimatyam
  • Patent number: 8831365
    Abstract: A system for processing a text capture operation is described. The system receives text captured from a rendered document in the text capture operation. The system also receives supplemental information distinct from the captured text. The system determines an action to perform in response to the text capture operation based upon both the captured text and the supplemental information.
    Type: Grant
    Filed: March 11, 2013
    Date of Patent: September 9, 2014
    Assignee: Google Inc.
    Inventors: Martin Towle King, Dale L. Grover, Clifford A. Kushler, James Quentin Stafford-Fraser
  • Patent number: 8830241
    Abstract: Conversion of text-based images to vector graphics (VG) is disclosed. The text-based images may include images of equations, custom typefaces, or other types of text that may not be included in a font selection of an optical character recognition (OCR) device or an application stored on a viewing device. A textual image may be converted from a raster graphics (RG) image to a VG image, which may enable resizing and alignment of the VG image with body text. In some aspects, the server may determine a body size of a reference character in the VG image. The server may determine a baseline of the VG image that may be used to align the image with the body text.
    Type: Grant
    Filed: November 30, 2009
    Date of Patent: September 9, 2014
    Assignee: Amazon Technologies, Inc.
    Inventor: Martin Gorner
  • Patent number: 8831364
    Abstract: An information processing apparatus of the present invention selects one language group, then selects one language from the selected language group, and performs OCR processing appropriate for the selected language on characters included in an image. From an obtained OCR processing result, a matching degree indicating a degree of similarity between the recognized characters in the image and the language selected for the OCR processing is calculated. Then, in a case where the matching degree is equal to or smaller than a particular value, a language belonging to a different language group is selected to further perform OCR processing. The efficiency of the OCR processing is improved. The information processing apparatus of the present invention allows improvement in the efficiency of the OCR processing.
    Type: Grant
    Filed: February 1, 2013
    Date of Patent: September 9, 2014
    Assignee: Canon Kabushiki Kaisha
    Inventor: Hiromasa Kawasaki
  • Patent number: 8825670
    Abstract: When a search keyword is characteristically used in an image, it is detected more easily.
    Type: Grant
    Filed: February 28, 2011
    Date of Patent: September 2, 2014
    Assignee: Rakuten, Inc.
    Inventor: Soh Masuko
  • Patent number: 8818111
    Abstract: Provided are an age estimation apparatus, an age estimation method, and an age estimation program capable of reducing the labor of labeling the image data used for age estimation. An age estimation apparatus for estimating an age of a person on image data includes a dimension compression unit for applying dimension compression to the image data to output low dimensional data; a clustering unit for performing clustering of the low dimensional data outputted; a labeling unit for labeling representative data of each cluster among the low dimensional data clustered; and an identification unit for estimating an age of a person on the basis of a learning result using a feature amount contained in labeled low dimensional data and unlabeled low dimensional data.
    Type: Grant
    Filed: April 14, 2010
    Date of Patent: August 26, 2014
    Assignees: NEC Soft, Ltd., Tokyo Institute of Technology
    Inventors: Kazuya Ueki, Masashi Sugiyama, Yasuyuki Ihara
  • Publication number: 20140226910
    Abstract: Techniques for identifying prohibited information within an image are described. For example, a machine accesses an image that depicts an item. The image may include prohibited information that is disallowed in accordance with a policy. The machine identifies the prohibited information within the image based on an analysis of the image. The machine initiates a response to the prohibited information based on the identifying of the prohibited information within the image.
    Type: Application
    Filed: April 17, 2014
    Publication date: August 14, 2014
    Applicant: eBay Inc.
    Inventors: Sanjay Pundlkrao Ghatare, Ali Dasdan, Naren Chittar
  • Patent number: 8805090
    Abstract: Systems and methods for measuring consistency between two objects based upon a rank of object elements instead of based upon the values of those object elements. Objects being compared can be represented by d-dimension feature vectors, U and V, where each dimension includes an associated value. U and V can be converted to rank vectors, P and Q, where values of U and V dimensions are replaced by an ordered rank or a function thereof. Analysis directed to the consistency between U and V can be accomplished by determining consistency between P and Q, which can be more efficient and more accurate, particularly with regard to illumination-invariant comparisons.
    Type: Grant
    Filed: February 7, 2012
    Date of Patent: August 12, 2014
    Assignee: Google Inc.
    Inventors: Jay Yagnik, Sergey Ioffe
  • Patent number: 8805095
    Abstract: A method for analyzing a character string, the method including: analyzing a character string to determine one of more characters of the character string; determining from a dictionary source, an alternative character string to the analyzed character string; comparing the analyzed character string with the alternative character string to determine a weighting factor for each of the characters of the analyzed character string relative to the positional arrangement of the characters in the alternative character string; and for each determined weighting factor, generating for each of the characters in the analyzed character string a corresponding character of a particular size as determined by the weighting factor.
    Type: Grant
    Filed: September 27, 2011
    Date of Patent: August 12, 2014
    Assignee: International Business Machines Corporation
    Inventor: Flemming Boegelund
  • Patent number: 8804141
    Abstract: A character output device includes: a character direction specification unit that specifies a drawing direction of characters for each processing-target region of a processing-target page; a character rotation determination unit that determines as to whether the processing-target page rotates or not based on a relation between the specified drawing direction and a direction of an output medium; and an output unit that performs an output processing based on the determination by the character rotation determination unit.
    Type: Grant
    Filed: July 27, 2009
    Date of Patent: August 12, 2014
    Assignee: Fuji Xerox Co., Ltd.
    Inventor: Masaya Miyazaki
  • Publication number: 20140219571
    Abstract: Provided are a method, computer program product and system for reporting time-based sentiment for a product. Text analysis is performed on at least one communication. At least one feature for the product is determined based on the text analysis. A sentiment value is generated for the at least one feature for the product. A date associated with the sentiment value is determined, and the sentiment value is reported for at least one feature over time.
    Type: Application
    Filed: February 4, 2013
    Publication date: August 7, 2014
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Dhruv A. BHATT, Kristin E. MCNEIL, Nitaben A. PATEL
  • Publication number: 20140212038
    Abstract: A method of detection of numbered captions in a document includes receiving a document including a sequence of document pages and identifying illustrations on pages of the document. For each identified illustration, associated text is identified. An imitation page is generated for each of the identified illustrations, each imitation page comprising a single illustration and its associated text. For a sequence of the imitation pages, a sequence of terms is identified. Each term is derived from a text fragment of the associate text of a respective imitation page. The terms of a sequence complying with at least one predefined numbering scheme which defines a form and an incremental state of the terms in a sequence. The terms of the identified sequence of terms are construed as being at least a part of a numbered caption for a respective illustration in the document.
    Type: Application
    Filed: January 29, 2013
    Publication date: July 31, 2014
    Applicant: XEROX CORPORATION
    Inventors: Herve Dejean, Jean-Luc Meunier
  • Publication number: 20140212056
    Abstract: A code that stores a history of what has been done to it and where it has been. The history can be stored in a local memory. The code can be changed based on that history.
    Type: Application
    Filed: March 31, 2014
    Publication date: July 31, 2014
    Applicant: HARRIS TECHNOLOGY, LLC
    Inventor: Scott C. Harris
  • Patent number: 8793162
    Abstract: An action plan data structure for one or more selected rendered documents is described. The data structure contains information specifying an action to perform automatically in response to a text capture from any of the selected rendered documents.
    Type: Grant
    Filed: May 5, 2010
    Date of Patent: July 29, 2014
    Assignee: Google Inc.
    Inventors: Martin T. King, Dale L. Grover, Clifford A. Kushler, James Q. Stafford-Fraser
  • Publication number: 20140207631
    Abstract: A system and method for management and processing a plurality of types of invoices at a user's site involving importing the plurality of types of invoices to provide comparable invoices and auditing the comparable invoices by performing an automated reasonability test on the comparable invoices. The system and method also provide a means for approving, processing and reporting on the comparable invoices.
    Type: Application
    Filed: January 23, 2013
    Publication date: July 24, 2014
    Inventor: Jason M. Fisher
  • Patent number: 8787681
    Abstract: Various embodiments of the invention provide systems and methods for classifying physical documents that have been converted to digital documents. Specifically, some embodiments are configured to classify digital documents that belong to a document classification whose representative members that lack structure or have varying structure, either of which makes automatic classification of such documents using conventional methods difficult. For example, certain systems and methods according to the invention can be used to classify physical real estate documents that have been converted to digital real estate documents, especially those that lack a discernable document structure.
    Type: Grant
    Filed: March 21, 2011
    Date of Patent: July 22, 2014
    Assignee: First American Data Tree LLC
    Inventor: Christopher Lawrence Rubio
  • Patent number: 8787673
    Abstract: A system and method is provided for automatically recognizing building numbers in street level images. In one aspect, a processor selects a street level image that is likely to be near an address of interest. The processor identifies those portions of the image that are visually similar to street numbers, and then extracts the numeric values of the characters displayed in such portions. If an extracted value corresponds with the building number of the address of interest such as being substantially equal to the address of interest, the extracted value and the image portion are displayed to a human operator. The human operator confirms, by looking at the image portion, whether the image portion appears to be a building number that matches the extracted value. If so, the processor stores a value that associates that building number with the street level image.
    Type: Grant
    Filed: July 12, 2011
    Date of Patent: July 22, 2014
    Assignee: Google Inc.
    Inventors: Bo Wu, Alessandro Bissacco, Raymond W. Smith, Kong man Cheung, Andrea Frome, Shlomo Urbach
  • Patent number: 8781258
    Abstract: An image processing apparatus, includes: a transforming area setter, operable to set at least a part of an area of a target image as a transforming area; a transforming area divider, operable to arrange a plurality of dividing points in the transforming area and to divide the transforming area into a plurality of small areas by using a line connecting the dividing points; and a transforming processor, operable to move a position of at least one of the dividing points to transform at least one of the small areas, thereby transforming an image in the transforming area.
    Type: Grant
    Filed: March 25, 2008
    Date of Patent: July 15, 2014
    Assignee: Seiko Epson Corporation
    Inventors: Ikuo Hayaishi, Akio Yamazaki
  • Patent number: 8768061
    Abstract: A method and system are disclosed for post optical character recognition font size determination. Optical character recognition output from an optical character recognition engine that includes character and bounding box information is aggregated into character strings. Measurements are then collected from each character in each character string that correspond to alignment heights of the top or bottom of the character with an ascender-line, a cap-line, a digit-line, a mean-line, a base-line, or a descender-line. Histograms are formed for each of these heights for each character string from the collected measurements. Based on the histograms, a pivot height is selected and used to determine the relative font size of the character string. The relative font size is normalized using a preselected factor associated with the selected pivot height. The normalized font size is then output as the font size of characters in the optical character recognition output.
    Type: Grant
    Filed: May 2, 2012
    Date of Patent: July 1, 2014
    Assignee: Xerox Corporation
    Inventor: Jean-Luc Meunier
  • Patent number: 8768047
    Abstract: Script detection service techniques are described. In an implementation, a corpora of text is analyzed to determine which strings in the corpora of text are to be included in a targeted dictionary that is usable for language detection services. The targeted dictionary is populated with strings that are individually associated with a human language. The strings include individual text characters associated with values that correspond to a particular subset of values in a table that associates subsets of values with individual human writing systems.
    Type: Grant
    Filed: July 9, 2013
    Date of Patent: July 1, 2014
    Assignee: Microsoft Corporation
    Inventors: Dimiter Georgiev, Shenghua Ye, Gerardo Villarreal Guzman, Kieran Snyder, Ryan M. Cavalcante, Tarek M. M. Sayed, Yaniv Feinberg, Yung-Shin Lin
  • Patent number: 8761435
    Abstract: Systems, devices, features, and methods for detecting geographic features in images, such as, for example, to develop a navigation database are disclosed. For example, a method of detecting text from collected images includes collecting a plurality of images of geographic areas along a road or path. An image of the plurality of images is selected. Components that represent a feature about the road or path in the selected image are determined. In one embodiment, the components are independent or invariant to scale of the feature. The determined components are compared to reference components in a data library. If the determined components substantially match with the reference components, the feature in the selected image is identified to be a text character (e.g., of a road sign) corresponding to at least some of the reference components in the data library.
    Type: Grant
    Filed: June 24, 2009
    Date of Patent: June 24, 2014
    Assignee: Navteq B.V.
    Inventor: Xin Chen
  • Publication number: 20140161365
    Abstract: An electronic device and method identify regions that are likely to be text in a natural image or video frame, followed by processing as follows: lines that are nearly vertical are automatically identified in a selected text region, oriented relative to the vertical axis within a predetermined range ?max_theta to +max_theta, followed by determination of an angle ? of the identified lines, followed by use of the angle ? to perform perspective correction by warping the selected text region. After perspective correction in this manner, each text region is processed further, to recognize text therein, by performing OCR on each block among a sequence of blocks obtained by slicing the potential text region. Thereafter, the result of text recognition is used to display to the user, either the recognized text or any other information obtained by use of the recognized text.
    Type: Application
    Filed: March 15, 2013
    Publication date: June 12, 2014
    Applicant: QUALCOMM Incorporated
    Inventors: Hemanth P. Acharya, Pawan Kumar Baheti
  • Publication number: 20140141836
    Abstract: An improved method for entering text or objects into fields is provided. Instead of a keyboard, a viewfinder provides text segmenting, text selecting and text recognizing (optical character recognition—OCR) functionalities. Text at a marker (e.g., a cursor or crosshairs) associated with the viewfinder is recognized and insertion of the recognized text is performed. The current frame is generally not captured by a user. As the user moves the camera to position a new word at the marker, the view finder is updated to provide results of recognition associated with the new word. A user is able to identify an area of interest, select text or other object of interest, and insert the same into one or more fields. The viewfinder may operate in conjunction with a camera of the electronic device on which the viewfinder is operating. Other mechanisms and variations are described.
    Type: Application
    Filed: November 7, 2012
    Publication date: May 22, 2014
    Applicant: ABBYY SOFTWARE LTD.
    Inventors: Michael Rozumyanskiy, Boris Samoylov
  • Publication number: 20140133767
    Abstract: A method for converting digital images to words includes receiving a digital image comprising text, generating a binary image from the digital image for each of N binarization threshold values to provide N binary images, converting each of the N binary images to text, and aligning the text from the N binary images to provide a word lattice for the digital image. Aligning the text may include prioritizing the text from the N binary images according to error rates on a training set. The training set may be a synthetic training set. An apparatus corresponding to the above method is also disclosed herein.
    Type: Application
    Filed: November 8, 2013
    Publication date: May 15, 2014
    Applicant: Brigham Young University
    Inventors: William B. Lund, Eric K. Ringger
  • Patent number: 8717377
    Abstract: A shell texture image shared for use among multilayer shell polygons and including a plurality of areas having transparency different from one another, is distorted to varying degrees depending on a position of each layer of the shell polygons, so as to texture-map each layer of the shell polygons therewith and so as to position the multilayer shell polygons in a virtual three-dimensional space. Thus, it is possible to realistically represent hair, grass, and the like with little effort in three-dimensional image processing.
    Type: Grant
    Filed: August 23, 2007
    Date of Patent: May 6, 2014
    Assignee: Nintendo Co., Ltd.
    Inventors: Yasuki Tawaraishi, Yusuke Kurahashi
  • Patent number: 8705878
    Abstract: An imaging apparatus configured to associate character information detected from an imaged picture displayed on a screen with a recorded picture displayed on the same screen including an imaging unit configured to image a subject and output an imaged picture, a picture recording unit configured to record pictures, a playback unit configured to play back a recorded picture recorded in the picture recording unit, and a character information detection unit configured to detect character information from the imaged picture outputted by the imaging unit.
    Type: Grant
    Filed: March 25, 2009
    Date of Patent: April 22, 2014
    Assignee: Sony Corporation
    Inventor: Tatsunobu Koike
  • Patent number: 8699794
    Abstract: Using methods, computer-readable storage media, and apparatuses for computer-implemented processing, a passage of text may be variably rendered. For each glyph in the passage of text, a glyph representation is varied according to a geometric transformation that was determined from statistical measurements of at least one geometric property from an ensemble of representations of the current glyph. Each varied glyph representation is included in renderable output data, such that when the passage of text is rendered to an output device, a given rendered representation of a given glyph subtly differs from other rendered representations of the given glyph.
    Type: Grant
    Filed: January 7, 2013
    Date of Patent: April 15, 2014
    Assignee: Gracious Eloise, Inc.
    Inventors: Eloise Bune D'Agostino, Michael Bennett D'Agostino, Bryan Michael Minor, Tamas Frajka, Michel Francois Pettigrew
  • Patent number: 8699789
    Abstract: A training system, training method, and a system and method of use of a trained classification system are provided. A classifier may be trained with a first “cheap” view but not using a second “costly” view of each of the training samples, which is not available at test time. The two views of samples are each defined in a respective original feature space. An embedding function is learned for embedding at least the first view of the training samples into a common feature space in which the second view can also be embedded or is the same as the second view original feature space. Labeled training samples (first view only) for training the classifier are embedded into the common feature space using the learned embedding function. The trained classifier can be used to predict labels for test samples for which the first view has been embedded in the common feature space with the embedding function.
    Type: Grant
    Filed: September 12, 2011
    Date of Patent: April 15, 2014
    Assignee: Xerox Corporation
    Inventors: Albert Gordo, Florent C. Perronnin