Patents by Inventor Yongmian Zhang
Yongmian Zhang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250005779Abstract: Apparatus and method to detect concealed weapons on a person from a distance coordinates movement of a field of view (FOV) of a wide angle view (WAV) camera with movement of an FOV of a millimeter wave (MMW) camera. In an embodiment, the apparatus identifies and tracks a plurality of individuals within the WAV camera FOV. The MMW camera is focused on a first of the identified plurality of individuals, and takes an image of the individual within the MMW camera FOV. In an embodiment, the first individual is centered in the MMW camera FOV. The image is analyzed to detect whether the individual is carrying any concealed weapons. If concealed weapons are detected, the apparatus may signal a need to detain and/or apprehend the individual.Type: ApplicationFiled: June 30, 2023Publication date: January 2, 2025Inventor: Yongmian ZHANG
-
Publication number: 20240005687Abstract: Method and apparatus to identify field labels from filled forms using image processing to compare two or a few copies of the same kind of form, possibly filled out differently. Filled forms are processed to provide text strings, one for each line of text on each filled form. The text strings are converted to vectors. Vectors from different filled forms are compared to identify common words which are indicative of field labels. In an embodiment, a histogram may be generated to show frequency of occurrence of characters and words, the histogram values also being indicative of field labels.Type: ApplicationFiled: July 1, 2022Publication date: January 4, 2024Inventor: Yongmian ZHANG
-
Patent number: 11386686Abstract: Method and apparatus to match bounding boxes around text to align forms. The approach is less computationally intensive, and less prone to error than text recognition. For purposes of achieving alignment, information per se is not as important as information location. Information within the bounding boxes is not as critical as is the location of the area which the bounding boxes occupy. Scanning artifacts, missing characters, or noise generally do not affect bounding boxes themselves so much as they do the contents of the bounding boxes. Thus, for purposes of form alignment, the bounding boxes themselves are sufficient. Using bounding boxes also avoids misalignment issues that can result from stray marks on a page, for example, from holes punched in a sheet, or from handwritten notations.Type: GrantFiled: March 31, 2020Date of Patent: July 12, 2022Assignee: Konica Minolta Business Solutions U.S.A., Inc.Inventor: Yongmian Zhang
-
Publication number: 20210303839Abstract: Method and apparatus to match bounding boxes around text to align forms. The approach is less computationally intensive, and less prone to error than text recognition. For purposes of achieving alignment, information per se is not as important as information location. Information within the bounding boxes is not as critical as is the location of the area which the bounding boxes occupy. Scanning artifacts, missing characters, or noise generally do not affect bounding boxes themselves so much as they do the contents of the bounding boxes. Thus, for purposes of form alignment, the bounding boxes themselves are sufficient. Using bounding boxes also avoids misalignment issues that can result from stray marks on a page, for example, from holes punched in a sheet, or from handwritten notations.Type: ApplicationFiled: March 31, 2020Publication date: September 30, 2021Applicant: KONICA MINOLTA BUSINESS SOLUTIONS U.S.A., INC.Inventor: Yongmian Zhang
-
Patent number: 10853695Abstract: A method, a computer readable medium, and a system for cell annotation are disclosed. The method includes receiving at least one new cell image for cell detection; extracting cell features from the at least one new cell image; comparing the extracted cell features to a matrix of cell features of each class to predict a closest class, wherein the matrix of cell features has been generated from at least initial training data comprising at least one cell image; detecting cell pixels from the extracted cell features of the at least one new cell image using the predicted closest class to generate a likelihood map; extracting individual cells from the at least one cell image by segmenting the individual cells from the likelihood map; and performing a machine annotation on the extracted individual cells from the at least one new cell image to identify cells, non-cell pixels, and/or cell boundaries.Type: GrantFiled: June 27, 2017Date of Patent: December 1, 2020Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian Zhang, Jingwen Zhu
-
Patent number: 10846566Abstract: An artificial neural network system for image classification, formed of multiple independent individual convolutional neural networks (CNNs), each CNN being configured to process an input image patch to calculate a classification for the center pixel of the patch. The multiple CNNs have different receptive field of views for processing image patches of different sizes centered at the same pixel. A final classification for the center pixel is calculated by combining the classification results from the multiple CNNs. An image patch generator is provided to generate the multiple input image patches of different sizes by cropping them from the original input image. The multiple CNNs have similar configurations, and when training the artificial neural network system, one CNN is trained first, and the learned parameters are transferred to another CNN as initial parameters and the other CNN is further trained. The classification includes three classes, namely background, foreground, and edge.Type: GrantFiled: August 9, 2017Date of Patent: November 24, 2020Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Jingwen Zhu, Yongmian Zhang
-
Publication number: 20200311411Abstract: A text recognition method and system involves computing a text matching score between an input text and an output candidate text. The text matching score is computed by evaluating respective N-grams of the input text and the output candidate text. The N-grams are compared in pairs for visual similarity by determining N-gram pair scores, which are used to compute the text matching score. The N-gram pair scores are determined using a set of probabilities of confusion between characters contained in the N-grams. The described approach can address inconsistent results that arise from conventional text similarity quantifiers.Type: ApplicationFiled: March 28, 2019Publication date: October 1, 2020Inventors: Shubham AGARWAL, Yongmian ZHANG
-
Publication number: 20200311413Abstract: Image processing is performed on an input image generated from scanning a filled-in document form. The input image is evaluated against a blank version of various document forms in order to identify the form type of the filled-in document form. The evaluation results in identifying one of the blank document forms as a match to the filled-in document form. Each document form has a set of keywords. The evaluation uses a vector of keyword matches in the filled-in document form. Once a blank document form is identified to be match, the filled-in document form may be categorized according to that document form and/or data extracted from the filled-in document may be stored in association with keywords of that document form.Type: ApplicationFiled: March 28, 2019Publication date: October 1, 2020Inventors: Yongmian ZHANG, Shubham AGARWAL
-
Patent number: 10521697Abstract: A local connectivity feature transform (LCFT) is applied to binary document images containing text characters, to generate transformed document images which are then input into a bi-directional Long Short Term Memory (LSTM) neural network to perform character/word recognition. The LCFT transformed image is a gray scale image where the pixel values encode local pixel connectivity information of corresponding pixels in the original binary image. The transform is one that provides a unique transform score for every possible shape represented as a 3×3 block. In one example, the transform is computed using a 3×3 weight matrix that combines bit coding with a zigzag pattern to assign weights to each element of the 3×3 block, and by summing up the weights for the non-zero elements of the 3×3 block shape.Type: GrantFiled: September 29, 2017Date of Patent: December 31, 2019Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Shubham Agarwal, Maral Mesmakhosroshahi, Yongmian Zhang
-
Patent number: 10473919Abstract: A 3D scanning system includes a base stand, two circular arc shaped support tracks, a mounting assembly for mounting the support tracks to the base stand with one or more degrees of rotational freedom, two sensor holders mounted on the respective support track for holding two depth sensors, and a drive mechanism for driving the sensor holders to move along the respective support tracks. The mounting assembly supports relative rotation of the two support tracks and pitch and roll rotations of the support tracks. To perform a 3D scan, a stationary object is placed in front of the two depth sensors. The sensor holders are moved along the respective support tracks to different positions to obtain depth images of the objects from different angles, from which a 3D surface of the object is constructed. Prior to scanning, the two depth sensors are calibrated relative to each other.Type: GrantFiled: December 29, 2017Date of Patent: November 12, 2019Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian Zhang, Wei Ming
-
Publication number: 20190266443Abstract: In an optical character recognition (OCR) method for digitizing printed text images using a long-short term memory (LSTM) network, text images are pre-processed using a stroke-aware max-min pooling method before being fed into the network, for both network training and OCR prediction. During training, an average stroke thickness is computed from the training dataset. Stroke-aware max-min pooling is applied to each text line image, where minimum pooling is applied if the stroke thickness of the line is greater than the average stroke thickness, while max pooling is applied if the stroke thickness is less than or equal to the average stroke thickness. The pooled images are used for network training. During prediction, stroke-aware max-min pooling is applied to each input text line image, and the pooled image is fed to the trained LSTM network to perform character recognition.Type: ApplicationFiled: February 28, 2018Publication date: August 29, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian Zhang, Shubham Agarwal
-
Patent number: 10373022Abstract: In an optical character recognition (OCR) method for digitizing printed text images using a long-short term memory (LSTM) network, text images are pre-processed using a stroke-aware max-min pooling method before being fed into the network, for both network training and OCR prediction. During training, an average stroke thickness is computed from the training dataset. Stroke-aware max-min pooling is applied to each text line image, where minimum pooling is applied if the stroke thickness of the line is greater than the average stroke thickness, while max pooling is applied if the stroke thickness is less than or equal to the average stroke thickness. The pooled images are used for network training. During prediction, stroke-aware max-min pooling is applied to each input text line image, and the pooled image is fed to the trained LSTM network to perform character recognition.Type: GrantFiled: February 28, 2018Date of Patent: August 6, 2019Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian Zhang, Shubham Agarwal
-
Publication number: 20190236411Abstract: An artificial neural network system for image classification, formed of multiple independent individual convolutional neural networks (CNNs), each CNN being configured to process an input image patch to calculate a classification for the center pixel of the patch. The multiple CNNs have different receptive field of views for processing image patches of different sizes centered at the same pixel. A final classification for the center pixel is calculated by combining the classification results from the multiple CNNs. An image patch generator is provided to generate the multiple input image patches of different sizes by cropping them from the original input image. The multiple CNNs have similar configurations, and when training the artificial neural network system, one CNN is trained first, and the learned parameters are transferred to another CNN as initial parameters and the other CNN is further trained. The classification includes three classes, namely background, foreground, and edge.Type: ApplicationFiled: August 9, 2017Publication date: August 1, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Jingwen ZHU, Yongmian ZHANG
-
Publication number: 20190228268Abstract: An artificial neural network system for image classification, including multiple independent individual convolutional neural networks (CNNs) connected in multiple stages, each CNN configured to process an input image to calculate a pixelwise classification. The output of an earlier stage CNN, which is a class score image having identical height and width as its input image and a depth of N representing the probabilities of each pixel of the input image belonging to each of N classes, is input into the next stage CNN as input image. When training the network system, the first stage CNN is trained using first training images and corresponding label data; then second training images are forward propagated by the trained first stage CNN to generate corresponding class score images, which are used along with label data corresponding to the second training images to train the second stage CNN.Type: ApplicationFiled: August 9, 2017Publication date: July 25, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian ZHANG, Jingwen ZHU
-
Publication number: 20190204588Abstract: A 3D scanning system includes a base stand, two circular arc shaped support tracks, a mounting assembly for mounting the support tracks to the base stand with one or more degrees of rotational freedom, two sensor holders mounted on the respective support track for holding two depth sensors, and a drive mechanism for driving the sensor holders to move along the respective support tracks. The mounting assembly supports relative rotation of the two support tracks and pitch and roll rotations of the support tracks. To perform a 3D scan, a stationary object is placed in front of the two depth sensors. The sensor holders are moved along the respective support tracks to different positions to obtain depth images of the objects from different angles, from which a 3D surface of the object is constructed. Prior to scanning, the two depth sensors are calibrated relative to each other.Type: ApplicationFiled: December 29, 2017Publication date: July 4, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Yongmian Zhang, Wei Ming
-
Publication number: 20190205758Abstract: Pathological analysis needs instance-level labeling on a histologic image with high accurate boundaries required. To this end, embodiments of the present invention provide a deep model that employs the DeepLab basis and the multi-layer deconvolution network basis in a unified model. The model is a deeply supervised network that allows to represent multi-scale and multi-level features. It achieved segmentation on the benchmark dataset at a level of accuracy which is significantly beyond all top ranking methods in the 2015 MICCAI Gland Segmentation Challenge. Moreover, the overall performance of the model surpasses the state-of-the-art Deep Multi-channel Neural Networks published most recently, and the model is structurally much simpler, more computational efficient and weight-lighted to learn.Type: ApplicationFiled: December 13, 2017Publication date: July 4, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Jingwen ZHU, Yongmian ZHANG
-
Publication number: 20190180147Abstract: A method, a computer readable medium, and a system for cell annotation are disclosed. The method includes receiving at least one new cell image for cell detection; extracting cell features from the at least one new cell image; comparing the extracted cell features to a matrix of cell features of each class to predict a closest class, wherein the matrix of cell features has been generated from at least initial training data comprising at least one cell image; detecting cell pixels from the extracted cell features of the at least one new cell image using the predicted closest class to generate a likelihood map; extracting individual cells from the at least one cell image by segmenting the individual cells from the likelihood map; and performing a machine annotation on the extracted individual cells from the at least one new cell image to identify cells, non-cell pixels, and/or cell boundaries.Type: ApplicationFiled: June 27, 2017Publication date: June 13, 2019Applicant: Konica Minolta Laboratory U.S.A., Inc.Inventors: Yongmian Zhang, Jingwen Zhu
-
Patent number: 10318803Abstract: In a text line segmentation process, connected components (CCs) in document image are categorized into three subsets (normal, large, small) based on their sizes. The centroids of the normal size CCs are used to perform line detection using Hough transform. Among the detected candidate lines, those with line bounding box heights greater than a certain height are removed. For each normal size CC, if its bounding box does not overlap the bounting box of any line with an overlap area greater than a predefined fraction of the CC bounding box, a new line is added for this CC, which passes through the centroid of the CC and has an average slant angle. Each large size CCs are broken into two or more CCs. All CCs are then assigned to the nearest lines. A refinement method is also described, which can take any text line segmentation result and refine it.Type: GrantFiled: November 30, 2017Date of Patent: June 11, 2019Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Shubham Agarwal, Yongmian Zhang
-
Publication number: 20190163971Abstract: In a text line segmentation process, connected components (CCs) in document image are categorized into three subsets (normal, large, small) based on their sizes. The centroids of the normal size CCs are used to perform line detection using Hough transform. Among the detected candidate lines, those with line bounding box heights greater than a certain height are removed. For each normal size CC, if its bounding box does not overlap the bounting box of any line with an overlap area greater than a predefined fraction of the CC bounding box, a new line is added for this CC, which passes through the centroid of the CC and has an average slant angle. Each large size CCs are broken into two or more CCs. All CCs are then assigned to the nearest lines. A refinement method is also described, which can take any text line segmentation result and refine it.Type: ApplicationFiled: November 30, 2017Publication date: May 30, 2019Applicant: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Shubham Agarwal, Yongmian Zhang
-
Patent number: 10282589Abstract: An artificial neural network system implemented on a computer for cell segmentation and classification of biological images. It includes a deep convolutional neural network as a feature extraction network, a first branch network connected to the feature extraction network to perform cell segmentation, and a second branch network connected to the feature extraction network to perform cell classification using the cell segmentation map generated by the first branch network. The feature extraction network is a modified VGG network where each convolutional layer uses multiple kernels of different sizes. The second branch network takes feature maps from two levels of the feature extraction network, and has multiple fully connected layers to independently process multiple cropped patches of the feature maps, the cropped patches being located at a centered and multiple shifted positions relative to the cell being classified; a voting method is used to determine the final cell classification.Type: GrantFiled: August 29, 2017Date of Patent: May 7, 2019Assignee: KONICA MINOLTA LABORATORY U.S.A., INC.Inventors: Maral Mesmakhosroshahi, Shubham Agarwal, Yongmian Zhang