Patents by Inventor Yun Zhai
Yun Zhai has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9299162Abstract: Multi-mode video event indexing includes determining a quality of object distinctiveness with respect to images from a video stream input. A high-quality analytic mode is selected from multiple modes and applied to video input images via a hardware device to determine object activity within the video input images if the determined level of detected quality of object distinctiveness meets a threshold level of quality, else a low-quality analytic mode is selected and applied to the video input images via a hardware device to determine object activity within the video input images, wherein the low-quality analytic mode is different from the high-quality analytic mode.Type: GrantFiled: July 31, 2015Date of Patent: March 29, 2016Assignee: International Business Machines CorporationInventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Arun Hampapur, Yun Zhai
-
Patent number: 9299229Abstract: Techniques for detecting one or more events are provided. The techniques include identifying one or more segments in a video sequence as one or more candidates for one or more events by a temporal ordering of the one or more candidates, and analyzing one or more motion patterns of the one or more candidates to detect the one or more events.Type: GrantFiled: November 29, 2008Date of Patent: March 29, 2016Assignee: Toshiba Global Commerce Solutions Holdings CorporationInventors: Russell Patrick Bobbitt, Quanfu Fan, Arun Hampapur, Frederik Kjeldsen, Sharathchandra Umapathirao Pankanti, Akira Yanagawa, Yun Zhai
-
Publication number: 20160034766Abstract: Transaction units of video data and transaction data captured from different checkout lanes are prioritized as a function of lane priority values of respective ones of the different checkout lanes from which the transaction units are acquired. Each of the checkout lanes has a different lane priority value. The individual transaction units are processed in the prioritized processing order to automatically detect irregular activities indicated by the transaction unit video and the transaction data of the processed individual transaction units.Type: ApplicationFiled: October 16, 2015Publication date: February 4, 2016Inventors: RUSSELL P. BOBBITT, QUANFU FAN, SACHIKO MIYAZAWA, SHARATHCHANDRA U. PANKANTI, YUN ZHAI
-
Patent number: 9251425Abstract: Automatic object retrieval from input video is based on learned, complementary detectors created for each of a plurality of different motionlet clusters. The motionlet clusters are partitioned from a dataset of training vehicle images as a function of determining that vehicles within each of the scenes of the images in each cluster share similar two-dimensional motion direction attributes within their scenes. To train the complementary detectors, a first detector is trained on motion blobs of vehicle objects detected and collected within each of the training dataset vehicle images within the motionlet cluster via a background modeling process; a second detector is trained on each of the training dataset vehicle images within the motionlet cluster that have motion blobs of the vehicle objects but are misclassified by the first detector; and the training repeats until all of the training dataset vehicle images have been eliminated as false positives or correctly classified.Type: GrantFiled: February 12, 2015Date of Patent: February 2, 2016Assignee: International Business Machines CorporationInventors: Ankur Datta, Rogerio S. Feris, Sharathchandra U. Pankanti, Yun Zhai
-
Publication number: 20150379357Abstract: Local models learned from anomaly detection are used to rank detected anomalies. The local models include image feature values extracted from an image field of video image data with respect to different predefined spatial and temporal local units, wherein anomaly results are determined by failures to fit to applied anomaly detection module local models. Image features values extracted from the image field local units associated with anomaly results are normalized, and image feature values extracted from the image field local units are clustered. Weights for anomaly results are learned as a function of the relations of the normalized extracted image feature values to the clustered image feature values. The normalized values are multiplied by the learned weights to generate ranking values to rank the anomalies.Type: ApplicationFiled: September 4, 2015Publication date: December 31, 2015Inventors: ANKUR DATTA, BALAMANOHAR PALURI, SHARATHCHANDRA U. PANKANTI, YUN ZHAI
-
Publication number: 20150379729Abstract: Field of view overlap among multiple cameras are automatically determined as a function of the temporal overlap of object tracks determined within their fields-of-view. Object tracks with the highest similarity value are assigned into pairs, and portions of the assigned object track pairs having a temporally overlapping period of time are determined. Scene entry points are determined from object locations on the tracks at a beginning of the temporally overlapping period of time, and scene exit points from object locations at an ending of the temporally overlapping period of time. Boundary lines for the overlapping fields-of-view portions within the corresponding camera fields-of-view are defined as a function of the determined entry and exit points in their respective fields-of-view.Type: ApplicationFiled: September 14, 2015Publication date: December 31, 2015Inventors: Ankur Datta, Rogerio S. Feris, Sharathchandra U. Pankanti, Yun Zhai
-
Publication number: 20150379768Abstract: Objects within two-dimensional video data are modeled by three-dimensional models as a function of object type and motion through manually calibrating a two-dimensional image to the three spatial dimensions of a three-dimensional modeling cube. Calibrated three-dimensional locations of an object in motion in the two-dimensional image field of view of a video data input are determined and used to determine a heading direction of the object as a function of the camera calibration and determined movement between the determined three-dimensional locations. The two-dimensional object image is replaced in the video data input with an object-type three-dimensional polygonal model having a projected bounding box that best matches a bounding box of an image blob, the model oriented in the determined heading direction. The bounding box of the replacing model is then scaled to fit the object image blob bounding box, and rendered with extracted image features.Type: ApplicationFiled: September 9, 2015Publication date: December 31, 2015Inventors: Ankur Datta, Rogerio S. Feris, Yun Zhai
-
Patent number: 9224049Abstract: Foreground object image features are extracted from input video via application of a background subtraction mask, and optical flow image features from a region of the input video image data defined by the extracted foreground object image features. If estimated movement features indicate that the underlying object is in motion, a dominant moving direction of the underlying object is determined. If the dominant moving direction is parallel to an orientation of the second, crossed thoroughfare, an event alarm indicating that a static object is blocking travel on the crossing second thoroughfare is not generated. If the estimated movement features indicate that the underlying object is static, or that its determined dominant moving direction is not parallel to the second thoroughfare, an appearance of the foreground object region is determined and a static-ness timer run while the foreground object region comprises the extracted foreground object image features.Type: GrantFiled: March 5, 2015Date of Patent: December 29, 2015Assignee: International Business Machines CorporationInventors: Rogerio S. Feris, Yun Zhai
-
Publication number: 20150356352Abstract: Long-term understanding of background modeling includes determining first and second dimension gradient model derivatives of image brightness data of an image pixel along respective dimensions of two-dimensional, single channel image brightness data of a static image scene. The determined gradients are averaged with previous determined gradients of the image pixels, and with gradients of neighboring pixels as a function of their respective distances to the image pixel, the averaging generating averaged pixel gradient models for each of a plurality of pixels of the video image data of the static image scene that each have mean values and weight values. Background models for the static image scene are constructed as a function of the averaged pixel gradients and weights, wherein the background model pixels are represented by averaged pixel gradient models having similar orientation and magnitude and weights meeting a threshold weight requirement.Type: ApplicationFiled: August 11, 2015Publication date: December 10, 2015Inventors: Rogerio S. Feris, Yun Zhai
-
Publication number: 20150356745Abstract: Multi-mode video event indexing includes determining a quality of object distinctiveness with respect to images from a video stream input. A high-quality analytic mode is selected from multiple modes and applied to video input images via a hardware device to determine object activity within the video input images if the determined level of detected quality of object distinctiveness meets a threshold level of quality, else a low-quality analytic mode is selected and applied to the video input images via a hardware device to determine object activity within the video input images, wherein the low-quality analytic mode is different from the high-quality analytic mode.Type: ApplicationFiled: August 19, 2015Publication date: December 10, 2015Inventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Arun Hampapur, Yun Zhai
-
Publication number: 20150339831Abstract: Multi-mode video event indexing includes determining a quality of object distinctiveness with respect to images from a video stream input. A high-quality analytic mode is selected from multiple modes and applied to video input images via a hardware device to determine object activity within the video input images if the determined level of detected quality of object distinctiveness meets a threshold level of quality, else a low-quality analytic mode is selected and applied to the video input images via a hardware device to determine object activity within the video input images, wherein the low-quality analytic mode is different from the high-quality analytic mode.Type: ApplicationFiled: July 31, 2015Publication date: November 26, 2015Inventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Arun Hampapur, Yun Zhai
-
Patent number: 9197868Abstract: Transaction units of video data and transaction data captured from different checkout lanes are prioritized as a function of lane priority values of respective ones of the different checkout lanes from which the transaction units are acquired. Each of the checkout lanes has a different lane priority value. The individual transaction units are processed in the prioritized processing order to automatically detect irregular activities indicated by the transaction unit video and the transaction data of the processed individual transaction units.Type: GrantFiled: September 10, 2013Date of Patent: November 24, 2015Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Russell P. Bobbitt, Quanfu Fan, Sachiko Miyazawa, Sharathchandra U. Pankanti, Yun Zhai
-
Patent number: 9165375Abstract: Field of view overlap among multiple cameras are automatically determined as a function of the temporal overlap of object tracks determined within their fields-of-view. Object tracks with the highest similarity value are assigned into pairs, and portions of the assigned object track pairs having a temporally overlapping period of time are determined. Scene entry points are determined from object locations on the tracks at a beginning of the temporally overlapping period of time, and scene exit points from object locations at an ending of the temporally overlapping period of time. Boundary lines for the overlapping fields-of-view portions within the corresponding camera fields-of-view are defined as a function of the determined entry and exit points in their respective fields-of-view.Type: GrantFiled: November 4, 2014Date of Patent: October 20, 2015Assignee: International Business Machines CorporationInventors: Ankur Datta, Rogerio S. Feris, Sharathchandra U. Pankanti, Yun Zhai
-
Patent number: 9158972Abstract: Objects within two-dimensional video data are modeled by three-dimensional models as a function of object type and motion through manually calibrating a two-dimensional image to the three spatial dimensions of a three-dimensional modeling cube. Calibrated three-dimensional locations of an object in motion in the two-dimensional image field of view of a video data input are determined and used to determine a heading direction of the object as a function of the camera calibration and determined movement between the determined three-dimensional locations. The two-dimensional object image is replaced in the video data input with an object-type three-dimensional polygonal model having a projected bounding box that best matches a bounding box of an image blob, the model oriented in the determined heading direction. The bounding box of the replacing model is then scaled to fit the object image blob bounding box, and rendered with extracted image features.Type: GrantFiled: September 5, 2014Date of Patent: October 13, 2015Assignee: International Business Machines CorporationInventors: Ankur Datta, Rogerio S. Feris, Yun Zhai
-
Patent number: 9158976Abstract: Local models learned from anomaly detection are used to rank detected anomalies. The local models include image feature values extracted from an image field of video image data with respect to different predefined spatial and temporal local units, wherein anomaly results are determined by failures to fit to applied anomaly detection module local models. Image features values extracted from the image field local units associated with anomaly results are normalized, and image feature values extracted from the image field local units are clustered. Weights for anomaly results are learned as a function of the relations of the normalized extracted image feature values to the clustered image feature values. The normalized values are multiplied by the learned weights to generate ranking values to rank the anomalies.Type: GrantFiled: May 18, 2011Date of Patent: October 13, 2015Assignee: International Business Machines CorporationInventors: Ankur Datta, Balamanohar Paluri, Sharathchandra U. Pankanti, Yun Zhai
-
Publication number: 20150278631Abstract: Machine logic that pre-processes and post-processes images for visual object detection by performing the following steps: receiving a set of image(s); filtering the set of image(s) using a set of multimodal integral filter(s), thereby removing at least a portion of the set of image(s) and resulting in a filtered set of image(s); performing object detection on the filtered set of image(s) to generate a set of object-detected image(s); assembling a first plurality of object-detected image(s) from the set of object-detected image(s); and upon assembling the first plurality of object-detected image(s), performing non-maximum suppression on the assembled first plurality of object-detected image(s).Type: ApplicationFiled: March 23, 2015Publication date: October 1, 2015Inventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Sharathchandra U. Pankanti, Yun Zhai
-
Patent number: 9147259Abstract: Multi-mode video event indexing includes determining a quality of object distinctiveness with respect to images from a video stream input. A high-quality analytic mode is selected from multiple modes and applied to video input images via a hardware device to determine object activity within the video input images if the determined level of detected quality of object distinctiveness meets a threshold level of quality, else a low-quality analytic mode is selected and applied to the video input images via a hardware device to determine object activity within the video input images, wherein the low-quality analytic mode is different from the high-quality analytic mode.Type: GrantFiled: August 21, 2013Date of Patent: September 29, 2015Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Arun Hampapur, Yun Zhai
-
Patent number: 9129380Abstract: Long-term understanding of background modeling includes determining first and second dimension gradient model derivatives of image brightness data of an image pixel along respective dimensions of two-dimensional, single channel image brightness data of a static image scene. The determined gradients are averaged with previous determined gradients of the image pixels, and with gradients of neighboring pixels as a function of their respective distances to the image pixel, the averaging generating averaged pixel gradient models for each of a plurality of pixels of the video image data of the static image scene that each have mean values and weight values. Background models for the static image scene are constructed as a function of the averaged pixel gradients and weights, wherein the background model pixels are represented by averaged pixel gradient models having similar orientation and magnitude and weights meeting a threshold weight requirement.Type: GrantFiled: January 21, 2014Date of Patent: September 8, 2015Assignee: International Business Machines CorporationInventors: Rogerio S. Feris, Yun Zhai
-
Patent number: 9123129Abstract: Multi-mode video event indexing includes determining a quality of object distinctiveness with respect to images from a video stream input. A high-quality analytic mode is selected from multiple modes and applied to video input images via a hardware device to determine object activity within the video input images if the determined level of detected quality of object distinctiveness meets a threshold level of quality, else a low-quality analytic mode is selected and applied to the video input images via a hardware device to determine object activity within the video input images, wherein the low-quality analytic mode is different from the high-quality analytic mode.Type: GrantFiled: August 21, 2013Date of Patent: September 1, 2015Assignee: International Business Machines CorporationInventors: Russell P. Bobbitt, Lisa M. Brown, Rogerio S. Feris, Arun Hampapur, Yun Zhai
-
Publication number: 20150242692Abstract: Foreground feature data and motion feature data is determined for frames of video data acquired from a train track area region of interest. The frames are labeled as “train present” if the determined foreground feature data value meets a threshold value, else as “train absent; and as “motion present” if the motion feature data meets a motion threshold, else as “static.” The labels are used to classify segments of the video data comprising groups of consecutive video frames, namely as within a “no train present” segment for groups with “train absent” and “static” labels; within a “train present and in transition” segment for groups “train present” and “motion present” labels; and within a “train present and stopped” segment for groups with “train present” and “static” labels. The presence or motion state of a train at a time of inquiry is thereby determined from the respective segment classification.Type: ApplicationFiled: May 14, 2015Publication date: August 27, 2015Inventors: RUSSELL P. BOBBITT, ROGERIO S. FERIS, YUN ZHAI