Patents by Inventor Ophir Azulai
Ophir Azulai has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11948382Abstract: A method for synthesizing negative training data associated with training models to detect text within documents and images. The method includes one or more computer processors receiving a set of dictates associated with generating one or more negative training datasets for training a set of models to classify a plurality of features found within a data source. The method further includes identifying a set of rules related to generating negative training data to detect text based on the received set of dictates. The method further includes compiling one or more arrays of elements of hard-negative training data into a negative training data dataset based on the identified set of rules and one or more dictates. The method further includes determining metadata corresponding an array of elements of hard-negative training data.Type: GrantFiled: December 18, 2020Date of Patent: April 2, 2024Assignee: International Business Machines CorporationInventors: Ophir Azulai, Udi Barzelay
-
Publication number: 20230343124Abstract: Described are techniques for font attribute detection. The techniques include receiving a document having different font attributes amongst a plurality of words respectively comprised of at least one character. The techniques further include generating a dense image document from the document by setting the plurality of words to a predefined size, removing blank spaces from the document, and altering an order of characters relative to the document. The techniques further include determining characteristics of the characters in the dense image document and aggregating the characteristics for at least one word. The techniques further include annotating the at least one word with a font attribute based on the aggregated characteristics.Type: ApplicationFiled: April 26, 2022Publication date: October 26, 2023Inventors: Ophir Azulai, Daniel Nechemia Rotman, Udi Barzelay
-
Patent number: 11776287Abstract: An approach to identifying text within an image may be presented. The approach can receive an image. The approach can classify an image on a pixel-by-pixel basis whether the pixel is text. The approach can generate bounding boxes around groups of pixels that are classified as text. The approach can mask sections of an image that where pixels are not classified as text. The approach may be used as a pre-processing technique for optical character recognition in documents, scanned images, or still images.Type: GrantFiled: April 27, 2021Date of Patent: October 3, 2023Assignee: International Business Machines CorporationInventors: Udi Barzelay, Ophir Azulai, Inbar Shapira
-
Publication number: 20230298373Abstract: An example system includes a processor to receive detected chart regions in a page of a document. The processor is to produce, via a graphical elements detector, predicted heatmaps and bounding boxes for graphical objects in the detected chart regions. The processor is also to apply chart type specific analysis algorithm to the predicted heatmaps and bounding boxes, to extract tabular chart data. The processor can then generate an output data file and a visualization based on the predicted heatmap and the extracted tabular chart data.Type: ApplicationFiled: March 21, 2022Publication date: September 21, 2023Inventors: Joseph SHTOK, Leonid KARLINSKY, Sivan HARARY, Ophir AZULAI
-
Patent number: 11741732Abstract: In some examples, a system for detecting text in an image includes a memory device to store a text detection model trained using images of up-scaled text, and a processor configured to perform text detection on an image to generate original bounding boxes that identify potential text in the image. The processor is also configured to generate a secondary image that includes up-scaled portions of the image associated with bounding boxes below a threshold size, and perform text detection on the secondary image to generate secondary bounding boxes that identify potential text in the secondary image. The processor is also configured to compare the original bounding boxes with the secondary bounding boxes to identify original bounding boxes that are false positives, and generate an image file that includes the original bounding boxes, wherein those original bounding boxes that are identified as false positives are removed.Type: GrantFiled: December 22, 2021Date of Patent: August 29, 2023Assignee: International Business Machines CorporationInventors: Ophir Azulai, Udi Barzelay, Oshri Pesah Naparstek
-
Publication number: 20230245481Abstract: A method, computer system, and a computer program product for text detection is provided. The present invention may include training a text detection model. The present invention may include performing text detection on an inputted image using the trained text detection model. The present invention may include determining whether at least one of a plurality of bounding boxes generated using the inputted image has an aspect ratio above a threshold. The present invention may include based upon determining that at least one of the plurality of bounding boxes generated using the inputted image has the aspect ratio above the threshold, upscaling any text within the at least one bounding box and performing text detection on a new image using the trained text detection model. The present invention may include outputting an output image.Type: ApplicationFiled: January 31, 2022Publication date: August 3, 2023Inventors: Ophir Azulai, Udi Barzelay, Oshri Pesah Naparstek
-
Publication number: 20230196807Abstract: In some examples, a system for detecting text in an image includes a memory device to store a text detection model trained using images of up-scaled text, and a processor configured to perform text detection on an image to generate original bounding boxes that identify potential text in the image. The processor is also configured to generate a secondary image that includes up-scaled portions of the image associated with bounding boxes below a threshold size, and perform text detection on the secondary image to generate secondary bounding boxes that identify potential text in the secondary image. The processor is also configured to compare the original bounding boxes with the secondary bounding boxes to identify original bounding boxes that are false positives, and generate an image file that includes the original bounding boxes, wherein those original bounding boxes that are identified as false positives are removed.Type: ApplicationFiled: December 22, 2021Publication date: June 22, 2023Inventors: Ophir AZULAI, Udi BARZELAY, Oshri Pesah NAPARSTEK
-
Publication number: 20220343103Abstract: An approach to identifying text within an image may be presented. The approach can receive an image. The approach can classify an image on a pixel-by-pixel basis whether the pixel is text. The approach can generate bounding boxes around groups of pixels that are classified as text. The approach can mask sections of an image that where pixels are not classified as text. The approach may be used as a pre-processing technique for optical character recognition in documents, scanned images, or still images.Type: ApplicationFiled: April 27, 2021Publication date: October 27, 2022Inventors: Udi Barzelay, Ophir Azulai, Inbar Shapira
-
Publication number: 20220318555Abstract: Approaches presented herein enable action recognition. More specifically, a plurality of video segments having one or more action representations is received. One or more sub-action representations in the plurality of video segments are learned. An embedding in a space of a distance metric learning (DML) network for each of the one or more sub-action representations is determined. A set of respective trajectory distances between each of the one or more sub-action representations and one or more class representatives in the space of the DML network based on the embedding is computed, and the one or more action representations based on the set of respective trajectory distances are classified.Type: ApplicationFiled: March 31, 2021Publication date: October 6, 2022Inventors: Rami Ben-Ari, Ophir Azulai, Udi Barzelay, Mor Shpigel Nacson
-
Publication number: 20220198186Abstract: A method for synthesizing negative training data associated with training models to detect text within documents and images. The method includes one or more computer processors receiving a set of dictates associated with generating one or more negative training datasets for training a set of models to classify a plurality of features found within a data source. The method further includes identifying a set of rules related to generating negative training data to detect text based on the received set of dictates. The method further includes compiling one or more arrays of elements of hard-negative training data into a negative training data dataset based on the identified set of rules and one or more dictates. The method further includes determining metadata corresponding an array of elements of hard-negative training data.Type: ApplicationFiled: December 18, 2020Publication date: June 23, 2022Inventors: Ophir Azulai, Udi Barzelay
-
Patent number: 11157257Abstract: Automatic cloning of a PYTHON CONDA environment into a DOCKER image, such that at least one CONDA container that functions the same as the PYTHON CONDA environment can be started from the DOCKER image. The automatic cloning may include: First, creating a Dockerfile that comprises commands to: install a PYTHON ANACONDA environment or obtain a PYTHON ANACONDA environment image, copy the PYTHON CONDA environment into the DOCKER image, and run a CONDA command, in the ANACONDA environment, to create a cloned PYTHON CONDA environment from the copied PYTHON CONDA environment. Second, building the DOCKER image from the Dockerfile.Type: GrantFiled: January 7, 2020Date of Patent: October 26, 2021Assignee: International Business Machines CorporationInventors: Ophir Azulai, Ofer Lavi, Eran Raichstein
-
Patent number: 11115661Abstract: A method of disarming and reconstructing an encoded video stream to nullify malicious agents potentially embedded in the encoded video stream comprising decoding a received encoded video stream to obtain a decoded video stream, extracting, from the encoded video stream, encoding information calculated by an originating encoder to create the encoded video stream, adjusting the encoding information by replacing one or more quantization parameters defined in the encoding information with respective adjusted quantization parameter(s) calculated based on a random selection of a value from a range of quantization parameter values, encoding the decoded video stream using the adjusted encoding information to produce a modified encoded video stream and transmitting the modified encoded video stream.Type: GrantFiled: March 17, 2019Date of Patent: September 7, 2021Assignee: International Business Machines CorporationInventor: Ophir Azulai
-
Publication number: 20210208862Abstract: Automatic cloning of a Python Conda environment into a Docker image, such that at least one Conda container that functions the same as the Python Conda environment can be started from the Docker image. The automatic cloning may include: First, creating a Dockerfile that comprises commands to: install a Python Anaconda environment or obtain a Python Anaconda environment image, copy the Python Conda environment into the Docker image, and run a Conda command, in the Anaconda environment, to create a cloned Python Conda environment from the copied Python Conda environment. Second, building the Docker image from the Dockerfile.Type: ApplicationFiled: January 7, 2020Publication date: July 8, 2021Inventors: Ophir Azulai, Ofer Lavi, ERAN RAICHSTEIN
-
Publication number: 20200296373Abstract: A method of disarming and reconstructing an encoded video stream to nullify malicious agents potentially embedded in the encoded video stream comprising decoding a received encoded video stream to obtain a decoded video stream, extracting, from the encoded video stream, encoding information calculated by an originating encoder to create the encoded video stream, adjusting the encoding information by replacing one or more quantization parameters defined in the encoding information with respective adjusted quantization parameter(s) calculated based on a random selection of a value from a range of quantization parameter values, encoding the decoded video stream using the adjusted encoding information to produce a modified encoded video stream and transmitting the modified encoded video stream.Type: ApplicationFiled: March 17, 2019Publication date: September 17, 2020Inventor: Ophir Azulai
-
Patent number: 10628703Abstract: Technology for matching images (for example, video images, still images) of an identical infrastructure object (for example, a tower component of a tower supporting power lines) for purposes of comparing the infrastructure object to itself at different points in time to detect a potential anomaly and the potential need for maintenance of the infrastructure object. In some embodiments, this matching of images is done using creation of a three dimensional (#D) computer model of the infrastructure object and by tagging captured images with location on the 3D model across multiple videos taken at different points in time.Type: GrantFiled: December 19, 2017Date of Patent: April 21, 2020Assignee: International Business Machines CorporationInventors: Udi Barzelay, Ophir Azulai, Yochay Tzur
-
Patent number: 10607107Abstract: Technology for matching images (for example, video images, still images) of an identical infrastructure object (for example, a tower component of a tower supporting power lines) for purposes of comparing the infrastructure object to itself at different points in time to detect a potential anomaly and the potential need for maintenance of the infrastructure object. In some embodiments, this matching of images is done using creation of a three dimensional (# D) computer model of the infrastructure object and by tagging captured images with location on the 3D model across multiple videos taken at different points in time.Type: GrantFiled: September 11, 2019Date of Patent: March 31, 2020Assignee: International Business Machines CorporationInventors: Udi Barzelay, Ophir Azulai, Yochay Tzur
-
Publication number: 20200005077Abstract: Technology for matching images (for example, video images, still images) of an identical infrastructure object (for example, a tower component of a tower supporting power lines) for purposes of comparing the infrastructure object to itself at different points in time to detect a potential anomaly and the potential need for maintenance of the infrastructure object. In some embodiments, this matching of images is done using creation of a three dimensional (#D) computer model of the infrastructure object and by tagging captured images with location on the 3D model across multiple videos taken at different points in time.Type: ApplicationFiled: September 11, 2019Publication date: January 2, 2020Inventors: Udi Barzelay, Ophir Azulai, Yochay Tzur
-
Patent number: 10417273Abstract: A computer implemented method of mapping multimedia analytics of multimedia objects into a resilient distributed dataset (RDD), comprising one or more processors adapted to obtain an RDD of a cluster computing framework executed by a cluster comprising a plurality of computing nodes, the RDD comprises a plurality of entries each comprising a pointer to one of a plurality of multimedia objects stored in a shared storage, instruct each of a plurality of framework tasks executed by at least some members of the cluster to apply a docker operator for retrieving and executing one of a plurality of multimedia containers each associated with a respective one of the multimedia objects and comprising a multimedia processing algorithm for processing the respective multimedia object, receive from the framework tasks multimedia analytics results generated simultaneously by the multimedia containers and map the multimedia analytics results into the RDD.Type: GrantFiled: January 5, 2017Date of Patent: September 17, 2019Assignee: International Business Machines CorporationInventors: Gal Ashour, Ophir Azulai, Roy Levin
-
Publication number: 20190188521Abstract: Technology for matching images (for example, video images, still images) of an identical infrastructure object (for example, a tower component of a tower supporting power lines) for purposes of comparing the infrastructure object to itself at different points in time to detect a potential anomaly and the potential need for maintenance of the infrastructure object. In some embodiments, this matching of images is done using creation of a three dimensional (#D) computer model of the infrastructure object and by tagging captured images with location on the 3D model across multiple videos taken at different points in time.Type: ApplicationFiled: December 19, 2017Publication date: June 20, 2019Inventors: Udi Barzelay, Ophir Azulai, Yochay Tzur
-
Publication number: 20180189296Abstract: A computer implemented method of mapping multimedia analytics of multimedia objects into a resilient distributed dataset (RDD), comprising one or more processors adapted to obtain an RDD of a cluster computing framework executed by a cluster comprising a plurality of computing nodes, the RDD comprises a plurality of entries each comprising a pointer to one of a plurality of multimedia objects stored in a shared storage, instruct each of a plurality of framework tasks executed by at least some members of the cluster to apply a docker operator for retrieving and executing one of a plurality of multimedia containers each associated with a respective one of the multimedia objects and comprising a multimedia processing algorithm for processing the respective multimedia object, receive from the framework tasks multimedia analytics results generated simultaneously by the multimedia containers and map the multimedia analytics results into the RDD.Type: ApplicationFiled: January 5, 2017Publication date: July 5, 2018Inventors: GAL ASHOUR, Ophir Azulai, Roy Levin