Network Structures Patents (Class 382/158)
-
Patent number: 12257727Abstract: A classifying sensing system, a classifying method performed using a sensing system, a tactile sensor, and a method of fabricating a tactile sensor. The classifying sensing system comprises a first spiking neural network, SNN, encoder configured for encoding an event-based output of a vision sensor into individual vision modality spiking representations with a first output size; a second SNN encoder configured for encoding an event-based output of a tactile sensor into individual tactile modality spiking representations with a second output size; a combination layer configured for merging the vision modality spiking representations and the tactile modality spiking representations; and a task SNN configured to receive the merged vision modality spiking representations and tactile modality spiking representations and output vision-tactile modality spiking representations with a third output size for classification.Type: GrantFiled: June 15, 2021Date of Patent: March 25, 2025Assignee: NATIONAL UNIVERSITY OF SINGAPOREInventors: Chee Keong Tee, Hian Hian See, Brian Lim, Soon Hong Harold Soh, Tasbolat Taunyazov, Weicong Sng, Sheng Yuan Jethro Kuan, Abdul Fatir Ansari
-
Patent number: 12236676Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating realistic extensions of images. In one aspect, a method comprises providing an input that comprises a provided image to a generative neural network having a plurality of generative neural network parameters. The generative neural network processes the input in accordance with trained values of the plurality of generative neural network parameters to generate an extended image. The extended image has (i) more rows, more columns, or both than the provided image, and (ii) is predicted to be a realistic extension of the provided image. The generative neural network is trained using an adversarial loss objective function.Type: GrantFiled: July 19, 2019Date of Patent: February 25, 2025Assignee: Google LLCInventors: Mikael Pierre Bonnevie, Aaron Maschinot, Aaron Sarna, Shuchao Bi, Jingbin Wang, Michael Spencer Krainin, Wenchao Tong, Dilip Krishnan, Haifeng Gong, Ce Liu, Hossein Talebi, Raanan Sayag, Piotr Teterwak
-
Patent number: 12154307Abstract: A sequence of patch tokens representing an image can be received. A network can be trained to learn informative patch tokens and uninformative patch tokens in the sequence of patch tokens, in learning to recognize an object in the image. The sequence of patch tokens can be reduced by removing the uninformative patch tokens from the sequence of patch tokens. The reduced sequence of patch tokens can be input to an attention-based deep learning neural network. The attention-based deep learning neural network can be fine-tuned to recognize the object in the image using the reduced sequence of patch tokens.Type: GrantFiled: December 22, 2021Date of Patent: November 26, 2024Assignees: International Business Machines Corporation, Massachusetts Institute of TechnologyInventors: Bowen Pan, Rameswar Panda, Rogerio Schmidt Feris, Aude Jeanne Oliva
-
Patent number: 12125244Abstract: A method includes generating multiple attention map for an image from outputs of a corresponding different one of multiple convolutional hidden layers of a neural network. A different weighted attention map is then generated from each of the attention map. The weighted attention maps are input to a first fully-connected neural network layer to generate a colorfulness metric, which may be used to augment human-perceived colorfulness of the image.Type: GrantFiled: February 15, 2022Date of Patent: October 22, 2024Assignee: KYOCERA Document Solutions, Inc.Inventors: Kilho Shin, Dongpei Su
-
Patent number: 12112569Abstract: Embodiments include systems and methods that may be performed by a processor of a computing device. Embodiments may be applied for keypoint detection in an image. In embodiments, the processor of the computing device may apply to an image a first-stage neural network to define and output a plurality of regions, apply to each of the plurality of regions a respective second-stage neural network to output a plurality of keypoints in each of the plurality of regions, and apply to the plurality of keypoints a third-stage neural network to determine a correction for each of the plurality of keypoints to provide corrected keypoints.Type: GrantFiled: November 19, 2021Date of Patent: October 8, 2024Assignee: QUALCOMM IncorporatedInventors: Upal Mahbub, Rakesh Nattoji Rajaram, Vasudev Bhaskaran
-
Patent number: 12086717Abstract: An all-optical Diffractive Deep Neural Network (D2NN) architecture learns to implement various functions or tasks after deep learning-based design of the passive diffractive or reflective substrate layers that work collectively to perform the desired function or task. This architecture was successfully confirmed experimentally by creating 3D-printed D2NNs that learned to implement handwritten classifications and lens function at the terahertz spectrum. This all-optical deep learning framework can perform, at the speed of light, various complex functions and tasks that computer-based neural networks can implement, and will find applications in all-optical image analysis, feature detection and object classification, also enabling new camera designs and optical components that can learn to perform unique tasks using D2NNs. In alternative embodiments, the all-optical D2NN is used as a front-end in conjunction with a trained, digital neural network back-end.Type: GrantFiled: May 12, 2023Date of Patent: September 10, 2024Assignee: THE REGENTS OF THE UNIVERSITY OF CALIFORNIAInventors: Aydogan Ozcan, Yair Rivenson, Xing Lin, Deniz Mengu, Yi Luo
-
Patent number: 12033338Abstract: A method for determining a trajectory of a vehicle within a physical space based at least on an aligned point cloud may include generating a fused feature map by concatenating a first feature map corresponding to a source point cloud and a second feature map corresponding to a target point cloud. A machine learning model may be applied to determine, based at least on the fused feature map, a relative transform aligning the target point cloud to the source point cloud. An aligned target point cloud may be generated by transforming the target point cloud in accordance with the relative transform. Furthermore, a trajectory of a vehicle within the physical space may be determined based on at least the first relative transform. Related systems and computer program products are also provided.Type: GrantFiled: March 31, 2023Date of Patent: July 9, 2024Assignee: Motional AD LLCInventors: Sherif Ahmed Morsy Nekkah, Nicole Alexandra Camous, Sergi Adipraja Widjaja, Venice Erin Baylon Liong, Xiaogang Wang
-
Patent number: 11994408Abstract: A method for localization performed by an agent includes receiving a query image of a current environment of the agent captured by a sensor integrated with the agent. The method also includes receiving a target image comprising a first set of keypoints matching a second set of keypoints of the query image. The first set of keypoints may be generated based on a task specified for the agent. The method still further includes determining a current location based on the target image.Type: GrantFiled: April 14, 2021Date of Patent: May 28, 2024Assignee: TOYOTA RESEARCH INSTITUTE, INC.Inventors: Jiexiong Tang, Rares Andrei Ambrus, Hanme Kim, Vitor Guizilini, Adrien David Gaidon, Xipeng Wang, Jeff Walls, Sudeep Pillai
-
Patent number: 11967150Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for parallel processing of video frames using neural networks. One of the methods includes receiving a video sequence comprising a respective video frame at each of a plurality of time steps; and processing the video sequence using a video processing neural network to generate a video processing output for the video sequence, wherein the video processing neural network includes a sequence of network components, wherein the network components comprise a plurality of layer blocks each comprising one or more neural network layers, wherein each component is active for a respective subset of the plurality of time steps, and wherein each layer block is configured to, at each time step at which the layer block is active, receive an input generated at a previous time step and to process the input to generate a block output.Type: GrantFiled: February 13, 2023Date of Patent: April 23, 2024Assignee: DeepMind Technologies LimitedInventors: Simon Osindero, Joao Carreira, Viorica Patraucean, Andrew Zisserman
-
Patent number: 11954591Abstract: This application provides a method of generating a description for a picture set performed at a computer device, and a storage medium. The method includes: acquiring a picture set to be processed; performing picture feature extraction on each picture in the picture set to acquire a picture feature corresponding to the picture, and forming a picture feature sequence corresponding to the picture set by using the picture features corresponding to the pictures; performing scene feature extraction on a picture feature sub-sequence corresponding to each scene in the picture feature sequence to acquire a scene feature corresponding to the scene, and forming a scene feature sequence corresponding to the picture set by using the scene features corresponding to the scenes; and generating textual description information of the picture set according to the picture feature sequence and the scene feature sequence that correspond to the picture set.Type: GrantFiled: August 11, 2020Date of Patent: April 9, 2024Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITEDInventors: Bairui Wang, Lin Ma, Wei Liu
-
Patent number: 11915474Abstract: Techniques and apparatus for analyzing visual content using a visual transformer are described. An example technique includes generating a first set of tokens based on a visual content item. Each token in the first set of tokens is associated with a regional feature from a different region of a plurality of regions of the visual content item. A second set of tokens is generated based on the visual content item. Each token in the second set of tokens is associated with a local feature from one of the plurality of regions of the visual content item. At least one feature map is generated for the visual content item, based on analyzing the first set of tokens and the second set of tokens separately using a hierarchical vision transformer. At least one vision task is performed based on the at least one feature map.Type: GrantFiled: May 31, 2022Date of Patent: February 27, 2024Assignee: International Business Machines CorporationInventors: Richard Chen, Rameswar Panda, Quanfu Fan
-
Patent number: 11854211Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: GrantFiled: January 26, 2022Date of Patent: December 26, 2023Assignee: Microsoft Technology Licensing, LLC.Inventors: Ishani Chakraborty, Jonathan C. Hanzelka, Lu Yuan, Pedro Urbina Escos, Thomas M. Soemo
-
Patent number: 11810341Abstract: A computer-implemented method of identifying filters for use in determining explainability of a trained neural network. The method comprises obtaining a dataset comprising the input image and an annotation of an input image, the annotation indicating at least one part of the input image which is relevant for inferring classification of the input image, determining an explanation filter set by iteratively: selecting a filter of the plurality of filters; adding the filter to the explanation filter set; computing an explanation heatmap for the input image by resizing and combining an output of each filter in the explanation filter set to obtain the explanation heatmap, the explanation heatmap having a spatial resolution of the input image; and computing a similarity metric by comparing the explanation heatmap to the annotation of the input image; until the similarity metric is greater than or equal to a similarity threshold; and outputting the explanation filter set.Type: GrantFiled: January 8, 2021Date of Patent: November 7, 2023Assignee: ROBERT BOSCH GMBHInventor: Andres Mauricio Munoz Delgado
-
Patent number: 11790523Abstract: A device receives an input image of a portion of a patient's body, and applies the input image to a feature extraction model, the feature extraction model comprising a trained machine learning model that is configured to generate an output that comprises, for each respective location of a plurality of locations in the input image, an indication that the input image contains an object of interest that is indicative of a presence of a disease state at the respective location. The device applies the output of the feature extraction model to a diagnostic model, the diagnostic model comprising a trained machine learning model that is configured to output a diagnosis of a disease condition in the patient based on the output of the feature extraction model. The device outputs the determined diagnosis of a disease condition in the patient obtained from the diagnostic model.Type: GrantFiled: October 30, 2018Date of Patent: October 17, 2023Assignee: Digital Diagnostics Inc.Inventors: Meindert Niemeijer, Ryan Amelon, Warren Clarida, Michael D. Abramoff
-
Patent number: 11768912Abstract: A computer-implemented method according to one embodiment includes receiving historical two-dimensional (2D) multivariate time series data; transforming the historical 2D multivariate time series data into a three-dimensional (3D) temporal tensor; training one or more deep volumetric 3D convolutional neural networks (CNNs), utilizing the 3D temporal tensor; and predicting future values for additional multivariate time series data, utilizing the one or more trained deep volumetric 3D CNNs.Type: GrantFiled: July 12, 2019Date of Patent: September 26, 2023Assignee: International Business Machines CorporationInventors: Mu Qiao, Yuya Jeremy Ong, Divyesh Jadav
-
Patent number: 11756284Abstract: An apparatus of labeling for object detection according to an embodiment of the present disclosure includes an image selector that determines a plurality of labeling target images from among a plurality of unlabeled images, and determines a labeling order of the plurality of labeling target images, a feedback obtainer that obtains label inspection information on the plurality of labeling target images from a user, and a model trainer that learns the label inspection information input from the user by using the labeling target images, obtains a pseudo label for supervised learning based on a learning result using the label inspection information, and re-determines the labeling order of the labeling target images based on the pseudo label.Type: GrantFiled: February 17, 2021Date of Patent: September 12, 2023Assignee: SAMSUNG SDS CO., LTD.Inventors: Seung Ho Shin, Ji Hoon Kim, Se Won Woo, Kyoung Jin Oh, Seong Won Park
-
Patent number: 11741734Abstract: Aspects of the disclosure provide for mechanisms for identification of blocks of associated words in documents using neural networks. A method of the disclosure includes obtaining a plurality of words of a document, the document having a first block of associated words, determining a plurality of vectors representative of the plurality of words, processing the plurality of vectors using a first neural network to obtain a plurality of recalculated vectors having values based on the plurality of vectors, determining a plurality of association values corresponding to a connections between at least two words of the document, and identifying, using the plurality of recalculated vectors and the plurality of association values, the first block of associated symbol sequences.Type: GrantFiled: January 13, 2022Date of Patent: August 29, 2023Assignee: ABBYY Development Inc.Inventor: Stanislav Semenov
-
Patent number: 11741723Abstract: A system and method for performing intersection scenario retrieval that includes receiving a video stream of a surrounding environment of an ego vehicle. The system and method also include analyzing the video stream to trim the video stream into video clips of an intersection scene associated with the travel of the ego vehicle. The system and method additionally include annotating the ego vehicle, dynamic objects, and their motion paths that are included within the intersection scene with action units that describe an intersection scenario. The system and method further include retrieving at least one intersection scenario based on a query of an electronic dataset that stores a combination of action units to operably control a presentation of at least one intersection scenario video clip that includes the at least one intersection scenario.Type: GrantFiled: June 29, 2020Date of Patent: August 29, 2023Assignee: HONDA MOTOR CO., LTD.Inventors: Yi-Ting Chen, Nakul Agarwal, Behzad Dariush, Ahmed Taha
-
Patent number: 11735174Abstract: A method of training a natural language neural network comprises obtaining at least one constituency span; obtaining a training video input; applying a multi-modal transform to the video input, thereby generating a transformed video input; comparing the at least one constituency span and the transformed video input using a compound Probabilistic Context-Free Grammar (PCFG) model to match the at least one constituency span with corresponding portions of the transformed video input; and using results from the comparison to learn a constituency parser.Type: GrantFiled: October 12, 2022Date of Patent: August 22, 2023Assignee: TENCENT AMERICA LLCInventor: Linfeng Song
-
Patent number: 11734347Abstract: A video retrieval method and apparatus, a device and a storage medium are provided. The method comprises the following steps: acquiring a comparison video clip from a video library according to the duration of a to-be-tested video (S110); determining the similarity between the to-be-tested video and the comparison video clip by a target spatio-temporal neural network, a spatio-temporal convolutional layer of the target spatio-temporal neural network being configured to be capable of performing two-dimensional convolution and temporal dimension information processing, respectively (S120); and traversing the video library, and outputting a retrieval result according to the similarity (S130).Type: GrantFiled: July 7, 2021Date of Patent: August 22, 2023Assignee: LYNXI TECHNOLOGIES CO., LTD.Inventors: Zhenzhi Wu, Yaolong Zhu
-
Patent number: 11694082Abstract: An all-optical Diffractive Deep Neural Network (D2NN) architecture learns to implement various functions or tasks after deep learning-based design of the passive diffractive or reflective substrate layers that work collectively to perform the desired function or task. This architecture was successfully confirmed experimentally by creating 3D-printed D2NNs that learned to implement handwritten classifications and lens function at the terahertz spectrum. This all-optical deep learning framework can perform, at the speed of light, various complex functions and tasks that computer-based neural networks can implement, and will find applications in all-optical image analysis, feature detection and object classification, also enabling new camera designs and optical components that can learn to perform unique tasks using D2NNs. In alternative embodiments, the all-optical D2NN is used as a front-end in conjunction with a trained, digital neural network back-end.Type: GrantFiled: June 17, 2022Date of Patent: July 4, 2023Assignee: THE REGENTS OF THE UNIVERSITY OF CALIFORNIAInventors: Aydogan Ozcan, Yair Rivenson, Xing Lin, Deniz Mengu, Yi Luo
-
Patent number: 11580736Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for parallel processing of video frames using neural networks. One of the methods includes receiving a video sequence comprising a respective video frame at each of a plurality of time steps; and processing the video sequence using a video processing neural network to generate a video processing output for the video sequence, wherein the video processing neural network includes a sequence of network components, wherein the network components comprise a plurality of layer blocks each comprising one or more neural network layers, wherein each component is active for a respective subset of the plurality of time steps, and wherein each layer block is configured to, at each time step at which the layer block is active, receive an input generated at a previous time step and to process the input to generate a block output.Type: GrantFiled: January 7, 2019Date of Patent: February 14, 2023Assignee: DeepMind Technologies LimitedInventors: Simon Osindero, Joao Carreira, Viorica Patraucean, Andrew Zisserman
-
Patent number: 11557026Abstract: A technique for detecting a glitch in an image is provided. The technique includes providing an image to a plurality of individual classifiers to generate a plurality of individual classifier outputs and providing the plurality of individual classifier outputs to an ensemble classifier to generate a glitch classification.Type: GrantFiled: September 23, 2020Date of Patent: January 17, 2023Assignees: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Nicholas Malaya, Max Kiehn, Stanislav Ivashkevich
-
Patent number: 11521605Abstract: A method of training a natural language neural network comprises obtaining at least one constituency span; obtaining a training video input; applying a multi-modal transform to the video input, thereby generating a transformed video input; comparing the at least one constituency span and the transformed video input using a compound Probabilistic Context-Free Grammar (PCFG) model to match the at least one constituency span with corresponding portions of the transformed video input; and using results from the comparison to learn a constituency parser.Type: GrantFiled: February 22, 2021Date of Patent: December 6, 2022Assignee: TENCENT AMERICA LLCInventor: Linfeng Song
-
Patent number: 11496635Abstract: An information processing system acquires, using a reading device, a read image from an original on which a handwritten character is written; acquires, based on the read image, a partial image that is a partial region of the read image and a binarized image that expresses the partial image by two tones; performs learning of a learning model based on learning data that uses the partial image as a correct answer image and the binarized image as an input image; acquires print data including a font character; generates conversion image data including a gradation character obtained by inputting the font character to the learning model; and causes an image forming device to form an image based on the generated conversion image data.Type: GrantFiled: April 12, 2021Date of Patent: November 8, 2022Assignee: Canon Kabushiki KaishaInventor: Keisui Okuma
-
Patent number: 11475658Abstract: Embodiments described herein relate generally to a methodology of efficient object classification within a visual medium. The methodology utilizes a first neural network to perform an attention based object localization within a visual medium to generate a visual mask. The visual mask is applied to the visual medium to generate a masked visual medium. The masked visual medium may be then fed into a second neural network to detect and classify objects within the visual medium.Type: GrantFiled: February 18, 2021Date of Patent: October 18, 2022Assignee: ANCESTRY.COM OPERATIONS INC.Inventors: Mohammad K. Ebrahimpour, Yen-Yun Yu, Jiayun Li, Jack Reese, Azadeh Moghtaderi
-
Patent number: 11475660Abstract: One embodiment facilitates recognizing parts of a vehicle. A convolution module is configured to generate a convolution feature map of a vehicle image. A region proposal module is configured to determine, based on the convolution feature map, one or more proposed regions, wherein a respective proposed region corresponds to a target of a respective vehicle part. A classification module is configured to determine a class and a bounding box of a vehicle part corresponding to a proposed region based on a feature of the proposed region. A conditional random field module is configured to optimize classes and bounding boxes of the vehicle parts based on correlated features of the corresponding proposed regions. A reporting module is configured to generate a result which indicates a list including an insurance claim item and corresponding damages based on the optimized classes and bounding boxes of the vehicle parts.Type: GrantFiled: August 29, 2019Date of Patent: October 18, 2022Assignee: Advanced New Technologies Co., LTD.Inventor: Qingpei Guo
-
Patent number: 11429103Abstract: Systems, methods, devices, and techniques for planning travel of an autonomous robot. A system identifies one or more obstacles that are located in proximity of at least a portion of a planned route for the autonomous robot. For each obstacle, the system: (i) determines a semantic class of the obstacle, including selecting the semantic class from a library that defines a set of multiple possible semantic classes for obstacles, and (ii) selects a planning policy for the obstacle that corresponds to the semantic class of the obstacle. The system can generate a trajectory along the at least the portion of the planned route using the selected planning policies. The robot can then initiate travel according to the trajectory.Type: GrantFiled: March 27, 2020Date of Patent: August 30, 2022Assignee: X Development LLCInventors: David Millard, Mikael Persson
-
Patent number: 11417007Abstract: A method of controlling an electronic apparatus includes acquiring an image and depth information of the acquired image; inputting the acquired image into a neural network model trained to acquire information on objects included in the acquired image; acquiring an intermediate feature value output by an intermediate layer of the neural network model; identifying a feature area for at least one object among the objects included in the acquired image based on the intermediate feature value; and acquiring distance information between the electronic apparatus and the at least one object based on the feature area for the at least one object and the depth information.Type: GrantFiled: November 6, 2020Date of Patent: August 16, 2022Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Seongmin Kang, Heungwoo Han
-
Patent number: 11403511Abstract: In some implementations at an electronic device, training a dual EDNN includes defining a data structure of attributes corresponding to defined parts of a task, processing a first instance of an input using a first EDNN to produce a first output while encoding a first set of the attributes in a first latent space, and processing a second instance of the input using a second EDNN to produce a second output while encoding attribute differences from attribute averages in a second latent space. The device then determines a second set of the attributes based on the attribute differences and the attribute averages. The device then adjusts parameters of the first and second EDNNs based on comparing the first instance of the input to the first output, the second instance of the input to the second output, and the first set of attributes to the second set of attributes.Type: GrantFiled: July 18, 2019Date of Patent: August 2, 2022Assignee: Apple Inc.Inventors: Peter Meier, Tanmay Batra
-
Method and apparatus for performing block retrieval on block to be processed of urine sediment image
Patent number: 11386340Abstract: The inventive concepts herein relate to performing block retrieval on a block to be processed of a urine sediment image. The method comprises: using a plurality of decision trees to perform block retrieval on the block to be processed, wherein each of the plurality of decision trees comprises a judgment node and a leaf node, and the judgment node judges the block to be processed to make it reach the leaf node by using a block retrieval feature in a block retrieval feature set to form a block retrieval result at the leaf node, and at least two decision trees in the plurality of decision trees are different in structures thereof and/or judgments performed by the judgment nodes thereof by using the block retrieval feature; and integrating the block retrieval results of the plurality of decision trees so as to form a final block retrieval result.Type: GrantFiled: August 7, 2020Date of Patent: July 12, 2022Assignee: SIEMENS HEALTHCARE DIAGNOSTIC INC.Inventors: Tian Shen, Juan Xu, XiaoFan Zhang -
Low probability transitions and boundary crossing into disallowed states for a more optimal solution
Patent number: 11308402Abstract: Artificial intelligence (AI) techniques that map disallowed states and enable access to those states under certain conditions through a search algorithm are disclosed. In other words, scenario boundaries may be crossed by jumping from one scenario that is less desirable or even has no solution to another scenario that is more desirable.Type: GrantFiled: July 2, 2020Date of Patent: April 19, 2022Assignee: THE AEROSPACE CORPORATIONInventors: Terence Yeoh, Nehal Desai -
Patent number: 11270084Abstract: A method for generating a human-like response to a voice or text command includes receiving an input sequence of words and processing the input sequence of words to generate a trigger word that is indicative of a desired nature of the human-like response. The method further includes encoding a neural network using the trigger word and generating the human-like response using an output of the neural network. The method enables implementation of voice command functionality in various types of devices with only a small amount of training data.Type: GrantFiled: October 12, 2018Date of Patent: March 8, 2022Assignee: Johnson Controls Tyco IP Holdings LLPInventors: Viswanath Ramamurti, Young M. Lee
-
Patent number: 11232299Abstract: Aspects of the disclosure provide for mechanisms for identification of blocks of associated words in documents using neural networks. A method of the disclosure includes obtaining a plurality of words of a document, the document having a first block of associated words, determining a plurality of vectors representative of the plurality of words, processing the plurality of vectors using a first neural network to obtain a plurality of recalculated vectors having values based on the plurality of vectors, determining a plurality of association values corresponding to a connections between at least two words of the document, and identifying, using the plurality of recalculated vectors and the plurality of association values, the first block of associated symbol sequences.Type: GrantFiled: December 18, 2019Date of Patent: January 25, 2022Assignee: ABBYY Production LLCInventor: Stanislav Semenov
-
Patent number: 11200056Abstract: A parallel union control device includes: at least one memory storing a set of instructions; and at least one processor configured to execute the set of instructions to cause each of the plurality of arithmetic units included in an parallel computer including a vector register to: successively compare input elements of a pair of input sets to undergo union processing, the pair being stored in an input operand register in the vector register; select one of the input elements as an output element of an output set, based on a comparison result; and store the output element into an output operand register in the vector register; shift a pointer pointing to the input element; load the input sets into the input operand register from a memory; store the output sets into the memory from the output operand register; and determine whether union processing performed in parallel is ended.Type: GrantFiled: February 5, 2019Date of Patent: December 14, 2021Assignee: NEC CORPORATIONInventors: Harumichi Yokoyama, Takuya Araki, Haoran Li
-
Patent number: 11164087Abstract: There is provided a system including a non-transitory memory storing an executable code and a hardware processor executing the executable code to receive an input sentence including a first predicate and at least a first argument depending from the first predicate, identify the first predicate, identify the first argument based on the first predicate, apply a dependency multiplication to determine a semantic role of the first argument based on the first predicate, and assign the first argument to an argument cluster including one or more similar arguments based on the semantic role of the first argument.Type: GrantFiled: May 20, 2016Date of Patent: November 2, 2021Assignee: Disney Enterprises, Inc.Inventors: Boyang Li, Yi Luan
-
Patent number: 11132605Abstract: Cardinal sine function used as an activation function for a hierarchical classifier. Application of a sine function, or a cardinal sine function, for hierarchical classification of a subject within subject matter domains and sub-domains. Hierarchical classification or multi-level classification is improved through use of the cardinal sine function or even standard sine function. Some embodiments of the present invention focus on the usage of cardinal sine function as activation function and how to apply this cardinal sine function for hierarchical classification of a subject. Some embodiments include a technique by which hierarchical classification or multi-level classification can benefit from application of a cardinal sine function.Type: GrantFiled: November 20, 2017Date of Patent: September 28, 2021Assignee: International Business Machines CorporationInventor: Abhishek Dasgupta
-
Patent number: 11113840Abstract: A method configured to implemented on at least one image processing device for detecting objects in images includes obtaining an image including an object. The method also includes generating one or more feature vectors related to the image based on a first convolutional neural network, wherein the one or more feature vectors includes a plurality of parameters. The method further includes determining the position of the object based on at least one of the plurality of parameters. The method still further includes determining a category associated with the object based on at least one the plurality of parameters.Type: GrantFiled: June 28, 2019Date of Patent: September 7, 2021Assignee: ZHEJIANG DAHUA TECHNOLOGY CO., LTD.Inventors: Xin Ye, Songlin Yang
-
Patent number: 11113836Abstract: Embodiments of object detection method, device, apparatus and a computer-readable storage medium are provided. The method can include: obtaining an enclosing frame of a target object in an input image; according to the enclosing frame, determining a reference frame from a predetermined candidate frame set comprising a plurality of candidate frames; generating a size-related feature according to a size of the reference frame and a size of the enclosing frame; and detecting an object in the input image by applying the size-related feature in a machine learning model. In an embodiment of the present application, the object detection is performed by using a feature related to an object size, that is, the prediction criterion related to the object size is added to an original feature in a machine learning model, thereby further improving the accuracy of the object detection.Type: GrantFiled: July 15, 2019Date of Patent: September 7, 2021Assignee: Baidu Online Network Technology (Beijing) Co., Ltd.Inventors: Xuehui Wang, Ming Li, Tian Xia
-
Patent number: 10977842Abstract: A method for processing a multi-directional X-ray computed tomography (CT) image using a neural network and an apparatus therefor are provided. The method includes receiving a predetermined number of multi-directional X-ray CT data and reconstructing an image for the multi-directional X-ray CT data using a neural network learned in each of an image domain and a sinogram domain.Type: GrantFiled: June 4, 2019Date of Patent: April 13, 2021Assignee: Korea Advanced Institute of Science and TechnologyInventors: JongChul Ye, Yoseob Han
-
Patent number: 10964061Abstract: A deep neural network (DNN) system learns a map representation for estimating a camera position and orientation (pose). The DNN is trained to learn a map representation corresponding to the environment, defining positions and attributes of structures, trees, walls, vehicles, etc. The DNN system learns a map representation that is versatile and performs well for many different environments (indoor, outdoor, natural, synthetic, etc.). The DNN system receives images of an environment captured by a camera (observations) and outputs an estimated camera pose within the environment. The estimated camera pose is used to perform camera localization, i.e., recover the three-dimensional (3D) position and orientation of a moving camera, which is a fundamental task in computer vision with a wide variety of applications in robot navigation, car localization for autonomous driving, device localization for mobile navigation, and augmented/virtual reality.Type: GrantFiled: May 12, 2020Date of Patent: March 30, 2021Assignee: NVIDIA CorporationInventors: Jinwei Gu, Samarth Manoj Brahmbhatt, Kihwan Kim, Jan Kautz
-
Patent number: 10948966Abstract: The disclosed computer-implemented method may include (i) identifying an artificial neural network that processes each input to the artificial neural network in a fixed number of operations, (ii) performing an analysis on the artificial neural network to determine an execution metric that represents the fixed number of operations performed by the artificial neural network to process each input, (iii) determining a quality-of-service metric for an executing system that executes the artificial neural network, and (iv) optimizing power consumption of the executing system by configuring, based on the execution metric and the quality-of-service metric, a processing throughput of at least one physical processor of the executing system, thereby causing the executing system to execute the artificial neural network at a rate that satisfies the quality-of-service metric while limiting the power consumption of the executing system. Various other methods, systems, and computer-readable media are also disclosed.Type: GrantFiled: March 7, 2018Date of Patent: March 16, 2021Assignee: Facebook, Inc.Inventors: Nadav Rotem, Abdulkadir Utku Diril, Mikhail Smelyanskiy, Jong Soo Park
-
Patent number: 10943105Abstract: A system and method for invoice field detection and parsing includes the steps of extracting character bounding blocks using optical character recognition (OCR) or digital character extraction (DCE), enhancing the image quality, analyzing the document layout based on imaging techniques, detecting the invoice field based on the machine learning techniques, and parsing the invoice field value based on the content information.Type: GrantFiled: September 24, 2019Date of Patent: March 9, 2021Assignee: The Neat Company, Inc.Inventors: Shuo Chen, Venkataraman Pranatharthiharan
-
Patent number: 10879946Abstract: Methods and systems for processing a noisy time series input to detect a signal, generate a de-noising mask, and/or output a de-noised time series output are provided. The input is transformed into one or more datagrams, such as real and imaginary time-frequency grams. The datagrams are stacked and provided as first and second channel inputs to a neural network. A neural network is trained to detect signals within the input. Alternatively or in addition, the network is trained to generate a de-noise mask, and/or to output a de-noised time series output. Implementation of the method and systems can include the use of multiple deep neural networks (DNNs), such as convolutional neural networks (CNN's), that are provided with inputs in the form of RF spectrograms. Embodiments of the present disclosure can be applied to various RF devices, such as communication devices, including but not limited to multiple inputs multiple output (MIMO) devices and 5G communication system devices, and RADAR devices.Type: GrantFiled: October 30, 2019Date of Patent: December 29, 2020Assignee: Ball Aerospace & Technologies Corp.Inventor: James M. Shima
-
Patent number: 10843835Abstract: Medicine packaging apparatuses and methods for accurately determining a remaining sheet amount of a medicine packaging sheet are described. The apparatus includes: a roll support section to which a core tube of a medicine packaging sheet roll is attached; a sensor disposed in the roll support section for outputting a count value according to a rotation amount; a wireless reader-writer unit for writing information to a core tube IC tag and reading said information; an information generation section for generating information to be written to the core tube IC tag; a remaining sheet amount estimation section for estimating a current amount of remaining sheet based on the information and dimensional information of the core tube; and a controller which selectively performs an operation if a reference time-point count value is not yet written to the core tube IC tag and another operation if the count value is already written thereto.Type: GrantFiled: January 30, 2019Date of Patent: November 24, 2020Assignee: YUYAMA MFG. CO., LTD.Inventors: Katsunori Yoshina, Tomohiro Sugimoto, Noriyoshi Fujii
-
Patent number: 10832124Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for generating descriptions of input images. One of the methods includes obtaining an input image; processing the input image using a first neural network to generate an alternative representation for the input image; and processing the alternative representation for the input image using a second neural network to generate a sequence of a plurality of words in a target natural language that describes the input image.Type: GrantFiled: August 12, 2019Date of Patent: November 10, 2020Assignee: Google LLCInventors: Samy Bengio, Oriol Vinyals, Alexander Toshkov Toshev, Dumitru Erhan
-
Patent number: 10796152Abstract: Embodiments described herein relate generally to a methodology of efficient object classification within a visual medium. The methodology utilizes a first neural network to perform an attention based object localization within a visual medium to generate a visual mask. The visual mask is applied to the visual medium to generate a masked visual medium. The masked visual medium may be then fed into a second neural network to detect and classify objects within the visual medium.Type: GrantFiled: September 17, 2019Date of Patent: October 6, 2020Assignee: ANCESTRY.COM OPERATIONS INC.Inventors: Mohammad K. Ebrahimpour, Yen-Yun Yu, Jiayun Li, Jack Reese, Azadeh Moghtaderi
-
Patent number: 10796221Abstract: Systems and techniques for facilitating a deep learning architecture for automated image feature extraction are presented. In one example, a system includes a machine learning component. The machine learning component generates learned imaging output regarding imaging data based on a convolutional neural network that receives the imaging data. The machine learning component also performs a plurality of sequential and/or parallel downsampling and upsampling of the imaging data associated with convolutional layers of the convolutional neural network.Type: GrantFiled: December 27, 2017Date of Patent: October 6, 2020Assignee: General Electric CompanyInventors: Min Zhang, Gopal Biligeri Avinash
-
Patent number: 10776691Abstract: Methods, systems and apparatuses, including computer programs encoded on computer storage media, are provided for learning or optimizing an indirect encoding of a mapping from digitally-encoded input arrays to digitally-encoded output arrays, with numerous technical advantages in terms of efficiency and effectiveness.Type: GrantFiled: June 23, 2016Date of Patent: September 15, 2020Assignee: Uber Technologies, Inc.Inventors: Zoubin Ghahramani, Gary Marcus
-
Patent number: 10713533Abstract: Provided are neural network-based image processing method and apparatus, and a computer-readable storage medium. The image processing method includes: inputting an image into an optimized neural network; extracting, by the optimized neural network, image features of the image; and outputting the image features, wherein the optimized neural network is obtained by performing a first optimization process on at least one sub-layer in a pre-trained initial neural network, each sub-layer of the at least one sub-layer includes a convolutional layer, and the first optimization process comprises: for each sub-layer of the at least one sub-layer, determining one or more channels to be removed from a filter of the convolutional layer and removing said one or more channels, and optimizing parameters of remaining channels in the filter of the convolutional layer, so that error of output features of each optimized sub-layer is minimized.Type: GrantFiled: July 12, 2018Date of Patent: July 14, 2020Assignee: MEGVII (BEIJING) TECHNOLOGY CO., LTD.Inventors: Xiangyu Zhang, Yihui He