Patents by Inventor Zhaowen Wang
Zhaowen Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9965717Abstract: Embodiments of the present invention relate to learning image representation by distilling from multi-task networks. In implementation, more than one single-task network is trained with heterogeneous labels. In some embodiments, each of the single-task networks is transformed into a Siamese structure with three branches of sub-networks so that a common triplet ranking loss can be applied to each branch. A distilling network is trained that approximates the single-task networks on a common ranking task. In some embodiments, the distilling network is a Siamese network whose ranking function is optimized to approximate an ensemble ranking of each of the single-task networks. The distilling network can be utilized to predict tags to associate with a test image or identify similar images to the test image.Type: GrantFiled: November 13, 2015Date of Patent: May 8, 2018Assignee: Adobe Systems IncorporatedInventors: Zhaowen Wang, Xianming Liu, Hailin Jin, Chen Fang
-
Publication number: 20180114097Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: ApplicationFiled: December 22, 2017Publication date: April 26, 2018Applicant: Adobe Systems IncorporatedInventors: Zhaowen Wang, Luoqi Liu, Hailin Jin
-
Publication number: 20180089151Abstract: Font recognition and similarity determination techniques and systems are described. For example, a computing device receives an image including a font and extracts font features corresponding to the font. The computing device computes font feature distances between the font and fonts from a set of training fonts. The computing device calculates, based on the font feature distances, similarity scores for the font and the training fonts used for calculating features distances. The computing device determines, based on the similarity scores, final similarity scores for the font relative to the training fonts.Type: ApplicationFiled: September 29, 2016Publication date: March 29, 2018Inventors: Zhaowen Wang, Hailin Jin
-
Publication number: 20180082156Abstract: Font replacement based on visual similarity is described. In one or more embodiments, a font descriptor includes multiple font features derived from a visual appearance of a font by a font visual similarity model. The font visual similarity model can be trained using a machine learning system that recognizes similarity between visual appearances of two different fonts. A source computing device embeds a font descriptor in a document, which is transmitted to a destination computing device. The destination compares the embedded font descriptor to font descriptors corresponding to local fonts. Based on distances between the embedded and the local font descriptors, at least one matching font descriptor is determined. The local font corresponding to the matching font descriptor is deemed similar to the original font. The destination computing device controls presentations of the document using the similar local font. Computation of font descriptors can be outsourced to a remote location.Type: ApplicationFiled: September 19, 2016Publication date: March 22, 2018Applicant: Adobe Systems IncorporatedInventors: Hailin Jin, Zhaowen Wang, Gavin Stuart Peter Miller
-
Publication number: 20180039605Abstract: Embodiments of the present invention are directed at providing a font similarity preview for non-resident fonts. In one embodiment, a font is selected on a computing device. In response to the selection of the font, a pre-computed font list is checked to determine what fonts are similar to the selected font. In response to a determination that similar fonts are not local to the computing device, a non-resident font list is sent to a font vendor. The font vendor sends back previews of the non-resident fonts based on entitlement information of a user. Further, a full non-resident font can be synced to the computing device. Other embodiments may be described and/or claimed.Type: ApplicationFiled: August 4, 2016Publication date: February 8, 2018Inventors: I-Ming Pao, Alan Lee Erickson, Yuyan Song, Seth Shaw, Hailin Jin, Zhaowen Wang
-
Patent number: 9875429Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: GrantFiled: October 6, 2015Date of Patent: January 23, 2018Assignee: ADOBE SYSTEMS INCORPORATEDInventors: Zhaowen Wang, Luoqi Liu, Hailin Jin
-
Patent number: 9824304Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: GrantFiled: October 6, 2015Date of Patent: November 21, 2017Assignee: Adobe Systems IncorporatedInventors: Zhaowen Wang, Hailin Jin
-
Patent number: 9811765Abstract: Techniques for image captioning with weak supervision are described herein. In implementations, weak supervision data regarding a target image is obtained and utilized to provide detail information that supplements global image concepts derived for image captioning. Weak supervision data refers to noisy data that is not closely curated and may include errors. Given a target image, weak supervision data for visually similar images may be collected from sources of weakly annotated images, such as online social networks. Generally, images posted online include “weak” annotations in the form of tags, titles, labels, and short descriptions added by users. Weak supervision data for the target image is generated by extracting keywords for visually similar images discovered in the different sources. The keywords included in the weak supervision data are then employed to modulate weights applied for probabilistic classifications during image captioning analysis.Type: GrantFiled: January 13, 2016Date of Patent: November 7, 2017Assignee: Adobe Systems IncorporatedInventors: Zhaowen Wang, Quanzeng You, Hailin Jin, Chen Fang
-
Patent number: 9799102Abstract: Smoothing images using machine learning is described. In one or more embodiments, a machine learning system is trained using multiple training items. Each training item includes a boundary shape representation and a positional indicator. To generate the training item, a smooth image is downscaled to produce a corresponding blocky image that includes multiple blocks. For a given block, the boundary shape representation encodes a blocky boundary in a neighborhood around the given block. The positional indicator reflects a distance between the given block and a smooth boundary of the smooth image. In one or more embodiments to smooth a blocky image, a boundary shape representation around a selected block is determined. The representation is encoded as a feature vector and applied to the machine learning system to obtain a positional indicator. The positional indicator is used to compute a location of a smooth boundary of a smooth image.Type: GrantFiled: December 2, 2015Date of Patent: October 24, 2017Assignee: Adobe Systems IncorporatedInventors: Nathan A. Carr, Zhaowen Wang, Duygu Ceylan, I-Chao Shen
-
Patent number: 9792534Abstract: Techniques for image captioning with word vector representations are described. In implementations, instead of outputting results of caption analysis directly, the framework is adapted to output points in a semantic word vector space. These word vector representations reflect distance values in the context of the semantic word vector space. In this approach, words are mapped into a vector space and the results of caption analysis are expressed as points in the vector space that capture semantics between words. In the vector space, similar concepts with have small distance values. The word vectors are not tied to particular words or a single dictionary. A post-processing step is employed to map the points to words and convert the word vector representations to captions. Accordingly, conversion is delayed to a later stage in the process.Type: GrantFiled: January 13, 2016Date of Patent: October 17, 2017Assignee: Adobe Systems IncorporatedInventors: Zhaowen Wang, Quanzeng You, Hailin Jin, Chen Fang
-
Publication number: 20170262414Abstract: Embodiments of the present invention are directed at providing a font similarity system. In one embodiment, a new font is detected on a computing device. In response to the detection of the new font, a pre-computed font list is checked to determine whether the new font is included therein. The pre-computed font list including feature representations, generated independently of the computing device, for corresponding fonts. In response to a determination that the new font is absent from the pre-computed font list, a feature representation for the new font is generated. The generated feature representation capable of being utilized for a similarity analysis of the new font. The feature representation is then stored in a supplemental font list to enable identification of one or more fonts installed on the computing device that are similar to the new font. Other embodiments may be described and/or claimed.Type: ApplicationFiled: March 10, 2016Publication date: September 14, 2017Inventors: I-Ming Pao, Zhaowen Wang, Hailin Jin, Alan Lee Erickson
-
Publication number: 20170200066Abstract: Techniques for image captioning with word vector representations are described. In implementations, instead of outputting results of caption analysis directly, the framework is adapted to output points in a semantic word vector space. These word vector representations reflect distance values in the context of the semantic word vector space. In this approach, words are mapped into a vector space and the results of caption analysis are expressed as points in the vector space that capture semantics between words. In the vector space, similar concepts with have small distance values. The word vectors are not tied to particular words or a single dictionary. A post-processing step is employed to map the points to words and convert the word vector representations to captions. Accordingly, conversion is delayed to a later stage in the process.Type: ApplicationFiled: January 13, 2016Publication date: July 13, 2017Inventors: Zhaowen Wang, Quanzeng You, Hailin Jin, Chen Fang
-
Publication number: 20170200065Abstract: Techniques for image captioning with weak supervision are described herein. In implementations, weak supervision data regarding a target image is obtained and utilized to provide detail information that supplements global image concepts derived for image captioning. Weak supervision data refers to noisy data that is not closely curated and may include errors. Given a target image, weak supervision data for visually similar images may be collected from sources of weakly annotated images, such as online social networks. Generally, images posted online include “weak” annotations in the form of tags, titles, labels, and short descriptions added by users. Weak supervision data for the target image is generated by extracting keywords for visually similar images discovered in the different sources. The keywords included in the weak supervision data are then employed to modulate weights applied for probabilistic classifications during image captioning analysis.Type: ApplicationFiled: January 13, 2016Publication date: July 13, 2017Inventors: Zhaowen Wang, Quanzeng You, Hailin Jin, Chen Fang
-
Publication number: 20170161876Abstract: Smoothing images using machine learning is described. In one or more embodiments, a machine learning system is trained using multiple training items. Each training item includes a boundary shape representation and a positional indicator. To generate the training item, a smooth image is downscaled to produce a corresponding blocky image that includes multiple blocks. For a given block, the boundary shape representation encodes a blocky boundary in a neighborhood around the given block. The positional indicator reflects a distance between the given block and a smooth boundary of the smooth image. In one or more embodiments to smooth a blocky image, a boundary shape representation around a selected block is determined. The representation is encoded as a feature vector and applied to the machine learning system to obtain a positional indicator. The positional indicator is used to compute a location of a smooth boundary of a smooth image.Type: ApplicationFiled: December 2, 2015Publication date: June 8, 2017Inventors: Nathan A. Carr, Zhaowen Wang, Duygu Ceylan, I-Chao Shen
-
Publication number: 20170140248Abstract: Embodiments of the present invention relate to learning image representation by distilling from multi-task networks. In implementation, more than one single-task network is trained with heterogeneous labels. In some embodiments, each of the single-task networks is transformed into a Siamese structure with three branches of sub-networks so that a common triplet ranking loss can be applied to each branch. A distilling network is trained that approximates the single-task networks on a common ranking task. In some embodiments, the distilling network is a Siamese network whose ranking function is optimized to approximate an ensemble ranking of each of the single-task networks. The distilling network can be utilized to predict tags to associate with a test image or identify similar images to the test image.Type: ApplicationFiled: November 13, 2015Publication date: May 18, 2017Inventors: ZHAOWEN WANG, XIANMING LIU, HAILIN JIN, CHEN FANG
-
Publication number: 20170098141Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: ApplicationFiled: October 6, 2015Publication date: April 6, 2017Inventors: Zhaowen Wang, Hailin Jin
-
Publication number: 20170098140Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: ApplicationFiled: October 6, 2015Publication date: April 6, 2017Inventors: Zhaowen Wang, Luoqi Liu, Hailin Jin
-
Publication number: 20170098138Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.Type: ApplicationFiled: October 6, 2015Publication date: April 6, 2017Inventors: Zhaowen Wang, Luoqi Liu, Hailin Jin
-
Patent number: 9607014Abstract: A system is configured to annotate an image with tags. As configured, the system accesses an image and generates a set of vectors for the image. The set of vectors may be generated by mathematically transforming the image, such as by applying a mathematical transform to predetermined regions of the image. The system may then query a database of tagged images by submitting the set of vectors as search criteria to a search engine. The querying of the database may obtain a set of tagged images. Next, the system may rank the obtained set of tagged images according to similarity scores that quantify degrees of similarity between the image and each tagged image obtained. Tags from a top-ranked subset of the tagged images may be extracted by the system, which may then annotate the image with these extracted tags.Type: GrantFiled: October 31, 2013Date of Patent: March 28, 2017Assignee: Adobe Systems IncorporatedInventors: Zhaowen Wang, Jianchao Yang, Zhe Lin, Jonathan Brandt
-
Patent number: 9436893Abstract: A system and method for distributed similarity learning for high-dimensional image features are described. A set of data features is accessed. Subspaces from a space formed by the set of data features are determined using a set of projection matrices. Each subspace has a dimension lower than a dimension of the set of data features. Similarity functions are computed for the subspaces. Each similarity function is based on the dimension of the corresponding subspace. A linear combination of the similarity functions is performed to determine a similarity function for the set of data features.Type: GrantFiled: November 27, 2013Date of Patent: September 6, 2016Assignee: ADOBE SYSTEMS INCORPORATEDInventors: Jianchao Yang, Zhaowen Wang, Zhe Lin, Jonathan Brandt