Patents by Inventor Changchuan Yin
Changchuan Yin has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12333238Abstract: Concepts and technologies disclosed herein are directed to embedding texts into high dimensional vectors in natural language processing (“NLP”). According to one aspect, an NLP system can receive an input text that includes n number of words. The NLP system can encode the input text into a first matrix using a word embedding algorithm, such as Word2Vec algorithm. The NLP system can encode the input text into the Word2Vec by embedding each word in the n number of words of the input text into a k-dimensional Word2Vec vector using the Word2Vec algorithm. The NLP system also can decode the first matrix into a second matrix using a text embedding algorithm. In some embodiments, the second matrix is a congruence derivative matrix. The NLP system can then output the second matrix to a machine learning module that implements a machine learning technique such as short text classification.Type: GrantFiled: May 26, 2022Date of Patent: June 17, 2025Assignee: AT&T Mobility II LLCInventors: Changchuan Yin, Shahzad Saeed
-
Patent number: 12008448Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.Type: GrantFiled: March 13, 2023Date of Patent: June 11, 2024Assignee: AT&T Intellect al P Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20240005082Abstract: Concepts and technologies disclosed herein are directed to embedding texts to high dimensional vectors in natural language processing (“NLP”). According to one aspect, an NLP system can receive an input text that includes n number of words. The NLP system can encode the input text into a first matrix using a word embedding algorithm, such as Word2Vec algorithm. The NLP system can encode the input text into the Word2Vec by embedding each word in the n number of words of the input text into a k-dimensional Word2Vec vector using the Word2Vec algorithm. The NLP system also can decode the first matrix into a second matrix using a text embedding algorithm. In some embodiments, the second matrix is a congruence derivative matrix. The NLP system can then output the second matrix to a machine learning module that implements a machine learning technique such as short text classification.Type: ApplicationFiled: May 26, 2022Publication date: January 4, 2024Applicant: AT&T Mobility II LLCInventors: Changchuan Yin, Shahzad Saeed
-
Publication number: 20230222389Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.Type: ApplicationFiled: March 13, 2023Publication date: July 13, 2023Inventor: Changchuan Yin
-
Patent number: 11675965Abstract: An example method is provided for encoding text for language processing. The method may be executed by a processing system, and the method includes receiving text comprising a plurality of alphanumeric characters or symbols and converting the text into a numerical vector comprising a plurality of numerical values, by mapping each alphanumeric character or symbol of the text to a vertex coordinate of one of a plurality of vertices of a hypercube, wherein a number of the plurality of vertices is equal to or greater than a number of the plurality of alphanumeric characters or symbols, wherein the numerical vector consumes less space in memory than the text. An amount of time consumed by language processing of the numerical vector may be less than an amount of time consumed by language processing of the text.Type: GrantFiled: April 7, 2021Date of Patent: June 13, 2023Assignee: AT&T Intellectual Property I, L.P.Inventors: Changchuan Yin, Sachin Lohe
-
Publication number: 20230104459Abstract: A method for minimizing algorithmic bias in machine learning models includes generating, using a first machine learning model, a first output representing a prediction, where the first output is generated in response to a first plurality of inputs including sensitive features and non-sensitive features, generating, using a second machine learning model, a second output that minimizes an influence of an algorithmic bias in the first output, where the second output is generated in response to a second plurality of inputs including the non-sensitive features and the first output, and generating a recommendation related to the prediction based on the second output.Type: ApplicationFiled: October 4, 2021Publication date: April 6, 2023Inventors: Changchuan Yin, Cheryl Brooks, Hemamalini Kosaraju, Sachin Lohe
-
Patent number: 11605020Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.Type: GrantFiled: July 23, 2019Date of Patent: March 14, 2023Assignee: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20230067842Abstract: A processing system including at least one processor may generate a plurality of subsequences of a time series data set, convert the plurality of subsequences to a plurality of frequency domain point sets, compute pairwise distances of the plurality of frequency domain point sets, project the plurality of frequency domain point sets into a lower dimensional space in accordance with the pairwise distances, where the projecting maps each of plurality of frequency domain point sets to a node of a plurality of nodes in the lower dimensional space, and generate a notification of at least one isolated node of the plurality of nodes, where the at least one isolated node represents at least one anomaly in the time series data set.Type: ApplicationFiled: September 1, 2021Publication date: March 2, 2023Inventors: Changchuan Yin, Sachin Lohe
-
Publication number: 20220358290Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.Type: ApplicationFiled: May 10, 2022Publication date: November 10, 2022Applicant: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20220327278Abstract: An example method is provided for encoding text for language processing. The method may be executed by a processing system, and the method includes receiving text comprising a plurality of alphanumeric characters or symbols and converting the text into a numerical vector comprising a plurality of numerical values, by mapping each alphanumeric character or symbol of the text to a vertex coordinate of one of a plurality of vertices of a hypercube, wherein a number of the plurality of vertices is equal to or greater than a number of the plurality of alphanumeric characters or symbols, wherein the numerical vector consumes less space in memory than the text. An amount of time consumed by language processing of the numerical vector may be less than an amount of time consumed by language processing of the text.Type: ApplicationFiled: April 7, 2021Publication date: October 13, 2022Inventors: Changchuan Yin, Sachin Lohe
-
Patent number: 11361159Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.Type: GrantFiled: April 27, 2021Date of Patent: June 14, 2022Assignee: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20220139011Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises four dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.Type: ApplicationFiled: November 18, 2021Publication date: May 5, 2022Applicant: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Patent number: 11210824Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises two dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.Type: GrantFiled: May 21, 2020Date of Patent: December 28, 2021Assignee: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20210366170Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises two dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.Type: ApplicationFiled: May 21, 2020Publication date: November 25, 2021Applicant: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20210248318Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.Type: ApplicationFiled: April 27, 2021Publication date: August 12, 2021Applicant: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Patent number: 11017170Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.Type: GrantFiled: September 27, 2018Date of Patent: May 25, 2021Assignee: AT&T Intellectual Property I, L.P.Inventor: Changchuan Yin
-
Publication number: 20210027190Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.Type: ApplicationFiled: July 23, 2019Publication date: January 28, 2021Inventor: Changchuan Yin
-
Publication number: 20200104358Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.Type: ApplicationFiled: September 27, 2018Publication date: April 2, 2020Inventor: Changchuan Yin