Patents by Inventor Changchuan Yin

Changchuan Yin has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12333238
    Abstract: Concepts and technologies disclosed herein are directed to embedding texts into high dimensional vectors in natural language processing (“NLP”). According to one aspect, an NLP system can receive an input text that includes n number of words. The NLP system can encode the input text into a first matrix using a word embedding algorithm, such as Word2Vec algorithm. The NLP system can encode the input text into the Word2Vec by embedding each word in the n number of words of the input text into a k-dimensional Word2Vec vector using the Word2Vec algorithm. The NLP system also can decode the first matrix into a second matrix using a text embedding algorithm. In some embodiments, the second matrix is a congruence derivative matrix. The NLP system can then output the second matrix to a machine learning module that implements a machine learning technique such as short text classification.
    Type: Grant
    Filed: May 26, 2022
    Date of Patent: June 17, 2025
    Assignee: AT&T Mobility II LLC
    Inventors: Changchuan Yin, Shahzad Saeed
  • Patent number: 12008448
    Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.
    Type: Grant
    Filed: March 13, 2023
    Date of Patent: June 11, 2024
    Assignee: AT&T Intellect al P Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20240005082
    Abstract: Concepts and technologies disclosed herein are directed to embedding texts to high dimensional vectors in natural language processing (“NLP”). According to one aspect, an NLP system can receive an input text that includes n number of words. The NLP system can encode the input text into a first matrix using a word embedding algorithm, such as Word2Vec algorithm. The NLP system can encode the input text into the Word2Vec by embedding each word in the n number of words of the input text into a k-dimensional Word2Vec vector using the Word2Vec algorithm. The NLP system also can decode the first matrix into a second matrix using a text embedding algorithm. In some embodiments, the second matrix is a congruence derivative matrix. The NLP system can then output the second matrix to a machine learning module that implements a machine learning technique such as short text classification.
    Type: Application
    Filed: May 26, 2022
    Publication date: January 4, 2024
    Applicant: AT&T Mobility II LLC
    Inventors: Changchuan Yin, Shahzad Saeed
  • Publication number: 20230222389
    Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.
    Type: Application
    Filed: March 13, 2023
    Publication date: July 13, 2023
    Inventor: Changchuan Yin
  • Patent number: 11675965
    Abstract: An example method is provided for encoding text for language processing. The method may be executed by a processing system, and the method includes receiving text comprising a plurality of alphanumeric characters or symbols and converting the text into a numerical vector comprising a plurality of numerical values, by mapping each alphanumeric character or symbol of the text to a vertex coordinate of one of a plurality of vertices of a hypercube, wherein a number of the plurality of vertices is equal to or greater than a number of the plurality of alphanumeric characters or symbols, wherein the numerical vector consumes less space in memory than the text. An amount of time consumed by language processing of the numerical vector may be less than an amount of time consumed by language processing of the text.
    Type: Grant
    Filed: April 7, 2021
    Date of Patent: June 13, 2023
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Changchuan Yin, Sachin Lohe
  • Publication number: 20230104459
    Abstract: A method for minimizing algorithmic bias in machine learning models includes generating, using a first machine learning model, a first output representing a prediction, where the first output is generated in response to a first plurality of inputs including sensitive features and non-sensitive features, generating, using a second machine learning model, a second output that minimizes an influence of an algorithmic bias in the first output, where the second output is generated in response to a second plurality of inputs including the non-sensitive features and the first output, and generating a recommendation related to the prediction based on the second output.
    Type: Application
    Filed: October 4, 2021
    Publication date: April 6, 2023
    Inventors: Changchuan Yin, Cheryl Brooks, Hemamalini Kosaraju, Sachin Lohe
  • Patent number: 11605020
    Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.
    Type: Grant
    Filed: July 23, 2019
    Date of Patent: March 14, 2023
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20230067842
    Abstract: A processing system including at least one processor may generate a plurality of subsequences of a time series data set, convert the plurality of subsequences to a plurality of frequency domain point sets, compute pairwise distances of the plurality of frequency domain point sets, project the plurality of frequency domain point sets into a lower dimensional space in accordance with the pairwise distances, where the projecting maps each of plurality of frequency domain point sets to a node of a plurality of nodes in the lower dimensional space, and generate a notification of at least one isolated node of the plurality of nodes, where the at least one isolated node represents at least one anomaly in the time series data set.
    Type: Application
    Filed: September 1, 2021
    Publication date: March 2, 2023
    Inventors: Changchuan Yin, Sachin Lohe
  • Publication number: 20220358290
    Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.
    Type: Application
    Filed: May 10, 2022
    Publication date: November 10, 2022
    Applicant: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20220327278
    Abstract: An example method is provided for encoding text for language processing. The method may be executed by a processing system, and the method includes receiving text comprising a plurality of alphanumeric characters or symbols and converting the text into a numerical vector comprising a plurality of numerical values, by mapping each alphanumeric character or symbol of the text to a vertex coordinate of one of a plurality of vertices of a hypercube, wherein a number of the plurality of vertices is equal to or greater than a number of the plurality of alphanumeric characters or symbols, wherein the numerical vector consumes less space in memory than the text. An amount of time consumed by language processing of the numerical vector may be less than an amount of time consumed by language processing of the text.
    Type: Application
    Filed: April 7, 2021
    Publication date: October 13, 2022
    Inventors: Changchuan Yin, Sachin Lohe
  • Patent number: 11361159
    Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.
    Type: Grant
    Filed: April 27, 2021
    Date of Patent: June 14, 2022
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20220139011
    Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises four dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.
    Type: Application
    Filed: November 18, 2021
    Publication date: May 5, 2022
    Applicant: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Patent number: 11210824
    Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises two dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.
    Type: Grant
    Filed: May 21, 2020
    Date of Patent: December 28, 2021
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20210366170
    Abstract: Aspects of the subject disclosure may include, for example, a method of transforming, by a processing system comprising a processor, text comprising a series of characters into a graphic representation, wherein the graphic representation comprises a series of dots arranged in a two-dimensional pattern, wherein the two-dimensional pattern comprises two dots per character, and wherein each dot in the series of dots is one unit away from a preceding dot; and plotting, by the processing system, the series of dots on a two-dimensional graph, thereby creating a unique encoded image of the text. Other embodiments are disclosed.
    Type: Application
    Filed: May 21, 2020
    Publication date: November 25, 2021
    Applicant: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20210248318
    Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.
    Type: Application
    Filed: April 27, 2021
    Publication date: August 12, 2021
    Applicant: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Patent number: 11017170
    Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.
    Type: Grant
    Filed: September 27, 2018
    Date of Patent: May 25, 2021
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Changchuan Yin
  • Publication number: 20210027190
    Abstract: A processing system including at least one processor may obtain a machine learning model, serialize the machine learning model into a serialized format, and embed a delimiter indicator into a documentation file comprising information regarding the use of the machine learning model, where the delimiter indicator is in a file position that is after an end-of-file indicator of the documentation file. The processing system may further embed the machine learning model in the serialized format into the documentation file in a file position that is after the delimiter indicator. The processing system may then store the documentation file with the delimiter indicator and the machine learning model in the serialized format that are embedded.
    Type: Application
    Filed: July 23, 2019
    Publication date: January 28, 2021
    Inventor: Changchuan Yin
  • Publication number: 20200104358
    Abstract: Text can be encoded into DNA sequences. Each word from a document or other text sample can be encoded in a DNA sequence or DNA sequences and the DNA sequences can be stored for later retrieval. The DNA sequences can be stored digitally, or actual DNA molecules containing the sequences can be synthesized and stored. In one example, the encoding technique makes use of a polynomial function to transform words based on the Latin alphabet into k-mer DNA sequences of length k. Because the whole bits required for the DNA sequences are smaller than the actual strings of words, storing documents using DNA sequences may compress the documents relative to storing the same documents using other techniques. In at least one example, the mapping between words and DNA sequences is one-to-one and the collision ratio for the encoding is low.
    Type: Application
    Filed: September 27, 2018
    Publication date: April 2, 2020
    Inventor: Changchuan Yin