Patents by Inventor Inseon Jang

Inseon Jang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230039546
    Abstract: An audio encoding/decoding apparatus and method using vector quantized residual error features are disclosed. An audio signal encoding method includes outputting a bitstream of a main codec by encoding an original signal, decoding the bitstream of the main codec, determining a residual error feature vector from a feature vector of a decoded signal and a feature vector of the original signal, and outputting a bitstream of additional information by encoding the residual error feature vector.
    Type: Application
    Filed: April 1, 2022
    Publication date: February 9, 2023
    Inventors: Inseon JANG, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Woo-taek LIM, Jongwon SHIN, Youngju CHEON, Sangwook HAN, Soojoong HWANG
  • Publication number: 20230038394
    Abstract: Disclosed are a method of encoding and decoding an audio signal and an encoder and a decoder performing the method. The method of encoding an audio signal includes identifying an input signal, and generating a bitstring of each encoding layer by applying, to the input signal, an encoding model including a plurality of successive encoding layers that encodes the input signal, in which a current encoding layer among the encoding layers is trained to generate a bitstring of the current encoding layer by encoding an encoded signal which is a signal encoded in a previous encoding layer and quantizing an encoded signal which is a signal encoded in the current encoding layer.
    Type: Application
    Filed: July 30, 2021
    Publication date: February 9, 2023
    Inventors: Woo-taek LIM, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Inseon JANG, Minje KIM
  • Patent number: 11562757
    Abstract: An audio signal encoding method performed by an encoder includes identifying a time-domain audio signal in a unit of blocks, quantizing a linear prediction coefficient extracted from a combined block in which a current original block of the audio signal and a previous original block chronologically adjacent to the current original block using frequency-domain linear predictive coding (LPC), generating a temporal envelope by dequantizing the quantized linear prediction coefficient, extracting a residual signal from the combined block based on the temporal envelope, quantizing the residual signal by one of time-domain quantization and frequency-domain quantization, and transforming the quantized residual signal and the quantized linear prediction coefficient into a bitstream.
    Type: Grant
    Filed: July 15, 2021
    Date of Patent: January 24, 2023
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Seung Kwon Beack, Jongmo Sung, Mi Suk Lee, Tae Jin Lee, Woo-taek Lim, Inseon Jang, Jin Soo Choi
  • Publication number: 20220375483
    Abstract: Disclosed are methods of encoding and decoding an audio signal, and an encoder and a decoder for performing the methods. The method of encoding an audio signal includes identifying an input signal corresponding to a low frequency band of the audio signal, windowing the input signal, generating a first latent vector by inputting the windowed input signal to a first encoding model, transforming the windowed input signal into a frequency domain, generating a second latent vector by inputting the transformed input signal to a second encoding model, generating a final latent vector by combining the first latent vector and the second latent vector, and generating a bitstream corresponding to the final latent vector.
    Type: Application
    Filed: November 8, 2021
    Publication date: November 24, 2022
    Inventors: Woo-taek LIM, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Inseon JANG, Jong-won SEOK, YUNSU KIM
  • Publication number: 20220358940
    Abstract: Disclosed are methods of encoding and decoding an audio signal using side information, and an encoder and a decoder for performing the methods. The method of encoding an audio signal using side information includes identifying an input signal, the input signal being an original audio signal, extracting side information from the input signal using a learning model trained to extract side information from a feature vector of the input signal, encoding the input signal, and generating a bitstream by combining the encoded input signal and the side information.
    Type: Application
    Filed: November 16, 2021
    Publication date: November 10, 2022
    Inventors: Woo-taek LIM, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Inseon JANG, Jong Won SHIN, Soojoong HWANG, Youngju CHEON, Sangwook HAN
  • Publication number: 20220335963
    Abstract: An audio signal encoding and decoding method using a neural network model, and an encoder and decoder for performing the same are disclosed. A method of encoding an audio signal using a neural network model, the method may include identifying an input signal, generating a quantized latent vector by inputting the input signal into a neural network model encoding the input signal, and generating a bitstream corresponding to the quantized latent vector, wherein the neural network model may include i) a feature extraction layer generating a latent vector by extracting a feature of the input signal, ii) a plurality of downsampling blocks downsampling the latent vector, and iii) a plurality of quantization blocks performing quantization of a downsampled latent vector.
    Type: Application
    Filed: February 11, 2022
    Publication date: October 20, 2022
    Inventors: Inseon JANG, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Woo-taek LIM, Hong-Goo KANG, Jihyun LEE, Chanwoo LEE, Hyungseob LIM
  • Publication number: 20220262378
    Abstract: An audio signal encoding and decoding method using a learning model, a training method of the learning model, and an encoder and decoder that perform the method, are disclosed. The audio signal decoding method may include extracting a first residual signal and a first linear prediction coefficient by decoding a bitstream received from an encoder, generating a first audio signal from the first residual signal using the first linear prediction coefficient, generating a second linear prediction coefficients and a second residual signal from the first audio signal, obtaining a third linear prediction coefficient by inputting the second linear prediction coefficient into a trained learning model, and generating a second audio signal from the second residual signal using the third linear prediction coefficient.
    Type: Application
    Filed: February 15, 2022
    Publication date: August 18, 2022
    Inventors: Woo-taek LIM, Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Inseon JANG
  • Publication number: 20220238126
    Abstract: Methods of encoding and decoding an audio signal using a learning model and an encoder and a decoder for performing the methods are disclosed. A method of encoding an audio signal using a learning model may include extracting pitch information of the audio signal, determining a dilation factor of a receptive field of a first expandable neural network block to extract a feature map from the audio signal based on the pitch information, generating a first feature map of the audio signal using the first expandable neural network block in which the dilation factor is determined, determining a second feature map by inputting the first feature map into a second expandable neural network block to process the first feature map, and converting the second feature map and the pitch information into a bitstream.
    Type: Application
    Filed: January 7, 2022
    Publication date: July 28, 2022
    Inventors: Jongmo SUNG, Seung Kwon BEACK, Tae Jin LEE, Woo-taek LIM, Inseon JANG
  • Publication number: 20220157326
    Abstract: A method of generating a residual signal performed by an encoder includes identifying an input signal including an audio sample, generating a first residual signal from the input signal using linear predictive coding (LPC), generating a second residual signal having a less information amount than the first residual signal by transforming the first residual signal, transforming the second residual signal into a frequency domain, and generating a third residual signal having a less information amount than the second residual signal from the transformed second residual signal using frequency-domain prediction (FDP) coding.
    Type: Application
    Filed: October 21, 2021
    Publication date: May 19, 2022
    Inventors: Seung Kwon BEACK, Jongmo SUNG, Tae Jin LEE, Woo-taek LIM, Inseon JANG
  • Publication number: 20220020385
    Abstract: An audio signal encoding method performed by an encoder includes identifying a time-domain audio signal in a unit of blocks, quantizing a linear prediction coefficient extracted from a combined block in which a current original block of the audio signal and a previous original block chronologically adjacent to the current original block using frequency-domain linear predictive coding (LPC), generating a temporal envelope by dequantizing the quantized linear prediction coefficient, extracting a residual signal from the combined block based on the temporal envelope, quantizing the residual signal by one of time-domain quantization and frequency-domain quantization, and transforming the quantized residual signal and the quantized linear prediction coefficient into a bitstream.
    Type: Application
    Filed: July 15, 2021
    Publication date: January 20, 2022
    Inventors: Seung Kwon Beack, Jongmo Sung, Mi Suk Lee, Tae Jin Lee, Woo-taek Lim, Inseon Jang, Jin Soo Choi
  • Publication number: 20220005487
    Abstract: An audio signal encoding and decoding method using a neural network model, a method of training the neural network model, and an encoder and decoder performing the methods are disclosed. The encoding method includes computing the first feature information of an input signal using a recurrent encoding model, computing an output signal from the first feature information using a recurrent decoding model, calculating a residual signal by subtracting the output signal from the input signal, computing the second feature information of the residual signal using a nonrecurrent encoding model, and converting the first feature information and the second feature information to a bitstream.
    Type: Application
    Filed: July 6, 2021
    Publication date: January 6, 2022
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Jongmo SUNG, Seung Kwon BEACK, Mi Suk LEE, Tae Jin LEE, Woo-taek LIM, Inseon JANG
  • Publication number: 20220005488
    Abstract: The encoding method includes computing the first feature information of an input signal using a recurrent encoding model, quantizing the first feature information and producing the first feature bitstream, computing the first output signal from the quantized first feature information using a recurrent decoding model, computing the second feature information of the input signal using a nonrecurrent encoding model, quantizing the second feature information and producing the second feature bitstream, computing the second output signal from the quantized second feature information using a nonrecurrent decoding model, determining an encoding mode based on the input signal, the first and second output signals, and the first and second feature bitstreams, and outputting an overall bitstream by multiplexing an encoding mode bit and one of the first feature bitstream and the second feature bitstream depending on the encoding mode.
    Type: Application
    Filed: July 6, 2021
    Publication date: January 6, 2022
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Jongmo SUNG, Seung Kwon BEACK, Mi Suk LEE, Tae Jin LEE, Woo-taek LIM, Inseon JANG
  • Publication number: 20210398547
    Abstract: An audio signal encoding method performed by an encoder includes identifying an audio signal of a time domain in units of a block, generating a combined block by combining i) a current original block of the audio signal and ii) a previous original block chronologically adjacent to the current original block, extracting a first residual signal of a frequency domain from the combined block using linear predictive coding of a time domain, overlapping chronologically adjacent first residual signals among first residual signals converted into a time domain, and quantizing a second residual signal of a time domain extracted from the overlapped first residual signal by converting the second residual signal of the time domain into a frequency domain using linear predictive coding of a frequency domain.
    Type: Application
    Filed: May 26, 2021
    Publication date: December 23, 2021
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Seung Kwon BEACK, Jongmo SUNG, Mi Suk LEE, Tae Jin LEE, Woo-taek LIM, Inseon JANG
  • Publication number: 20210390967
    Abstract: Disclosed is a method of encoding and decoding an audio signal using linear predictive coding (LPC) and an encoder and a decoder that perform the method. The method of encoding an audio signal to be performed by the encoder includes identifying a time-domain audio signal block-wise, quantizing a linear prediction coefficient obtained from a block of the audio signal through the LPC, generating an envelope based on the quantized linear prediction coefficient, extracting a residual signal based on the envelope and a result of converting the block into a frequency domain, grouping the residual signal by each sub-band and determining a scale factor for quantizing the grouped residual signal, quantizing the residual signal using the scale factor, and converting the quantized residual signal and the quantized linear prediction coefficient into a bitstream and transmitting the bitstream to a decoder.
    Type: Application
    Filed: April 28, 2021
    Publication date: December 16, 2021
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Seung Kwon Beack, Jongmo SUNG, Mi Suk LEE, Tae Jin LEE, Woo-taek LIM, Inseon JANG, Jin Soo CHOI
  • Publication number: 20210366497
    Abstract: Methods of encoding and decoding a speech signal using a neural network model that recognizes sound sources, and encoding and decoding apparatuses for performing the methods are provided. A method of encoding a speech signal includes identifying an input signal for a plurality of sound sources; generating a latent signal by encoding the input signal; obtaining a plurality of sound source signals by separating the latent signal for each of the plurality of sound sources; determining a number of bits used for quantization of each of the plurality of sound source signals according to a type of each of the plurality of sound sources; quantizing each of the plurality of sound source signals based on the determined number of bits; and generating a bitstream by combining the plurality of quantized sound source signals.
    Type: Application
    Filed: May 20, 2021
    Publication date: November 25, 2021
    Applicants: Electronics and Telecommunications Research Institute, The Trustees of Indiana University
    Inventors: Woo-taek LIM, Seung Kwon BEACK, Jongmo SUNG, Mi Suk LEE, Tae Jin LEE, Inseon JANG, Minje KIM, Haici YANG
  • Publication number: 20210090551
    Abstract: An emotional speech generating method and apparatus capable of adjusting an emotional intensity is disclosed. The emotional speech generating method includes generating emotion groups by grouping weight vectors representing a same emotion into a same emotion group, determining an internal distance between weight vectors included in a same emotion group, determining an external distance between weight vectors included in a same emotion group and weight vectors included in another emotion group, determining a representative weight vector of each of the emotion groups based on the internal distance and the external distance, generating a style embedding by applying the representative weight vector of each of the emotion groups to a style token including prosodic information for expressing an emotion, and generating an emotional speech expressing the emotion using the style embedding.
    Type: Application
    Filed: September 23, 2020
    Publication date: March 25, 2021
    Applicants: Electronics and Telecommunications Research Institute, Industry-Academic Cooperation Foundation, Yonsei University
    Inventors: Inseon JANG, Hong-Goo KANG, Chung Hyun AHN, Se-Yun UM, Sangshin OH, Tae Jin LEE
  • Patent number: 9680999
    Abstract: An acoustic echo removing apparatus detects space information representing a location of a sound source using a plurality of microphone input signals that are received through a plurality of microphones, and generates an acoustic echo estimation signal from a far-end talker voice signal using an adaptive filter coefficient. The acoustic echo removing apparatus detects a double talk segment using the space information, and determines update of the adaptive filter coefficient according to whether the double talk segment exists.
    Type: Grant
    Filed: April 30, 2014
    Date of Patent: June 13, 2017
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventor: Inseon Jang
  • Publication number: 20150133097
    Abstract: An acoustic echo removing apparatus detects space information representing a location of a sound source using a plurality of microphone input signals that are received through a plurality of microphones, and generates an acoustic echo estimation signal from a far-end talker voice signal using an adaptive filter coefficient. The acoustic echo removing apparatus detects a double talk segment using the space information, and determines update of the adaptive filter coefficient according to whether the double talk segment exists.
    Type: Application
    Filed: April 30, 2014
    Publication date: May 14, 2015
    Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventor: Inseon JANG
  • Patent number: 8976983
    Abstract: Provided are a method of generating and playing an object-based audio content that may effectively store preset information about an object-based audio content, and a computer-readable recording medium for storing data having a file format structure for an object-based audio service. The method of generating the object-based audio content may include: receiving a plurality of audio objects (310) generating at least one preset using the plurality of audio objects (320) and storing a preset parameter with respect to an attribute of the at least preset and the plurality of audio objects (330). The preset parameter may be stored in a form of a box that is defined in a media file format about the object-based audio content. Through this, it is possible to effectively store a preset about a plurality of audio objects.
    Type: Grant
    Filed: April 23, 2009
    Date of Patent: March 10, 2015
    Assignee: Electronics and Telecommunications Research Institute
    Inventors: Inseon Jang, Jeongil Seo, Hui Yong Kim, Tae Jin Lee, Kyeongok Kang, Jin Woo Hong, Jin Woong Kim, Chieteuk Ahn, SeungChul Ham
  • Patent number: 8340943
    Abstract: Provided is an apparatus of separating a musical sound source, which may re-construct mixed signals into target sound sources and other sound sources directly using sound source information performed using a predetermined musical instrument when the sound source information is present, thereby more effectively separating sound sources included in the mixed signal. The apparatus may include a Nonnegative Matrix Partial Co-Factorization (NMPCF) analysis unit to perform an NMPCF analysis on a mixed signal and a predetermined sound source signal using a sound source separation model, and to obtain a plurality of entity matrices based on the analysis result, and a target instrument signal separating unit to separate, from the mixed signal, a target instrument signal corresponding to the predetermined sound source signal by calculating an inner product between the plurality of entity matrices.
    Type: Grant
    Filed: August 12, 2010
    Date of Patent: December 25, 2012
    Assignees: Electronics and Telecommunications Research Institute, Postech Acadeny-Industry Foundation
    Inventors: Min Je Kim, Seungjin Choi, Jiho Yoo, Kyeongok Kang, Inseon Jang, Jin-Woo Hong