Patents by Inventor Jhing-Fa Wang

Jhing-Fa Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20130261879
    Abstract: A method for diagnosing breakdown cause of a vehicle is disclosed. In the method, several sound signals are sensed respectively with several sound sensing devices, which are respectively installed at several zones among the vehicle, from the vehicle. A current driving status of the vehicle is obtained through an electrical control unit (ECU) of the vehicle. Determine a sound source of the vehicle according to the sound signals. A breakdown cause of the vehicle is diagnosed according to the sound signals, the current driving status and the sound source.
    Type: Application
    Filed: May 21, 2012
    Publication date: October 3, 2013
    Applicant: Institute for Information Industry
    Inventors: Shi-Huang Chen, Jhing-Fa Wang, Tzyh-Jong Shang
  • Patent number: 8451292
    Abstract: A video summarized method based on mining the story structure and semantic relations among concept entities has steps of processing a video to generate multiple important shots that are annotated with respective keywords: Performing a concept expansion process by using the keywords to create expansion trees for the annotated shots; rearranging the keywords of the expansion trees and classifying to calculate relations thereof; applying a graph entropy algorithm to determine significant shots and edges interconnected with the shots. Based on the determined result of the graph entropy algorithm, a structured relational graph is built to display the significant shots and edges thereof. Consequently, users can more rapidly browse the content of a video and comprehend if different shots are related.
    Type: Grant
    Filed: November 23, 2009
    Date of Patent: May 28, 2013
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Bo-Wei Chen, Jia-Ching Wang, Chia-Hung Chang
  • Patent number: 8204114
    Abstract: A block intra prediction direction detection algorithm comprises acts of dividing a block, finding directions from edge assent rules, determining a main edge of the block, selecting prediction modes from the main edge, choosing base prediction modes and using all unique selected and base prediction modes in intra prediction. The algorithms comprise a 4×4 block intra prediction direction detection algorithm, a 16×16 luminance block intra prediction direction detection algorithm and an 8×8 chrominance block intra prediction direction detection algorithm.
    Type: Grant
    Filed: July 3, 2008
    Date of Patent: June 19, 2012
    Assignee: National Cheng-Kung University
    Inventors: Jhing-Fa Wang, Wei-Guang Lin, An-Chao Tsai
  • Publication number: 20110122137
    Abstract: A video summarized method based on mining the story structure and semantic relations among concept entities has steps of processing a video to generate multiple important shots that are annotated with respective keywords: Performing a concept expansion process by using the keywords to create expansion trees for the annotated shots; rearranging the keywords of the expansion trees and classifying to calculate relations thereof; applying a graph entropy algorithm to determine significant shots and edges interconnected with the shots. Based on the determined result of the graph entropy algorithm, a structured relational graph is built to display the significant shots and edges thereof. Consequently, users can more rapidly browse the content of a video and comprehend if different shots are related.
    Type: Application
    Filed: November 23, 2009
    Publication date: May 26, 2011
    Applicant: NATIONAL CHENG KUNG UNIVERSITY
    Inventors: Jhing-Fa WANG, Bo-Wei CHEN, Jia-Ching WANG, Chia-Hung CHANG
  • Patent number: 7774203
    Abstract: The present invention discloses an audio signal segmentation algorithm comprising the following steps. First, an audio signal is provided. Then, an audio activity detection (AAD) step is applied to divide the audio signal into at least one noise segment and at least one noisy audio segment. Then, an audio feature extraction step is used on the noisy audio segment to obtain multiple audio features. Then, a smoothing step is applied. Then, multiple speech frames and multiple music frames are discriminated. The speech frames and the music frames compose at least one speech segment and at least one music segment. Finally, the speech segment and the music segment are segmented from the noisy audio segment.
    Type: Grant
    Filed: October 31, 2006
    Date of Patent: August 10, 2010
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Chao-Ching Huang, Dian-Jia Wu
  • Patent number: 7707032
    Abstract: A method and system used to determine the similarity between an input speech data and a sample speech data is provided. First, the input speech data is segmented into a plurality of input speech frames and the sample speech data is segmented into a plurality of sample speech frames. Then, the input speech frames and the sample speech frames are used to build a matching matrix, wherein the matching matrix comprises the distance values between each of the input speech frames and each of the sample speech frames. Next, the distance values are used to calculate a matching score. Finally, the similarity between the input speech data and the sample speech data is determined according to this matching score.
    Type: Grant
    Filed: October 20, 2005
    Date of Patent: April 27, 2010
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Po-Chuan Lin, Li-Chang Wen
  • Patent number: 7613365
    Abstract: The present invention discloses a video summarization system and the method thereof. A similarity computing apparatus computes the similarity between each frame to obtain multiple similarity values. A key frame extracting apparatus chooses the key frames from the frames wherein the sum of the similarity values between the key frames is a minimum. A feature space mapping apparatus converts the sentences into multiple corresponding sentence vectors and computes the distance between each sentence vector to obtain multiple distance values. A clustering apparatus divides the sentences into multiple clusters according to the distance values and the importance of the sentences, and also applies a splitting step to split the cluster with the highest importance into multiple new clusters. A key sentence extracting apparatus chooses multiple key sentence from the clusters, wherein the sum of the importance of the key sentences is the maximum.
    Type: Grant
    Filed: July 14, 2006
    Date of Patent: November 3, 2009
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Chen-Yu Chen
  • Patent number: 7574360
    Abstract: A unit selection module for Chinese Text-to-Speech (TTS) synthesis includes a probabilistic context free grammar (PCFG) parser, a latent semantic indexing (LSI) module, and a modified variable-length unit selection scheme; any Chinese sentence is firstly input and then parsed into a context-free grammar (CFG) by the PCFG parser; wherein there are several possible CFGs for every Chinese sentence, and the CFG (or the syntactic structure) with the highest probability is then taken as the best CFG (or the syntactic structure) of the Chinese sentence; the LSI module is then used to calculate the structural distance between all the candidate synthesis units and the target unit in a corpus; through the modified variable-length unit selection scheme, tagged with the dynamic programming algorithm, the units are searched to find the best synthesis unit concatenation sequence.
    Type: Grant
    Filed: July 22, 2005
    Date of Patent: August 11, 2009
    Assignee: National Cheng Kung University
    Inventors: Chung Hsien Wu, Jiun Fu Chen, Chi Chun Hsia, Jhing Fa Wang
  • Publication number: 20090052534
    Abstract: A block intra prediction direction detection algorithm comprises acts of dividing a block, finding directions from edge assent rules, determining a main edge of the block, selecting prediction modes from the main edge, choosing base prediction modes and using all unique selected and base prediction modes in intra prediction. The algorithms comprise a 4×4 block intra prediction direction detection algorithm, a 16×16 luminance block intra prediction direction detection algorithm and an 8×8 chrominance block intra prediction direction detection algorithm.
    Type: Application
    Filed: July 3, 2008
    Publication date: February 26, 2009
    Applicant: NATIONAL CHENG-KUNG UNIVERSITY
    Inventors: Jhing-Fa WANG, Wei-Guang LIN, An-Chao TSAI
  • Patent number: 7418131
    Abstract: An image-capturing device and method for removing strangers from an image are described. First, a first image is input. Then, a control module determines if an unwanted object processing step is needed, and obtains a result. If the result is no, the first image is directly sent to an output module. If the result is yes, an image-identifying module begins to identify the target-image and the unwanted object in the first image, and then, an unwanted object processing module starts the step to process unwanted images. The unwanted object processing step can remove the unwanted object from an image and fill the left lacuna region. Afterwards, a second image is produced and sent to the output module.
    Type: Grant
    Filed: July 6, 2005
    Date of Patent: August 26, 2008
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Han-Jen Hsu, Shang-Chia Liao
  • Patent number: 7305337
    Abstract: The present invention includes a method for speech encoding and decoding and a design of speech coder and decoder. The characteristic of speech encoding method relies on the type of data with high compression rate after the whole speech data is compressed. The present invention is able to lower the bit rate of the original speech from 64 Kbps to 1.6 Kbps and provide a bit rate lower than the traditional compression method. It can provide good speech quality, and attain the function of storing the maximum speech data with minimum memory. As to the speech decoding method, some random noises are appropriated added into the exciting source, so that more speech characteristics can be simulated to produce various speech sounds. In addition, the present invention also discloses a coder and a decoder designed by application specific integrated circuit, and the structural design is optimized according to the software.
    Type: Grant
    Filed: December 24, 2002
    Date of Patent: December 4, 2007
    Assignee: National Cheng Kung University
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Yun-Fei Chao, Han-Chiang Chen, Ming-Chi Shih
  • Publication number: 20070271093
    Abstract: The present invention discloses an audio signal segmentation algorithm comprising the following steps. First, an audio signal is provided. Then, an audio activity detection (AAD) step is applied to divide the audio signal into at least one noise segment and at least one noisy audio segment. Then, an audio feature extraction step is used on the noisy audio segment to obtain multiple audio features. Then, a smoothing step is applied. Then, multiple speech frames and multiple music frames are discriminated. The speech frames and the music frames compose at least one speech segment and at least one music segment. Finally, the speech segment and the music segment are segmented from the noisy audio segment.
    Type: Application
    Filed: October 31, 2006
    Publication date: November 22, 2007
    Inventors: Jhing-Fa Wang, Chao-Ching Huang, Dian-Jia Wu
  • Publication number: 20070214418
    Abstract: The present invention discloses a video summarization system and the method thereof. A similarity computing apparatus computes the similarity between each frame to obtain multiple similarity values. A key frame extracting apparatus chooses the key frames from the frames wherein the sum of the similarity values between the key frames is a minimum. A feature space mapping apparatus converts the sentences into multiple corresponding sentence vectors and computes the distance between each sentence vector to obtain multiple distance values. A clustering apparatus divides the sentences into multiple clusters according to the distance values and the importance of the sentences, and also applies a splitting step to split the cluster with the highest importance into multiple new clusters. A key sentence extracting apparatus chooses multiple key sentence from the clusters, wherein the sum of the importance of the key sentences is the maximum.
    Type: Application
    Filed: July 14, 2006
    Publication date: September 13, 2007
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Chen-Yu Chen
  • Patent number: 7266496
    Abstract: The present invention discloses a complete speech recognition system having a training button and a recognition button, and the whole system uses the application specific integrated circuit (ASIC) architecture for the design, and also uses the modular design to divide the speech processing into 4 modules: system control module, autocorrelation and linear predictive coefficient module, cepstrum module, and DTW recognition module. Each module forms an intellectual product (IP) component by itself. Each IP component can work with various products and application requirements for the design reuse to greatly shorten the time to market.
    Type: Grant
    Filed: December 24, 2002
    Date of Patent: September 4, 2007
    Assignee: National Cheng-Kung University
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Tai-Lung Chen, Chin-Chan Chang
  • Publication number: 20070094020
    Abstract: A method and system used to determine the similarity between an input speech data and a sample speech data is provided. First, the input speech data is segmented into a plurality of input speech frames and the sample speech data is segmented into a plurality of sample speech frames. Then, the input speech frames and the sample speech frames are used to build a matching matrix, wherein the matching matrix comprises the distance values between each of the input speech frames and each of the sample speech frames. Next, the distance values are used to calculate a matching score. Finally, the similarity between the input speech data and the sample speech data is determined according to this matching score.
    Type: Application
    Filed: October 20, 2005
    Publication date: April 26, 2007
    Applicant: NATIONAL CHENG KUNG UNIVERSITY
    Inventors: Jhing-Fa Wang, Po-Chuan Lin, Li-Chang Wen
  • Publication number: 20060095264
    Abstract: This invention relates to a unit selection module for Chinese Text-to-Speech (TTS) synthesis, mainly comprising a probabilistic context free grammar (PCFG) parser, a latent semantic indexing (LSI) module, and a modified variable-length unit selection scheme; any Chinese sentence is firstly input and then parsed into a context-free grammar (CFG) by the PCFG parser; wherein there are several possible CFGs for every Chinese sentence, and the CFG (or the syntactic structure) with the highest probability is then taken as the best CFG (or the syntactic structure) of the Chinese sentence; the LSI module is then used to calculate the structural distance between all the candidate synthesis units and the target unit in a corpus; through the modified variable-length unit selection scheme, tagged with the dynamic programming algorithm, the units are searched to find the best synthesis unit concatenation sequence.
    Type: Application
    Filed: July 22, 2005
    Publication date: May 4, 2006
    Applicant: National Cheng Kung University
    Inventors: Chung-Hsien Wu, Jiun-Fu Chen, Chi-Chun Hsia, Jhing-Fa Wang
  • Publication number: 20060045372
    Abstract: An image-capturing device and method for removing strangers from an image are described. First, a first image is input. Then, a control module determines if an unwanted object processing step is needed, and obtains a result. If the result is no, the first image is directly sent to an output module. If the result is yes, an image-identifying module begins to identify the target-image and the unwanted object in the first image, and then, an unwanted object processing module starts the step to process unwanted images. The unwanted object processing step can remove the unwanted object from an image and fill the left lacuna region. Afterwards, a second image is produced and sent to the output module.
    Type: Application
    Filed: July 6, 2005
    Publication date: March 2, 2006
    Applicant: NATIONAL CHENG KUNG UNIVERSITY
    Inventors: Jhing-Fa Wang, Han-Jen Hsu, Shang-Chia Liao
  • Publication number: 20030139923
    Abstract: The present invention includes a method for speech encoding and decoding and a design of speech coder and decoder. The characteristic of speech encoding method relies on the type of data with high compression rate after the whole speech data is compressed. The present invention is able to lower the bit rate of the original speech from 64 Kbps to 1.6 Kbps and provide a bit rate lower than the traditional compression method. It can provide good speech quality, and attain the function of storing the maximum speech data with minimum memory. As to the speech decoding method, some random noises are appropriated added into the exciting source, so that more speech characteristics can be simulated to produce various speech sounds. In addition, the present invention also discloses a coder and a decoder designed by application specific integrated circuit, and the structural design is optimized according to the software.
    Type: Application
    Filed: December 24, 2002
    Publication date: July 24, 2003
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Yun-Fei Chao, Han-Chiang Chen, Ming-Chi Shih
  • Publication number: 20030125938
    Abstract: The present invention discloses a complete speech recognition system having a training button and a recognition button, and the whole system uses the application specific integrated circuit (ASIC) architecture for the design, and also uses the modular design to divide the speech processing into 4 modules: system control module, autocorrelation and linear predictive coefficient module, cepstrum module, and DTW recognition module. Each module forms an intellectual product (IP) component by itself. Each IP component can work with various products and application requirements for the design reuse to greatly shorten the time to market.
    Type: Application
    Filed: December 24, 2002
    Publication date: July 3, 2003
    Inventors: Jhing-Fa Wang, Jia-Ching Wang, Tai-Lung Chen, Chin-Chan Chang