Patents by Inventor Kye-hyeon KIM

Kye-hyeon KIM has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20200252770
    Abstract: A method for a V2V communication by using a radar module used for detecting objects nearby is provided. And the method includes steps of: (a) a computing device performing (i) a process of instructing the radar module to transmit 1-st transmitting signals by referring to at least one 1-st schedule and (ii) a process of generating RVA information by using (1-1)-st receiving signals, corresponding to the 1-st transmitting signals; and (b) the computing device performing a process of instructing the radar module to transmit 2-nd transmitting signals by referring to at least one 2-nd schedule.
    Type: Application
    Filed: December 31, 2019
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200249671
    Abstract: A learning method for providing a functional safety by warning a driver about a potential dangerous situation by using an explainable AI which verifies detection processes of a neural network for an autonomous driving is provided.
    Type: Application
    Filed: December 23, 2019
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200249675
    Abstract: A method for providing a dynamic adaptive deep learning model other than a fixed deep learning model, to thereby support at least one specific autonomous vehicle to perform a proper autonomous driving according to surrounding circumstances is provided. And the method includes steps of: (a) a managing device which interworks with autonomous vehicles instructing a fine-tuning system to acquire a specific deep learning model to be updated; (b) the managing device inputting video data and its corresponding labeled data to the fine-tuning system as training data, to thereby update the specific deep learning model; and (c) the managing device instructing an automatic updating system to transmit the updated specific deep learning model to the specific autonomous vehicle, to thereby support the specific autonomous vehicle to perform the autonomous driving by using the updated specific deep learning model other than a legacy deep learning model.
    Type: Application
    Filed: January 9, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250402
    Abstract: A method for face recognition by using a multiple patch combination based on a deep neural network is provided. The method includes steps of: a face-recognizing device, (a) if a face image with a 1-st size is acquired, inputting the face image into a feature extraction network, to allow the feature extraction network to generate a feature map by applying convolution operation to the face image with the 1-st size, and to generate multiple features by applying sliding-pooling operation to the feature map, wherein the feature extraction network has been learned to extract a feature using a face image for training having a 2-nd size and wherein the 2-nd size is smaller than the 1-st size; and (b) inputting the multiple features into a learned neural aggregation network, to allow the neural aggregation network to aggregate the multiple features and to output an optimal feature for the face recognition.
    Type: Application
    Filed: December 20, 2019
    Publication date: August 6, 2020
    Applicant: STRADVISION, INC.
    Inventors: Kye-Hyeon KIM, Yongjoong KIM, Hak-Kyoung KIM, Woonhyun NAM, SukHoon BOO, Myungchul SUNG, Dongsoo SHIN, Donghun YEO, Wooju RYU, Myeong-Chun LEE, Hyungsoo LEE, Taewoong JANG, Kyungjoong JEONG, Hongmo JE, Hojin CHO
  • Publication number: 20200250526
    Abstract: A method for achieving better performance in an autonomous driving while saving computing powers, by using confidence scores representing a credibility of an object detection which is generated in parallel with an object detection process is provided. And the method includes steps of: (a) a computing device acquiring at least one circumstance image on surroundings of a subject vehicle, through at least one image sensor installed on the subject vehicle; (b) the computing device instructing a Convolutional Neural Network (CNN) to apply at least one CNN operation to the circumstance image, to thereby generate initial object information and initial confidence information on the circumstance image; and (c) the computing device generating final object information on the circumstance image by referring to the initial object information and the initial confidence information with a support of a Reinforcement Learning (RL) agent, and through V2X communications with at least part of surrounding objects.
    Type: Application
    Filed: January 9, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200249699
    Abstract: A method for switching driving modes of a subject vehicle to support the subject vehicle to perform a platoon driving by using platoon driving information is provided. And the method includes steps of: (a) a basement server, which interworks with the subject vehicle driving in a first mode, acquiring first platoon driving information, to N-th platoon driving information by referring to a real-time platoon driving information DB; (b) the basement server (i) calculating a first platoon driving suitability score to an N-th platoon driving suitability score by referring to first platoon driving parameters to N-th platoon driving parameters and (ii) selecting a target platoon driving group to be including the subject vehicle; (c) the basement server instructing the subject vehicle to drive in a second mode.
    Type: Application
    Filed: December 23, 2019
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250974
    Abstract: A method for detecting emergency vehicles in real time, and managing subject vehicles to support the emergency vehicles to drive without interferences from the subject vehicles by referring to detected information on the emergency vehicles is provided. And the method includes steps of: (a) a management server generating metadata on the specific emergency vehicle by referring to emergency circumstance information; (b) the management server generating a circumstance scenario vector by referring to the emergency circumstance information and the metadata, comparing the circumstance scenario vector with reference scenario vectors, to thereby find a specific scenario vector whose similarity score with the circumstance scenario vector is larger than a threshold, and acquiring an emergency reaction command by referring to the specific scenario vector; (c) the management server transmitting the emergency reaction command to each of the subject vehicles.
    Type: Application
    Filed: January 10, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200247321
    Abstract: A method for adjusting a position of a driver assistance device according to a driver state is provided. The method includes steps of: a position adjusting device, (a) inputting an upper and a lower body images of a driver, acquired after the driver sits and starts a vehicle, into a pose estimation network, to acquire body keypoints, calculate body part lengths, and adjust a driver's seat position; and (b) while the vehicle is traveling, inputting the upper body image into a face detector to detect a facial part, inputting the facial part into an eye detector to detect an eye part, and inputting the adjusted driver's seat position and 2D coordinates of an eye into a 3D coordinates transforming device, to generate 3D coordinates of the eye referring to the 2D coordinates and the driver's seat position, and adjust a mirror position of the vehicle referring to the 3D coordinates.
    Type: Application
    Filed: January 9, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250541
    Abstract: A learning method for supporting a safer autonomous driving through a fusion of information acquired from images and communications is provided. And the method includes steps of: (a) a learning device instructing a first neural network and a second neural network to generate an image-based feature map and a communication-based feature map by using a circumstance image and circumstance communication information; (b) the learning device instructing a third neural network to apply a third neural network operation to the image-based feature map and the communication-based feature map to generate an integrated feature map; (c) the learning device instructing a fourth neural network to apply a fourth neural network operation to the integrated feature map to generate estimated surrounding motion information; and (d) the learning device instructing a first loss layer to train parameters of the first to the fourth neural networks.
    Type: Application
    Filed: January 9, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250468
    Abstract: A method for training a CNN by using a camera and a radar together, to thereby allow the CNN to perform properly even when an object depiction ratio of a photographed image acquired through the camera is low due to a bad condition of a photographing circumstance is provided. And the method includes steps of: (a) a learning device instructing a convolutional layer to apply a convolutional operation to a multichannel integrated image, to thereby generate a feature map; (b) the learning device instructing an output layer to apply an output operation to the feature map, to thereby generate estimated object information; and (c) the learning device instructing a loss layer to generate a loss by using the estimated object information and GT object information corresponding thereto, and to perform backpropagation by using the loss, to thereby learn at least part of parameters in the CNN.
    Type: Application
    Filed: December 31, 2019
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250848
    Abstract: A method for planning an autonomous driving by using a V2X communication and an image processing under a road circumstance where both vehicles capable of the V2X communication and vehicles incapable of the V2X communication exist is provided. And the method includes steps of: (a) a computing device, corresponding to a subject autonomous vehicle, instructing a planning module to acquire recognition information on surrounding vehicles including (i) first vehicles capable of a V2X communication and (ii) second vehicles incapable of the V2X communication; (b) the computing device instructing the planning module to select an interfering vehicle among the surrounding vehicles; and (c) the computing device instructing the planning module to generate a potential interference prediction model, and to modify current optimized route information in order to evade a potential interfering action, to thereby generate updated optimized route information of the subject autonomous vehicle.
    Type: Application
    Filed: December 23, 2019
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200247469
    Abstract: A method for delivering a steering intention of an autonomous driving module to a steering apparatus more accurately by using a reference map is provided. And the method includes steps of: (a) a computing device, if a subject intended steering signal inputted by the autonomous driving module at a current timing is acquired, instructing a signal adjustment module to select, by referring to the reference map, specific reference steering guide values corresponding to the subject intended steering signal; (b) the computing device (i) adjusting the subject intended steering signal by referring to the specific reference steering guide values, in order to generate a subject adjusted steering signal, and (ii) transmitting the subject adjusted steering signal to the steering apparatus, to thereby support the steering apparatus to rotate the subject vehicle by a specific steering angle corresponding to the subject intended steering signal.
    Type: Application
    Filed: January 10, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250499
    Abstract: A method for integrating images from vehicles performing a cooperative driving is provided. The method includes steps of: a main driving image integrating device on one main vehicle (a) inputting one main driving image into a main object detector to (1) generate one main feature map by applying convolution operation via a main convolutional layer, (2) generate main ROIs via a main region proposal network, (3) generate main pooled feature maps by applying pooling operation via a main pooling layer, and (4) generate main object detection information on the main objects by applying fully-connected operation via a main fully connected layer; (b) inputting the main pooled feature maps into a main confidence network to generate main confidences; and (c) acquiring sub-object detection information and sub-confidences from sub-vehicles, and integrating the main object detection information and the sub-object detection information using the main & the sub-confidences to generate object detection result.
    Type: Application
    Filed: January 10, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250514
    Abstract: A learning method for generating integrated object detection information by integrating first object detection information and second object detection information is provided. And the method includes steps of: (a) a learning device instructing a concatenating network to generate one or more pair feature vectors; (b) the learning device instructing a determining network to apply FC operations to the pair feature vectors, to thereby generate (i) determination vectors and (ii) box regression vectors; (c) the learning device instructing a loss unit to generate an integrated loss by referring to the determination vectors, the box regression vectors and their corresponding GTs, and performing backpropagation processes by using the integrated loss, to thereby learn at least part of parameters included in the DNN.
    Type: Application
    Filed: December 22, 2019
    Publication date: August 6, 2020
    Applicant: STRADVISION, INC.
    Inventors: Kye-Hyeon KIM, Yongjoong KIM, Hak-Kyoung KIM, Woonhyun NAM, SukHoon BOO, Myungchul SUNG, Dongsoo SHIN, Donghun YEO, Wooju RYU, Myeong-Chun LEE, Hyungsoo LEE, Taewoong JANG, Kyungjoong JEONG, Hongmo JE, Hojin CHO
  • Publication number: 20200249676
    Abstract: A method for providing an autonomous driving service platform for autonomous vehicles by using a competitive computing and information fusion is provided.
    Type: Application
    Filed: January 9, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200250492
    Abstract: A method for learning an automatic labeling device for auto-labeling a base image of a base vehicle using sub-images of nearby vehicles is provided. The method includes steps of: a learning device inputting the base image and the sub-images into previous trained dense correspondence networks to generate dense correspondences; and into encoders to output convolution feature maps, inputting the convolution feature maps into decoders to output deconvolution feature maps; with an integer k from 1 to n, generating a k-th adjusted deconvolution feature map by translating coordinates of a (k+1)-th deconvolution feature map using a k-th dense correspondence; generating a concatenated feature map by concatenating the 1-st deconvolution feature map and the adjusted deconvolution feature maps; and inputting the concatenated feature map into a masking layer to output a semantic segmentation image and instructing a 1-st loss layer to calculate 1-st losses and updating decoder weights and encoder weights.
    Type: Application
    Filed: January 10, 2020
    Publication date: August 6, 2020
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, Sukhoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Patent number: 10733511
    Abstract: A learning method for selecting specific information, to be used for updating an HD Map is provided. And the method includes steps of: (a) a learning device instructing a coordinate neural network to generate a local feature map and a global feature vector by applying a coordinate neural network operation to a coordinate matrix; (b) the learning device instructing a determination neural network to generate a first estimated suitability score to an N-th estimated suitability score by applying a determination neural network operation to the integrated feature map; (c) the learning device instructing a loss layer to generate a loss by referring to (i) the first estimated suitability score to the N-th estimated suitability score and (ii) a first Ground Truth(GT) suitability score to an N-th GT suitability score, and perform backpropagation by using the loss, to thereby learn parameters of the determination neural network and the coordinate neural network.
    Type: Grant
    Filed: December 23, 2019
    Date of Patent: August 4, 2020
    Assignee: StradVision, Inc.
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200242475
    Abstract: A method for on-device continual learning of a neural network which analyzes input data is provided to be used for smartphones, drones, vessels, or a military purpose. The method includes steps of: a learning device, (a) sampling new data to have a preset first volume, instructing an original data generator network, which has been learned, to repeat outputting synthetic previous data corresponding to a k-dimension random vector and previous data having been used for learning the original data generator network, such that the synthetic previous data has a second volume, and generating a batch for a current-learning; and (b) instructing the neural network to generate output information corresponding to the batch. The method can be performed by generative adversarial networks (GANs), online learning, and the like. Also, the present disclosure has effects of saving resources such as storage, preventing catastrophic forgetting, and securing privacy.
    Type: Application
    Filed: January 28, 2019
    Publication date: July 30, 2020
    Applicant: Stradvision, Inc.
    Inventors: Kye-Hyeon Kim, Yongjoong Kim, Insu Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Donghun Yeo, Wooju Ryu, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
  • Publication number: 20200241544
    Abstract: A learning method for performing a seamless parameter switch by using a location-specific algorithm selection for an optimized autonomous driving is provided. And the method includes steps of: (a) a learning device instructing a K-th convolutional layer to apply a convolution operation to K-th training images, to thereby generate K-th feature maps; (b) the learning device instructing a K-th output layer to apply a K-th output operation to the K-th feature maps, to thereby generate K-th estimated autonomous driving source information; (c) the learning device instructing a K-th loss layer to generate a K-th loss by using the K-th estimated autonomous driving source information and its corresponding GT, and then to perform backpropagation by using the K-th loss, to thereby learn K-th parameters of the K-th CNN; and (d) the learning device storing the K-th CNN in a database after tagging K-th location information to the K-th CNN.
    Type: Application
    Filed: December 31, 2019
    Publication date: July 30, 2020
    Applicant: STRADVISION, INC.
    Inventors: Kye-Hyeon KIM, Yongjoong KIM, Hak-Kyoung KIM, Woonhyun NAM, SukHoon BOO, Myungchul SUNG, Dongsoo SHIN, Donghun YEO, Wooju RYU, Myeong-Chun LEE, Hyungsoo LEE, Taewoong JANG, Kyungjoong JEONG, Hongmo JE, Hojin CHO
  • Publication number: 20200241526
    Abstract: A method for remotely controlling at least one autonomous vehicle is provided. The method includes steps of: an autonomous driving control device, (a) on condition that the autonomous driving control device detects driving environment by referring to sensor information and allows the autonomous vehicle to travel on an autonomous driving mode or a manual driving mode, determining whether the autonomous driving control device fails to establish a driving plan by using the driving environment and whether the autonomous driving control device fails to change to the manual driving mode by using the driving environment; and (b) if the autonomous driving control device fails to establish the driving plan or fails to change to the manual driving mode, selecting a remote control mode and transmitting request information to a remote control service providing server, to allow a remote driver to control the autonomous vehicle by using a specific remote vehicle.
    Type: Application
    Filed: December 31, 2019
    Publication date: July 30, 2020
    Applicant: STRADVISION, INC.
    Inventors: Kye-Hyeon KIM, Yongjoong KIM, Hak-Kyoung KIM, Woonhyun NAM, SukHoon BOO, Myungchul SUNG, Dongsoo SHIN, Donghun YEO, Wooju RYU, Myeong-Chun LEE, Hyungsoo LEE, Taewoong JANG, Kyungjoong JEONG, Hongmo JE, Hojin CHO