Abstract: According to one aspect of the invention, there is provided a method for providing a voice recognition trigger, comprising the steps of: estimating a first distance, which is a distance between a device and a user, on the basis of proximity information detected by the device, and estimating a second distance, which is a distance between the device and a location where a voice detected by the device is uttered, with reference to information on the voice detected by the device; and determining whether the voice detected by the device is an object of voice recognition, with reference to similarity between the first distance and the second distance.
Abstract: According to one aspect of the invention, there is provided a method for assisting object control, comprising the steps of: specifying, in a real-world coordinate system, a control vector corresponding to a gesture or manipulation performed by a user; determining an object indicated by the control vector as a target object to be controlled by the gesture or manipulation performed by the user; performing mapping for specifying a control coordinate point corresponding to a magnitude or direction of the control vector in a target object reference coordinate system defined on the basis of a characteristic of the target object; and determining a control command to be applied to the target object, with reference to the control coordinate point specified by the mapping.
Abstract: A remote control apparatus comprises an image acquisition unit for acquiring an image of the body of a user by photographing the image, a location calculation unit for calculating coordinate data representing the locations of a first coordinate (eye) and a second coordinate (a finger or a pointer) of the body of the user by means of the image received from the image acquisition unit, and a match confirmation unit for checking whether the locations of the first and second coordinates of the body of the user match the extended connecting straight line, the coordinates having been confirmed on the basis of the camera lens-based coordinate data calculated in the location calculation unit, and if the match is confirmed, then concluding the setup intention of the user to be an affirmative.
Abstract: A remote control apparatus comprises an image acquisition unit for acquiring an image of the body of a user by photographing the image, a location calculation unit for calculating coordinate data representing the locations of a first coordinate (eye) and a second coordinate (a finger or a pointer) of the body of the user by means of the image received from the image acquisition unit, and a match confirmation unit for checking whether the locations of the first and second coordinates of the body of the user match the extended connecting straight line, the coordinates having been confirmed on the basis of the camera lens-based coordinate data calculated in the location calculation unit, and if the match is confirmed, then concluding the setup intention of the user to be an affirmative.
Abstract: A control method for allowing a user to specify an electronic device and switch it to a speech recognition mode is provided. With the optimum control method and the electronic device utilizing the method, a voice command may be transmitted to the electronic device more quickly and effectively regardless of the circumstances, and the electronic device may be specified through gesture recognition to enable transmission of the voice command, so that the voice command may be effectively executed without needing a user to learn or memorize a name or the like of the electronic device in advance for speech recognition. Further, it is possible to more accurately recognize a gesture that is a preliminary step for transmitting a voice command to the electronic device, thereby improving the recognition rate and preventing malfunction.
Type:
Application
Filed:
September 18, 2018
Publication date:
January 17, 2019
Applicant:
VTOUCH CO., LTD.
Inventors:
Seokjoong KIM, Chunghoon KIM, So Yeon KIM
Abstract: Disclosed are a method and an apparatus for accurately detecting a gesture from a user's motion using a user-based spatial coordinate system. A method for detecting a gesture in a user-based spatial coordinate system comprises the steps of: setting a user-based spatial coordinate system using a first body coordinate corresponding to a first body part of the user as a starting point; analyzing the motion vector state of a second body coordinate corresponding to a second body part of the user in the user-based spatial coordinate system over time; and detecting the user's gesture on the basis of a change in the motion vector state.
Type:
Application
Filed:
December 8, 2017
Publication date:
June 21, 2018
Applicant:
VTOUCH CO., LTD
Inventors:
Seokjoong KIM, Chunghoon KIM, Beomho LEE
Abstract: Disclosed is a virtual touch remote control apparatus and method using user's movements (gestures) capable of control electronic devices remotely regardless of existence of a display on the device, by providing additional information display with a projector. The apparatus includes a 3D coordinates calculation unit, an electronic device detection unit, a user detection unit, an electronic device operation controller, and an electronic device information DB for storing device specific information.
Abstract: Disclosed is a remote control apparatus and method using a virtual touch of a three-dimensionally modeled electronic device. The remote control apparatus includes a 3D coordinate calculation unit, an electronic device detecting unit, a user detecting unit, an electronic device driving control unit, and an electronic device information database. The remote control apparatus enables the user to remotely control an operation of an electronic device using a virtual touch through a motion or gesture of the user's finger.
Abstract: Disclosed is a virtual touch remote control apparatus and method using user's movements (gestures) capable of control electronic devices remotely regardless of existence of a display on the device, by providing additional information display with a projector. The apparatus includes a 3D coordinates calculation unit, an electronic device detection unit, a user detection unit, an electronic device operation controller, and an electronic device information DB for storing device specific information.
Abstract: The present invention is to provide a transparent display virtual touch apparatus capable of exquisitely operating by a display portion worn on a user's face and located in front of an eye of a user and of identifying contents regardless of the direction and location of the user.
Abstract: Provided is a 3D game device using a virtual touch, the three-dimensional game device using a virtual touch includes a 3D game executing unit rendering a 3D stereoscopic game pre-stored in a game database and generating a 3D stereoscopic image regarding the rendered 3D game to provide the 3D stereoscopic image to a display unit, and a virtual touch unit generating spatial coordinate data of a specific point of a user and image coordinate data from a user's viewpoint using the 3D stereoscopic image provided from the display unit and comparing the generated spatial coordinate data and image coordinate data to verify whether or not a specific point of a user contacts or approaches the 3D stereoscopic image and thus recognize a touch of the 3D stereoscopic image.
Abstract: Disclosed is a remote control apparatus and method using a virtual touch of a three-dimensionally modeled electronic device. The remote control apparatus includes a 3D coordinate calculation unit, an electronic device detecting unit, a user detecting unit, an electronic device driving control unit, and an electronic device information database. The remote control apparatus enables the user to remotely control an operation of an electronic device using a virtual touch through a motion or gesture of the user's finger.
Abstract: Provided is a virtual touch device for remotely controlling electronic equipment having a display surface. The virtual touch apparatus include an image acquisition unit, a spatial coordinate calculation unit, a touch location calculation unit, and a virtual touch processing unit. The image acquisition unit includes two image sensors disposed at different locations and photographs a user's body at the front of the display surface. The spatial coordinate calculation unit calculates three-dimensional coordinate data of the user's body using an image from the image acquisition unit. The touch location calculation unit calculates a contact point coordinate where a straight line connecting between a first spatial coordinate and a second spatial coordinate meets the display surface using the first and second spatial coordinates received from the spatial coordinate calculation unit.