Abstract: The sharing of video editing techniques is performed by receiving a project description file, thumbnail graphics, and a location identifier at a first server, wherein the project description file is generated during editing of multimedia content by the user, and wherein the thumbnail graphics represent multimedia editing objects incorporated into the edited multimedia content. The edited multimedia content is retrieved from a second server based on the location identifier. The edited multimedia content is synchronized with the thumbnail graphics and multimedia editing objects specified by the project description file. The synchronized edited multimedia content, thumbnail graphics, and multimedia editing objects are displayed.
Abstract: A method for estimating noise in a series of video frames includes selecting a region of a first frame, calculating high frequency quantities for the selected region, calculating temporal-domain high frequency quantities for the selected region of the frame and a mapping region of a second frame, and generating a noise estimate of the noise according to a relationship between the high frequency quantities and the temporal-domain high frequency quantities.
Abstract: Various embodiments are disclosed for detecting facial regions in a plurality of images. In one embodiment, a method comprises assigning at least one of the possible facial regions an assurance factor, forming clusters of possible facial regions based on a connection factor between the facial regions, and determining facial regions from the possible facial regions based on the assurance factor and the clusters of possible facial regions.
Type:
Grant
Filed:
August 24, 2011
Date of Patent:
April 8, 2014
Assignee:
Cyberlink Corp.
Inventors:
Chung-Yi Weng, Shih-Min Tang, Ho Chao Huang
Abstract: Disclosed are systems, methods, and computer readable media having programs for classifying sports video. In one embodiment, a method includes: extracting, from an audio stream of a video clip, a plurality of key audio components contained therein; and classifying, using at least one of the plurality of key audio components, a sport type contained in the video clip. In one embodiment, a computer readable medium having a computer program for classifying ports video includes: logic configured to extract a plurality of key audio components from a video clip; and logic configured to classify a sport type corresponding to the video clip.
Abstract: Disclosed are various embodiments for the selection of an object boundary shown in an image. An initial boundary probability map of an image is generated by one or more computing devices. A first input path segment is obtained, and a selection path is generated based on the first input path segment and the initial boundary probability map. A weighted boundary probability map is generated based on the selection path. A second input path segment is obtained, and a replacement selection path is generated based on the weighted boundary probability map and the second input path segment.
Abstract: Various embodiments are disclosed for performing inpainting. One embodiment is a method for editing a digital image in an image editing device. The method comprises obtaining an inpainting region in the digital image, determining a target resolution for scaling a resolution of the digital image based on an original resolution of the digital image, and determining an intermediate resolution level for scaling a resolution of the digital image based on the target resolution. The method further comprises scaling the resolution of the digital image to the intermediate resolution level, performing partial inpainting of the inpainting region at the intermediate resolution, and performing inpainting on a remainder portion in the inpainting region at a final target resolution.
Abstract: Included are embodiments of a method for collaborative editing. At least one embodiment includes receiving an indication to include at least one media effect with at least one media segment and sending at least one media attribute associated with the at least one media segment to a remote server. Some embodiments include receiving data related to at least one determined media effect, the at least one determined media effect being determined from the at least one media attribute.
Abstract: Systems and methods for translating user input in a high definition video playback system are disclosed. A disc viewer is configured to execute at least one user interactive program stored on a high definition optical disc defining a user interface. The user interactive program defines a user interface and a plurality of keyboard event handlers for interaction with the user interface. The user interface includes a plurality of selectable user interface elements. A user interface renderer is configured to encode for display the user interface. A non-keyboard input translator is configured to receive a non-keyboard user input translates it into at least one keyboard event.
Abstract: A method implemented in a video editing device comprises retrieving media content and generating a user interface comprising a graphical representation of the retrieved media content on a first timeline component. The method further comprises analyzing the retrieved media content to extract attributes associated with the media content and generating a second timeline component in the user interface. At least a portion of the extracted attributes is arranged along the second timeline component with respect to time, and each of the portion of extracted attributes is represented by a corresponding graphical representation. Furthermore, each attribute corresponds to a segment in the media content. The method further comprises retrieving, based on the displayed attributes arranged along the graphical timeline component, a selection of at least one segment of the media content.
Type:
Application
Filed:
May 9, 2013
Publication date:
January 23, 2014
Applicant:
Cyberlink Corp.
Inventors:
Jau-Hsiung Huang, Yuong-Wei Lei, Min-Kuang Wang
Abstract: A stereoscopic cursor method comprising: generating a virtual plane based on a variable cursor scene depth of a stereoscopic cursor for a stereoscopic user interface, the stereoscopic user interface comprising plural stereoscopic buttons, wherein the virtual plane is positioned between a viewer and the plural stereoscopic buttons; and causing the stereoscopic cursor to move along the virtual plane responsive to viewer input, the movement occurring smoothly and gradually adjacent to the plural stereoscopic buttons and varying in depth as the stereoscopic cursor progresses along the virtual plane from one of the plural stereoscopic buttons to another, the generating and causing performed by a processor in a computing device.
Abstract: In one embodiment, a two-dimensional to stereoscopic conversion method, comprising: estimating a local motion region in a first image relative to one or more second images, the first and the one or more second images comprising two-dimensional images; generating a color model based on the local motion region; calculating a similarity value for each of at least one image pixel selected from the first image based on the color model; and assigning a depth value for each of the at least one image pixel selected from the first image based on the calculated similarity value to generate a stereoscopic image, the method performed by one or more processors.
Abstract: A stereoscopic cursor method comprising: calculating a cursor scene depth of a stereoscopic cursor for a stereoscopic user interface comprising plural stereoscopic buttons, wherein the stereoscopic cursor is positioned between a viewer and the plural stereoscopic buttons; constraining movement of the stereoscopic cursor between the viewer and the plural stereoscopic buttons at the cursor scene depth for input device movements by the viewer that navigate across the front of the plural stereoscopic buttons; receiving an input signal corresponding to viewer selection of one of the plural stereoscopic buttons; and responsive to receiving the input signal, causing movement of the stereoscopic cursor from one end of the cursor scene depth to the one of the plural stereoscopic buttons in a direction coincident with the cursor scene depth.
Abstract: To modify a facial feature region in a video bitstream, the video bitstream is received and a feature region is extracted from the video bitstream. An audio characteristic, such as frequency, rhythm, or tempo is retrieved from an audio bitstream, and the feature region is modified according to the audio characteristic to generate a modified image. The modified image is outputted.
Abstract: To modify a facial feature region in a video bitstream, the video bitstream is received and a feature region is extracted from the video bitstream. An audio characteristic, such as frequency, rhythm, or tempo is retrieved from an audio bitstream, and the feature region is modified according to the audio characteristic to generate a modified image. The modified image is outputted.
Abstract: A remote video editing system is provided for editing a video from a remote location without the need to locally save the video every time an edit is made. Briefly described, one embodiment, among others, includes the steps of analyzing characteristics of the video and embedding the information into the video; storing the video and the embedded information at an editing server; receiving edit instructions from a client at the editing server; creating a project file containing the edit instructions received; and editing the video based on the project file.
Abstract: One embodiment, among others, includes a method for editing video. The method comprises receiving a plurality of media files and receiving specified special effects corresponding to each of the plurality of media files. Based on the received plurality of media files and the specified special effects for each of the plurality of media files, a single output is generated whereby the special effects are incorporated and where the plurality of media files are unmodified.
Abstract: To modify a facial feature region in a video bitstream, the video bitstream is received and a feature region is extracted from the video bitstream. An audio characteristic, such as frequency, rhythm, or tempo is retrieved from an audio bitstream, and the feature region is modified according to the audio characteristic to generate a modified image. The modified image is outputted.
Abstract: To browse images grouped by person on a photo displaying system, a plurality of facial regions are generated from a plurality of images of a plurality of albums. The facial regions are grouped into a plurality of clusters of faces based on similarity between the facial regions. The clusters of faces are associated with the albums. A signal is received to select one of the facial regions. The selected facial region belongs to a specific cluster of faces and the specific cluster of faces is associated with a specific album of the albums. Images of the specific album and facial regions of the clusters of faces are displayed.
Abstract: To modify a facial feature region in a video bitstream, the video bitstream is received and a feature region is extracted from the video bitstream. An audio characteristic, such as frequency, rhythm, or tempo is retrieved from an audio bitstream, and the feature region is modified according to the audio characteristic to generate a modified image. The modified image is outputted.
Abstract: Various embodiments are disclosed for modifying stereoscopic images. One embodiment is a method implemented in an image processing device for modifying stereoscopic images. The method comprises retrieving, by an image processing device, a stereoscopic image having at least a first view image and a second view image and retrieving an orientation selection relating to the stereoscopic image, the orientation selection comprising a selection other than one of: a horizontal flip selection and a 180 degree rotation selection. The method further comprises calculating a depth map according to at least part of the stereoscopic image, rotating the first view image based on the orientation selection to obtain a rotated first view image, and generating a new second view image according to the depth map and the rotated first view image.