METHOD AND SYSTEM FOR SOCIAL REMIXING OF MEDIA CONTENT

- Yahoo

Systems and methods have been developed for manipulating media assets in a networked computing environment where processing power, bandwidth and/or storage capacity may be limited.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
RELATED APPLICATIONS

This application claims the benefit of U.S. Provisional Application No. 60/758,664, filed Jan. 13, 2006, which application is hereby incorporated herein by reference, and U.S. Provisional Application No. 60/790,569, filed Apr. 10, 2006, which application is hereby incorporated herein by reference.

COPYRIGHT NOTICE

A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by any-one of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.

BACKGROUND

In the current Internet there exist many different types of media assets in the form of digital files. Digital files may contain data representing one or more types of content, including but not limited to, audio, images, and videos. For example, media assets include file formats such as MPEG-1 Audio Layer 3 (“MP3”) for audio, Joint Photographic Experts Group (“JPEG”) for images, Motion Picture Experts Group (“MPEG-2” and “MPEG-4”) for video, Adobe Flash for animations, and executable files.

Such media assets are currently created and edited using applications executing locally on a dedicated computer. For example, in the case of digital video, popular applications for creating and editing media assets include Apple's iMovie and FinalCut Pro and Microsoft's MovieMaker. After creation and editing a media asset, one or more files may be transmitted to a computer (e.g., a server) located on a distributed network such as the Internet. The server may host the files for viewing by different users. Examples of companies operating such servers are YouTube (http://youtube.com) and Google Video (http://video.google.com).

Presently, users must create and/or edit media assets on their client computers before transmitting the media assets to a server. Many users are therefore unable able to edit media assets from another client where, for example, the user's client computer does not contain the appropriate application or media asset for editing. Moreover, editing applications are typically designed for professional or high-end consumer markets. Such applications do not address the needs of average consumers who lack dedicated computers with considerable processing power and/or storage capacity.

Additionally, average consumers typically do not have the transmission bandwidth necessary to transfer, share or access media assets that may be widespread across a network. Increasingly, many media assets are stored on computer connected to the Internet. For example, services such as Getty Images sell media assets (e.g., images) that are stored on computers connected to the Internet. Thus, when a user requests a media asset for manipulation or editing, the asset is typically transferred in its entirety over the network. Particularly in the case of digital video, such transfers may consume tremendous processing and transmission resources.

SUMMARY

Against this backdrop systems and methods have been developed for manipulating media assets in a networked computing environment where processing power, bandwidth and/or storage capacity may be limited. More specifically, systems and methods have been developed whereby low-resolution media assets may be created that are optimized for transfer over low bandwidth networks and for editing and manipulation in an environment with low processing power and low storage capacity, and a high-resolution media asset may be created for playback.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for editing a low-resolution media asset to generate a high-resolution edited media asset. The method includes receiving a request to edit a first high-resolution media asset from a requester. The method further includes transmitting a low-resolution media asset to the requestor, the low-resolution media asset based upon the first high-resolution media asset. The method includes receiving from the requester an edit instruction associated with the low-resolution media asset. The method further includes generating a second high-resolution media asset based upon the first high-resolution media asset and the edit instruction associated with the low-resolution media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method for editing a low-resolution media asset to generate a high-resolution edited media asset. The computer readable medium includes instructions for receiving a request to edit a first high-resolution media asset from a requestor. The computer readable medium further includes instructions for transmitting a low-resolution media asset to the requester, the low-resolution media asset based upon the first high-resolution media asset. The computer readable medium includes instructions for receiving from a requestor an edit instruction associated with the low-resolution media asset. The computer readable medium further includes instructions for generating a second high-resolution media asset based upon the first high-resolution media asset and the edit instruction associated with the low-resolution media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a system. The system includes a high-resolution media asset library. The system further includes a low-resolution media asset generator that generates low-resolution media assets from high-resolution media assets contained in the high-resolution media asset library. The system includes a high-resolution media asset editor that applies edits to a high-resolution media asset based on edits made to an associated low-resolution media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method. The method includes receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a keyframe master asset. The method further includes generating a first portion of the video asset, the first portion containing one or more keyframes associated with the starting frame, the keyframes obtained from the keyframe master asset. The method includes generating a second portion of the video asset, the second portion containing sets of the keyframes and optimized frames, the optimized frames obtained from an optimized master asset associated with the keyframe master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method. The method includes receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a master asset. The method further includes generating a first portion of the video asset, the first portion containing one or more keyframes associated with the starting frame, the keyframes obtained from a keyframe master asset corresponding to the master asset. The method includes generating a second portion of the video asset, the second portion containing sets of the keyframes and optimized frames, the optimized frames obtained from an optimized master asset corresponding to the master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method. The method includes receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in an optimized master asset. The method further includes generating a keyframe master asset, based upon the optimized master asset, the keyframe master asset including one or more keyframes corresponding to the starting frame. The method includes generating a first portion of the video asset, the first portion including at least the starting frame identified in the optimized master asset. The method further includes generating a second portion of the video asset, the second portion including sets of the keyframes and optimized frames, the optimized frames obtained from the optimized master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method. The computer readable medium includes instructions for receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a keyframe master asset. The computer readable medium further includes instructions for generating a first portion of the video asset, the first portion containing one or more keyframes associated with the starting frame, the keyframes obtained from the keyframe master asset. The computer readable medium includes instructions for generating a second portion of the video asset, the second portion containing sets of the keyframes and optimized frames, the optimized frames obtained from an optimized master asset associated with the keyframe master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method. The computer readable medium includes instructions for receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a master asset. The computer readable medium further includes instructions for generating a first portion of the video asset, the first portion containing one or more keyframes associated with the starting frame, the keyframes obtained from a keyframe master asset corresponding to the master asset. The computer readable medium includes instructions for generating a second portion of the video asset, the second portion containing sets of the keyframes and optimized frames, the optimized frames obtained from an optimized master asset corresponding to the master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method. The computer readable medium includes instructions for receiving a request to generate a video asset, the video asset identifying a starting frame and an ending frame in an optimized master asset. The computer readable medium further includes instructions for generating a keyframe master asset, based upon the optimized master asset, the keyframe master asset including one or more keyframes corresponding to the starting frame. The computer readable medium includes instructions for generating a first portion of the video asset, the first portion including at least the starting frame identified in the optimized master asset. The computer readable medium further includes instructions for generating a second portion of the video asset, the second portion including sets of the keyframes and optimized frames, the optimized frames obtained from the optimized master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a system. The system includes a master asset library that stores at least one high-resolution master asset. The system further includes a specification applicator that stores at least one edit specification for applying edits to the at least one high-resolution master asset. The system includes a master asset editor that applies the at least one edit specification to the at least one high-resolution master asset. The system further includes an edit asset generator that generates a low-resolution asset corresponding to the high-resolution master asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method. The method includes editing a low-resolution media asset, the low-resolution media asset corresponding to a master high-resolution media asset. The method further includes generating an edit specification based on the editing of the low-resolution media asset. The method includes applying the edit specification to the master high-resolution media asset to create an edited high-resolution media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer-readable medium having stored thereon a data structure. The computer readable medium includes a first data field comprising data identifying a high-resolution media asset. The computer readable medium further includes a second data field comprising data describing one or more edits made to a low-resolution media asset associated with the high-resolution media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for identifying edit information of a media asset. The method includes editing a low resolution media asset, the low resolution media asset containing at least a first portion corresponding to a first high-resolution master media asset and a second portion corresponding to a second high-resolution master media asset. The method further includes receiving a request to generate a high-resolution edited media asset, the request identifying the first high-resolution master media asset and the second high-resolution master media asset. The method includes generating the high-resolution edited media asset. The method further includes associating with the high-resolution edited media asset edit information that identifies the first high-resolution master media asset and the second high-resolution master media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method for identifying edit information of a media asset. The method includes editing a low resolution media asset, the low resolution media asset containing at least a first portion corresponding to a first high-resolution master media asset and a second portion corresponding to a second high-resolution master media asset. The method further includes receiving a request to generate a high-resolution edited media asset, the request identifying the first high-resolution master media asset and the second high-resolution master media asset. The method includes generating the high-resolution edited media asset. The method further includes associating with the high-resolution edited media asset edit information that identifies the first high-resolution master media asset and the second high-resolution master media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for rendering a media asset. The method includes receiving a command to render an aggregate media asset defined by an edit specification, the edit specification identifying at least a first media asset associated with at least one edit instruction. The method further includes retrieving the edit specification. The method includes retrieving the first media asset. The method further includes rendering, on a media asset rendering device, the first media asset of the aggregate media asset in accordance with the at least one edit instruction.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method for rendering a media asset. The method includes receiving a command to render an aggregate media asset defined by an edit specification, the edit specification identifying at least a first media asset associated with at least one edit instruction. The method further includes retrieving the edit specification. The method includes retrieving the first media asset. The method further includes rendering, on a media asset rendering device, the first media asset of the aggregate media asset in accordance with the at least one edit instruction.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for editing an aggregate media asset. The method includes, in a playback session, receiving a stream corresponding to an aggregate media asset from a remote computing device, the aggregate media asset comprised of at least one component media asset. The method further includes rendering the aggregate media asset on an image rendering device. The method includes receiving a user command to edit an edit specification associated with the aggregate media asset. The method further includes initiating an edit session for editing the edit specification associated with the aggregate media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a computer readable medium encoding or containing computer executable instructions for performing a method for editing an aggregate media asset. The method includes, in a playback session, receiving a stream corresponding to an aggregate media asset from a remote computing device, the aggregate media asset comprised of at least one component media asset. The method further includes rendering the aggregate media asset on an image rendering device. The method includes receiving a user command to edit an edit specification associated with the aggregate media asset. The method further includes initiating an edit session for editing the edit specification associated with the aggregate media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for storing an aggregate media asset. The method includes storing a plurality of component media assets. The method further includes storing a first aggregate edit specification, the first aggregate edit specification including at least one command for rendering the plurality of component media assets to generate a first aggregate media asset.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for publishing a composite media asset and receiving compensation therefor. The method includes receiving, from a first remote computing device, an edit specification for rendering a composite media asset, the edit specification identifying at least a first component media asset and a second component media asset that comprise the composite media asset. The method further includes storing the edit specification. The method includes receiving, from a second remote computing device, a request to render at least the first component media asset based upon information in the edit specification. The method further includes transmitting, to the second remote computing device, information identifying the first component media asset. The method includes receiving compensation for the transmitting step.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for inserting advertising into a composite media asset and receiving compensation therefor. The method includes receiving a composite media asset edit specification defining a composite media asset, the composite media asset edit specification identifying a first component media asset and a second component media asset. The method further includes receiving a request, by one or more vendors, to incorporate at least one advertisement with the composite media asset. The method includes incorporating the at least one advertisement with the composite media asset. The method further includes receiving compensation from the one or more vendors for the incorporating step.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for inserting advertising into a composite media asset and receiving compensation therefor. The method includes receiving a composite media asset edit specification defining a composite media asset, the composite media asset edit specification identifying a first component media asset and a second component media asset. The method further includes receiving a request, by one or more vendors, to incorporate at least one advertisement with the composite media asset. The method includes incorporating the advertisement as part of the composite media asset. The method further includes transmitting the composite media asset to a remote computing device. The method includes, upon rendering of at least a portion of the composite media asset at the remote computing device, receiving compensation from the one or more vendors.

In one example (which example is intended to be illustrative and not restrictive), the present invention may be considered a method for reusing component media assets and receiving compensation therefor. The method includes maintaining a library including at least a first edit specification corresponding to a first composite media asset, the first composite media asset including at least one component media asset. The method further includes receiving a request, from a remote computing device, to generate a second edit specification for a second composite media asset, the second edit specification for the second composite media asset identifying the at least one component media asset and the second composite media asset including the at least one component media asset. The method includes generating the second edit specification. The method further includes receiving compensation for the generating step.

These and various other features as well as advantages will be apparent from a reading of the following detailed description and a review of the associated drawings. Additional features are set forth in the description which follows, and in part will be apparent from the description, or may be learned by practice of the described embodiments. The benefits and features will be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.

It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of the invention as claimed.

BRIEF DESCRIPTION OF THE DRAWINGS

The following drawing figures, which form a part of this application, are illustrative of embodiments systems and methods described below and are not meant to limit the scope of the invention in any manner, which scope shall be based on the claims appended hereto.

FIG. 1 illustrates an embodiment of a system for manipulating a media asset in a networked computing environment.

FIG. 2 illustrates an embodiment of a system for manipulating a media asset in a networked computing environment.

FIG. 3 illustrates an embodiment of a method for editing a low-resolution media asset to generate a high-resolution edited media asset.

FIG. 4 illustrates an embodiment of a method for generating a media asset.

FIG. 5 illustrates an embodiment of a method for generating a media asset.

FIG. 6 illustrates an embodiment of a method for generating a media asset.

FIG. 7 illustrates an embodiment of a method for recording edits to media content.

FIG. 8 illustrates an embodiment of a method for identifying edit information of a media asset.

FIG. 9 illustrates an embodiment of a method for rendering a media asset.

FIG. 10 illustrates an embodiment of a method for storing an aggregate media asset.

FIG. 11 illustrates an embodiment of a method for editing an aggregate media asset.

FIG. 12 illustrates an embodiment of a method for publishing a composite media asset and receiving compensation therefor.

FIG. 13 illustrates an embodiment of a method for inserting advertising into a composite media asset and receiving compensation therefor.

FIG. 14 illustrates an embodiment of another method for inserting advertising into a composite media asset and receiving compensation therefor.

FIG. 15 illustrates a method for reusing component media assets and receiving compensation therefor.

DETAILED DESCRIPTION

FIG. 1 illustrates an embodiment of a system 100 for generating a media asset. In one embodiment, a system 100 is comprised of a master asset library 102. In one embodiment, a master asset library 102 may be a logical grouping of data, including but not limited to high-resolution and low-resolution media assets. In another embodiment, a master asset library 102 may be a physical grouping of data, including but not limited to high-resolution and low-resolution media assets. In an embodiment, a master asset library 102 may be comprised of one or more databases and reside on one or more servers. In one embodiment, master asset library 102 may be comprised of a plurality of libraries, including public, private, and shared libraries. In one embodiment, a master asset library 102 may be organized into a searchable library. In another embodiment, the one or more servers comprising master asset library 102 may include connections to one or more storage devices for storing digital files.

For purposes of this disclosure, the drawings associated with this disclosure, and the appended claims, the term “files” generally refers to a collection of information that is stored as a unit and that, among other things, may be retrieved, modified, stored, deleted or transferred. Storage devices may include, but are not limited to, volatile memory (e.g., RAM, DRAM), non-volatile memory (e.g., ROM, EPROM, flash memory), and devices such as hard disk drives and optical drives. Storage devices may store information redundantly. Storage devices may also be connected in parallel, in a series, or in some other connection configuration. As set forth in the present embodiment, one or more assets may reside within a master asset library 102.

For purposes of this disclosure, the drawings associated with this disclosure, and the appended claims, an “asset” refers to a logical collection of content that may be comprised within one or more files. For example, an asset may be comprised of a single file (e.g., an MPEG video file) that contains images (e.g., a still frame of video), audio, and video information. As another example, an asset may also may be comprised of a collection of files (e.g., JPEG image files) that collectively may be used to render an animation or video. As yet another example, an asset may also comprise an executable file (e.g., an executable vector graphics file, such as an SWF file or an FLA file). A master asset library 102 may include many types of assets, including but not limited to, video, images, animations, text, executable files, and audio. In one embodiment, master asset library 102 may include one or more high-resolution master assets. For the remainder of this disclosure, “master asset” will be disclosed as a digital file containing video content. One skilled in the art will recognize, however, that a master asset is not limited to containing video information, and as set forth previously, a master asset may contain many types of information including but not limited to images, audio, text, executable files, and/or animations.

In one embodiment, a media asset may be stored in a master asset library 102 so as to preserve the quality of the media asset. For example, in the case of a media asset comprising video information, two important aspects of video quality are spatial resolution and temporal resolution. Spatial resolution generally describes the clarity of lack of blurring in a displayed image, while temporal resolution generally describes the smoothness of motion. Motion video, like film, consists of a certain number of frames per second to represent motion in the scene. Typically, the first step in digitizing video is to partition each frame into a large number of picture elements, or pixels or pels for short. The larger the number of pixels, the higher the spatial resolution. Similarly, the more frames per second, the higher the temporal resolution.

In one embodiment, a media asset may be stored in a master asset library 102 as a master asset that is not directly manipulated. For example, a media asset may be preserved in a master asset library 102 in its original form, although it may still be used to create copies or derivative media assets (e.g., low-resolution assets). In one embodiment, a media asset may also be stored in a master asset library 102 with corresponding or associated assets. In one embodiment, a media asset stored in a master asset library 102 may be stored as multiple versions of the same media asset. For example, multiple versions of a media asset stored in master asset library 102 may include an all-keyframe version that does not take advantage of intra-frame similarities for compression purposes, and an optimized version that does take advantage of intra-frame similarities. In one embodiment, the original media asset may represent an all-keyframe version. In another embodiment, the original media asset may originally be in the form of an optimized version or stored as an optimized version. One skilled in the art will recognize that media assets may take many forms within a master asset library 102 that are within the scope of this disclosure.

In one embodiment, a system 100 is also comprised of an edit asset generator 104. In an embodiment, an edit asset generator 104 may be comprised of transcoding hardware and/or software that, among other things, may convert a media asset from one format into another format. For example, a transcoder may be used to convert an MPEG file into a Quicktime file. As another example, a transcoder may be used to convert a JPEG file into a bitmap (e.g., *.BMP) file. As yet another example, a transcoder may standardize media asset formats into an Flash video file (*.FLV) format. In one embodiment, a transcoder may create more than one versions of an original media asset. For example, upon receiving an original media asset, a transcoder may convert the original media asset into a high-resolution version and a low-resolution version. As another example, a transcoder may convert an original media asset into one or more files. In one embodiment, a transcoder may exist on a remote computing device. In another embodiment, a transcoder may exist on one or more connected computers. In one embodiment, an edit asset generator 104 may also be comprised of hardware and/or software for transferring and/or uploading media assets to one or more computers. In another embodiment, an edit asset generator 104 may be comprised of or connected to hardware and/or software used to capture media assets from external sources such as a digital camera.

In one embodiment, an edit asset generator 104 may generate a low-resolution version of a high-resolution media asset stored in a master asset library 102. In another embodiment, an edit asset generator 104 may transmit a low-resolution version of a media asset stored in a master asset library 102, for example, by converting the media asset in real-time and transmitting the media asset as a stream to a remote computing device. In another embodiment, an edit asset generator 104 may generate a low quality version of another media asset (e.g., a master asset), such that the low quality version preserves while still providing sufficient data to enable a user to apply edits to the low quality version.

In one embodiment, a system 100 may also be comprised of a specification applicator 106. In one embodiment, a specification applicator 106 may be comprised of one or more files or edit specifications that include instructions for editing and modifying a media asset (e.g., a high-resolution media asset). In one embodiment, a specification applicator 106 may include one or more edit specifications that comprise modification instructions for a high-resolution media asset based upon edits made to a corresponding or associated low-resolution media asset. In one embodiment, a specification applicator 106 may store a plurality of edit specifications in one or more libraries.

In one embodiment, a system 100 is also comprised of a master asset editor 108 that may apply one or more edit specifications to a media asset. For example, a master asset editor 108 may apply an edit specification stored in a specification applicator 106 library to a first high-resolution media asset and thereby creates another high-resolution media asset, e.g., a second high-resolution media asset. In one embodiment, a master asset editor 108 may apply an edit specification to a media asset in real-time. For example, a master asset editor 108 may modify a media asset as the media asset is transmitted to another location. In another embodiment, a master asset editor 108 may apply an edit specification to a media asset in non-real-time. For example, a master asset editor 108 may apply edit specifications to a media asset as part of a scheduled process. In one embodiment, a master asset editor 108 may be used to minimize the necessity of transferring large media assets over a network. For example, by storing edits in an edit specification, a master asset editor 108 may transfer small data files across a network to effectuate manipulations made on a remote computing device to higher quality assets stored on one or more local computers (e.g., computers comprising a master asset library).

In another embodiment, a master asset editor 108 may be responsive to commands from a remote computing device (e.g., clicking a “remix” button at a remote computing device may command the master asset editor 108 to apply an edit specification to a high-resolution media asset). For example, a master asset editor 108 may dynamically and/or interactively apply an edit specification to a media asset upon a user command issuing from a remote computing device. In one embodiment, a master asset editor 108 may dynamically apply an edit specification to a high-resolution to generate an edited high-resolution media asset for playback. In another embodiment, a master asset editor 108 may apply an edit specification to a media asset on a remote computing device and one or more computers connected by a network (e.g., Internet 114). For example, bifurcating the application of an edit specification may minimize the size of the edited high-resolution asset prior to transferring it to a remote computing device for playback. In another embodiment, a master asset editor 108 may apply an edit specification on a remote computing device, for example, to take advantage of vector-based processing that may be executed efficiently on a remote computing device at playtime.

In one embodiment, a system 100 is also comprised of an editor 110 that may reside on a remote computing device 112 that is connected to one or more networked computers, such as the Internet 114. In one embodiment, an editor 110 may be comprised of software. For example, an editor 110 may be a stand-alone program. As another example, an editor 110 may be comprised of one or more instructions that may be executed through another program such as an Internet 114 browser (e.g., Microsoft Internet Explorer). In one embodiment, an editor 110 may be designed with a user interface similar to other media-editing programs. In one embodiment, an editor 110 may contain connections to a master asset library 102, an edit asset library 104, a specification applicator 106 and/or a master asset editor 108. In one embodiment, an editor 110 may include pre-constructed or “default” edit specifications that may be applied by a remote computing device to a media asset. In one embodiment, an editor 110 may include a player program for displaying media assets and/or applying one or more instructions from an edit specification upon playback of a media asset. In another embodiment, an editor 110 may be connected to a player program (e.g., a standalone editor may be connected to a browser).

FIG. 2 illustrates an embodiment of a system 200 for generating a media asset. In one embodiment, the system 200 comprises a high-resolution media asset library 202. In one embodiment, the high-resolution media asset library 202 may be a shared library, a public library, and/or a private library. In one embodiment, the high-resolution media asset library 202 may include at least one video file. In another embodiment, the high resolution media asset library 202 may include at least one audio file. In yet another embodiment, the high-resolution media asset library 202 may include at least one reference to a media asset residing on a remote computing device 212. In one embodiment, the high-resolution media asset library 202 may reside on a plurality of computing devices.

In one embodiment, the system 200 further comprises a low-resolution media asset generator 204 that generates low-resolution media assets from high-resolution media assets contained in the high-resolution media asset library. For example, as discussed above, a low-resolution media asset generator 204 may convert a high-resolution media asset to a low-resolution media asset.

In one embodiment, the system 200 further comprises a low-resolution media asset editor 208 that transmits edits made to an associated low-resolution media asset to one or more computers via a network, such as the Internet 214. In another embodiment, the low-resolution media asset editor 208 may reside on a computing device remote from the high resolution media asset editor, for example, remote computing device 212. In another embodiment, the low-resolution media asset editor 208 may utilize a browser. For example, the low-resolution media asset editor 208 may store low-resolution media assets in the cache of a browser.

In one embodiment, the system 200 may also comprise an image rendering device 210 that displays the associated low-resolution media asset. In one embodiment, an image rendering device 210 resides on a computing device 212 remote from the high-resolution media asset editor 206. In another embodiment, an image rendering device 210 may utilize a browser.

In one embodiment, the system 200 further comprises a high-resolution media asset editor 206 that applies edits to a high-resolution media asset based on edits made to an associated low-resolution media asset.

FIG. 3 illustrates an embodiment of a method 300 for editing a low-resolution media asset to generate a high-resolution edited media asset. In the method 300, a request to edit a first high-resolution media is received from a requestor in a requesting operation 302. In one embodiment, the first high-resolution media asset may be comprised of a plurality of files and receiving a request to edit the first high-resolution media asset in requesting operation 302 may further comprise receiving a request to edit at least one of the plurality of files. In another embodiment, requesting operation 302 may further comprise receiving a request to edit at least one high-resolution audio or video file.

In the method 300, a low-resolution media asset based upon the first high-resolution media asset is transmitted to a requester in a transmitting operation 304. In one embodiment, transmitting operation 304 may comprise transmitting at least one low-resolution audio or video file. In another embodiment, transmitting operation 304 may further comprise converting at least one high-resolution audio or video file associated with a first high-resolution media asset from a first file format into at least one low-resolution audio or video file, respectively, having a second file format. For example, a high-resolution uncompressed audio file (e.g., a WAV file) may be converted into a compressed audio file (e.g., an MP3 file). As another example, a compressed file with a lesser compression ratio may be converted into a file of the same format, but formatted with a greater compression ratio.

The method 300 then comprises receiving from a requestor an edit instruction associated with a low-resolution media asset in receiving operation 306. In one embodiment, receiving operation 306 may further comprise receiving an instruction to modify a video presentation property of at least one high-resolution video file. For example, modification of a video presentation property may include receiving an instruction to modify an image aspect ratio, a spatial resolution value, a temporal resolution value, a bit rate value, or a compression value. In another embodiment, receiving operation 306 may further comprise receiving an instruction to modify a timeline (e.g., sequence of frames) of at least one high-resolution video file.

The method 300 further comprises generating a second high-resolution media asset based upon the first high-resolution media asset and the edit instruction associated with the low-resolution media asset in a generating operation 308. In one embodiment of generating operation 308, an edit specification is applied to at least one high-resolution audio or video file comprising the first high-resolution media asset. In a further embodiment, generating operation 308 generates at least one high-resolution audio or video file. In another embodiment, generating operation 308 further comprises the steps of: generating a copy of at least one high-resolution audio or video file associated with a first high-resolution media asset; applying the edit instruction, respectively, to the at least one high-resolution audio or video file; and saving the copy as a second high-resolution media asset.

In another embodiment of method 300, at least a portion of the second high-resolution media asset may be transmitted to a remote computing device. In still yet another embodiment of method 300, at least a portion of the second high-resolution media asset may be displayed by an image rendering device. For example, the image rendering device may take the form of a browser residing at a remote computing device.

FIG. 4 illustrates an embodiment of a method 400 for generating a media asset. In the method 400, a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a keyframe master asset, is received in receiving operation 402. For example, the request of receiving operation 402 may identify a first portion and/or a second portion of a video asset.

In a generating a first portion operation 404, the method 400 then comprises generating a first portion of the video asset where the first portion contains one or more keyframes associated with the starting frame and the keyframes are obtained from the keyframe master asset. For example, where the keyframe master asset comprises an uncompressed video file, one or more frames of the uncompressed video file may comprise the keyframes associated with the starting frame of the media asset.

In a generating a second portion operation 406, the method 400 further comprises generating a second portion of the video asset where the second portion contains sets of the keyframes and optimized frames and the optimized frames obtained from an optimized master asset associated with the keyframe master asset. For example, where the optimized master asset comprises a compressed video file, a set of frames that are compressed may be combined in a video asset with one or more uncompressed frames from an uncompressed video file.

In another embodiment of method 400, a library of master assets may be maintained such that a keyframe master asset and an optimized master asset may be generated corresponding to at least one of the library master assets. In still yet another embodiment of method 400, a request may identify a starting keyframe or ending keyframe in a keyframe master asset that corresponds, respectively, to a starting frame or ending frame.

FIG. 5 illustrates an embodiment of a method 500 for generating a media asset. In the method 500, a request to generate a video asset, the video asset identifying a starting frame and an ending frame in a master asset, is received in receiving operation 502. For example, the request of receiving operation 502 may identify a first portion and/or a second portion of a video asset.

In a generating a first portion operation 504, the method 500 then comprises generating a first portion of the video asset where the first portion contains one or more keyframes associated with the starting frame and the keyframes obtained from a keyframe master asset correspond to a master asset. For example,

In a generating a second portion operation 506, the method 500 then comprises generating a second portion of the video asset where the second portion contains sets of the keyframes and optimized frames and the optimized frames obtained from an optimized master asset correspond to a master asset. For example, where the optimized master asset comprises a compressed video file, a set of frames that are compressed may be combined in a video asset with one or more uncompressed keyframes from a keyframe master asset.

In another embodiment of method 500, a library of master assets may be maintained such that a keyframe master asset and an optimized master asset may be generated corresponding to at least one of the library master assets. In still yet another embodiment of method 500, a request may identify a starting keyframe or ending keyframe in a keyframe master asset that corresponds, respectively, to a starting frame or ending frame.

FIG. 6 illustrates an embodiment of a method 600 for generating a media asset. In the method 600, a request to generate a video asset where the video asset identifies a starting frame and an ending frame in an optimized master asset is received in a receiving operation 602. For example, the request of receiving operation 602 may identify a first portion and/or a second portion of a video asset.

The method 600 then comprises generating a keyframe master asset, based upon the optimized master asset, that includes one or more keyframes corresponding to the starting frame in a generating a keyframe operation 604. In a generating a first portion operation 606, the method 600 further comprises generating a first portion of the video asset where the first portion includes at least a starting frame identified in an optimized master asset. In a generating a second portion operation 608, the method 600 then further comprises generating a second portion of the video asset where the second portion includes sets of keyframes and optimized frames and the optimized frames are obtained from the optimized master asset.

In another embodiment of method 600, a library of master assets may be maintained such that a keyframe master asset and an optimized master asset may be generated corresponding to at least one of the library master assets. In still yet another embodiment of method 600, a request may identify a starting keyframe or ending keyframe in a keyframe master asset that corresponds, respectively, to a starting frame or ending frame.

FIG. 7 illustrates an embodiment of a method 700 for recording edits to media content. In the method 700, a low-resolution media asset corresponding to a master high-resolution media asset is edited in editing operation 702. In one embodiment, editing comprises modifying an image of a low-resolution media asset that corresponds to a master high-resolution media asset. For example, where an image includes pixel data, the pixels may be manipulated such that they appear in a different color or with a different brightness. In another embodiment, editing comprises modifying the duration of a low-resolution media asset corresponding to a duration of a master high-resolution media asset. For example, modifying a duration may include shortening a low-resolution media asset and the high-resolution media asset corresponding to the low-resolution media asset.

In a further embodiment, where the master high-resolution media asset and the low-resolution media asset comprise at least one or more frames of video information, the editing comprises modifying a transition property of the at least one or more frames of video information of a low-resolution media asset that corresponds to a master high-resolution media asset. For example, a transition such as a fade-in or fade-out transition may replace an image of one frame with an image of another frame. In another embodiment, editing comprises modifying a volume value of an audio component of a low-resolution media asset corresponding to a master high-resolution media asset. For example, a media asset including video information may include an audio track that may be played louder or softer depending upon whether a greater or lesser volume value is selected.

In another embodiment, where the master high-resolution media asset and the low-resolution media asset comprise at least two or more frames of sequential video information, editing comprises modifying the sequence of the at least two or more frames of sequential video information of a low-resolution media asset corresponding to a master high-resolution media asset. For example, a second frame may be sequenced prior to a first frame of a media asset comprising video information.

In still yet another embodiment, editing comprises modifying one or more uniform resource locators (e.g., URLs) associated with a low-resolution media asset corresponding to a master high-resolution media asset. In still another embodiment, editing comprises modifying a playback rate (e.g., 30 frames per second) of the low-resolution media asset corresponding to the master high-resolution media asset. In yet another embodiment, editing comprises modifying the resolution (e.g., the temporal or spatial resolution) of a low-resolution media asset corresponding to a master high-resolution media asset. In one embodiment, editing may occur on a remote computing device. For example, the edit specification itself may be created on a remote computing device. Similarly, for example, the edited high-resolution media asset may be transmitted to the remote computing device for rendering on an image rendering device such as a browser.

The method 700 then comprises generating an edit specification based on the editing of the low-resolution media asset in a generating operation 704. The method 700 further comprises applying the edit specification to the master high-resolution media asset to create an edited high-resolution media asset in an applying operation 706. In one embodiment, the method 700 further comprises rendering an edited high-resolution media asset on an image-rendering device. For example, rendering an edited high-resolution media asset may itself comprise applying a media asset filter to the edited high-resolution media asset. As another example, applying the media asset filter may comprise overlaying the edited high-resolution media asset with an animation. As yet another example, applying the media asset filter may further comprise changing a display property of the edited high-resolution media asset. Changing a display property may include, but is not limited to, changing a video presentation property. In this example, applying the media asset filter may comprise changing a video effect, a title, a frame rate, a trick-play effect (e.g., a media asset filter may change a fast-forward, pause, slow-motion and/or rewind operation), and/or a composite display (e.g., displaying at least a portion of two different media assets at the same time, such as in the case of picture-in-picture and/or green-screen compositions). In another embodiment, the method 700 may further comprise storing an edit specification. For example, an edit specification may be stored at a remote computing device or one or more computers connected via a network, such as via the Internet.

FIG. 8 illustrates an embodiment of a method 800 for identifying edit information of a media asset. In the method 800, a low-resolution media asset is edited in an editing operation 802 where the low-resolution media asset contains at least a first portion corresponding to a first high-resolution master media asset and a second portion corresponding to a second high-resolution master media asset. In one embodiment, editing operation 802 further comprises storing at least some of the edit information as metadata with a high-resolution edited media asset. In another embodiment, editing operation 802 may occur on a remote computing device.

In receiving operation 804, the method 800 then comprises receiving a request to generate a high-resolution edited media asset where the request identifies a first high-resolution master media asset and a second high-resolution master media asset. The method 800 then comprises generating a high-resolution edited media asset in a generating operation 806. The method 800 further comprises associating with a high-resolution edited media asset edit information that identifies the first high-resolution master media asset and the second high-resolution master media asset in an associating operation 808.

In one embodiment, method 800 further comprises retrieving either a first high-resolution master media asset or a second high-resolution master media asset. In yet another embodiment, method 800 still further comprises assembling a retrieved first high-resolution media asset and a retrieved second high-resolution media asset into a high-resolution edited media asset.

FIG. 9 illustrates an embodiment of a method 900 for rendering a media asset. In the method 900, a command to render an aggregate media asset defined by an edit specification, where the edit specification identifies at least a first media asset associated with at least one edit instruction, is received in receiving operation 902. In one embodiment, receiving operation 902 comprises an end-user command. In another embodiment, receiving operation 902 may comprise a command issued by a computing device, such as a remote computing device. In yet another embodiment, receiving operation 902 may be comprised of a series of commands that together represents a command to render an aggregate media asset defined by an edit specification.

In edit specification retrieving operation 904, an edit specification is retrieved. In an embodiment, retrieving operation 904 may comprise retrieving an edit specification from memory or some other storage device. In another embodiment, retrieving operation 904 may comprise retrieving an edit specification from a remote computing device. In yet another embodiment, retrieving an edit specification in retrieving operation 904 may comprise retrieving several edit specifications that collectively comprise a single related edit specification. For example, several edit specifications may be associated with different media assets (e.g., the acts of a play may each comprise a media asset) that together comprise a single related edit specification (e.g., for an entire play, inclusive of each act of the play). In one embodiment, the edit specification may identify a second media asset associated with a second edit instruction that may be retrieved and rendered on a media asset rendering device.

In media asset retrieving operation 906, a first media asset is retrieved. In one embodiment, retrieving operation 906 may comprise retrieving a first media asset from a remote computing device. In another embodiment, retrieving operation 906 may comprise retrieving a first media asset from memory or some other storage device. In yet another embodiment, retrieving operation 906 may comprise retrieving a certain portion (e.g., the header or first part of a file) of a first media asset. In another embodiment of retrieving operation 906, a first media asset may be comprised of multiple sub-parts. Following the example set forth in retrieving operation 904, a first media asset in the form of a video (e.g., a play with multiple acts) may be comprised of media asset parts (e.g., multiple acts represented as distinct media assets). In this example, the edit specification may contain information that links together or relates the multiple different media assets into a single related media asset.

In rendering operation 908, the first media asset of the aggregate media asset is rendered on a media asset rendering device in accordance with the at least one edit instruction. In one embodiment, the edit instruction may identify or point to a second media asset. In one embodiment, the media asset rendering device may be comprised of a display for video information and speakers for audio information. In an embodiment where there exists a second media asset, the second media asset may include information that is similar to the first media asset (e.g., both the first and second media assets may contain audio or video information) or different from the first media asset (e.g., the second media asset may contain audio information, such as a commentary of a movie, whereas the first media asset may contain video information, such as images and speech, for a movie). In another embodiment, rendering operation 908 may further include an edit instruction that modifies a transition property for transitioning from a first media asset to a second media asset, that overlays effects and/or titles on an asset, that combines two assets (e.g., combinations resulting from edit instructions directed towards picture-in-picture and/or green-screen capabilities), that modifies the frame rate and/or presentation rate of at least a portion of a media asset, that modifies the duration of the first media asset, that modifies a display property of the first media asset, or that modifies an audio property of the first media asset.

FIG. 10 illustrates an embodiment of a method 1000 for storing an aggregate media asset. In the method 1000, a plurality of component media assets are stored in storing operation 1002. For example, by way of illustration and not of limitation, storing operation 1002 may comprise caching at least one of the plurality of component media assets in memory. As another example, one or more component media assets may be cached in the memory cache reserved for a program such as an Internet browser.

In storing operation 1004, a first aggregate edit specification is stored where the first aggregate edit specification includes at least one command for rendering the plurality of component media assets to generate a first aggregate media asset. For example, an aggregate media asset may comprise one or more component media assets containing video information. In this example, the component videos may be ordered such that they may be rendered in a certain order as an aggregate video (e.g., a video montage). In one embodiment, storing operation 1004 comprises storing at least one command to display, in a sequence, a first portion of the plurality of component media assets. For example, the command to display may modify the playback duration of a component media asset including video information. In another embodiment of storing operation 1004, at least one command to render an effect corresponding to at least one of the plurality of component media assets may be stored. As one example, storing operation 1004 may include one or more effects that command transitions between component media assets. In still yet another embodiment of storing operation 1004, a second aggregate edit specification, the second aggregate edit specification including at least one command for rendering the plurality of component media assets to generate a second aggregate media asset may be stored.

FIG. 11 illustrates an embodiment of a method for editing an aggregate media asset. In the method 1100, a stream corresponding to an aggregate media asset from a remote computing device, the aggregate media asset comprised of at least one component media asset, is received in a playback session in receiving operation 1102. For example, a playback session may be comprised of a user environment that permits playback of a media asset. As another example, a playback session may be comprised of one or more programs that may display one or more files. Following this example, a playback session may be comprised of an Internet browser that is capable of receiving a streaming aggregate media asset. In this example, the aggregate media asset may be comprised of one or more component media assets residing on remote computing devices. The one or more component media assets may be streamed so as to achieve bandwidth and processing efficiency on a local computing device.

In a rendering operation 1104, the aggregate media asset is rendered on an image rendering device. For example, the aggregate media asset may be displayed such that pixel information from an aggregate media asset including video information is shown. In a receiving operation 1106, a user command to edit an edit specification associated with the aggregate media asset is received. As discussed previously, edit specifications may take many forms, including but not limited to one or more files containing metadata and other information associated with the component media assets that may be associated with an aggregate media asset.

In an initiating operation 1108, an edit session is initiated for editing the edit specification associated with the aggregate media asset. In one embodiment, initiating operation 1108 comprises displaying information corresponding to the edit specification associated with the aggregate media asset. For example, an editing session may permit a user to adjust the duration of a certain component media asset. In another embodiment, method 1100 further comprises modifying the edit specification associated with the aggregate media asset, thereby altering the aggregate media asset. Following the previous example, once a component media asset is edited in the editing session, the edits to the component media asset may be made to the aggregate media asset.

FIG. 12 illustrates an embodiment of a method 1200 for publishing a composite media asset and receiving compensation therefor. In the method 1200, an edit specification for rendering a composite media asset is received from a first remote computing device, where the edit specification identifies at least a first component media asset and a second component media asset that comprise the composite media asset, in receiving operation 1202. In storing operation 1204, the edit specification is stored. For example, the edit specification may be stored in a library residing on or connected to one or more computing devices. As another example, the edit specification may be stored in memory.

In receiving operation 1206, a second remote computing device receives a request to render at least the first component media asset based upon information in the edit specification. In one embodiment, receiving the request to render the first component media asset may comprise receiving at least one search term corresponding to the information in the edit specification. For example, one or more search term words may correspond to information that is part of or associated with (e.g., synonymous with information stored in an edit specification) an edit specification. In another embodiment, receiving the request to render the first component media asset may comprise receiving a request to render the composite media asset. For example, a request to render a composite media asset may take the form of a request to display one or more files (e.g., high-resolution or low-resolution files). As another example, a request may take the form of a signal to display the first portion of a composite media asset.

In transmitting operation 1208, information identifying the first component media asset is transmitted to the second remote computing device. In one embodiment, the information identifying the first component media asset may comprise identifying a location of the first component media asset. For example, the edit specification itself may point to an address or some other location where the first component media asset may be accessed. As another example, at least a portion of the first component media asset may be transmitted to the second remote computing device. In yet another embodiment, the information identifying the first component media asset may comprise identifying a file type of the first component media asset. For example, the information identifying the first component media asset may identify whether a certain requested file is formatted according to the MPEG-2 specification. In still yet another embodiment, transmitting information identifying the first component media asset may further comprise receiving a portion of the first component media asset and rendering the first component media asset. One skilled in the art will recognize that rendering the first component media asset may take many forms that are within the scope of this disclosure. In another embodiment, the information identifying the first media asset component may comprise a digital right that is detectable by an image rendering device. One skilled in the art will recognize that many types of digital rights exist that are within the scope of this disclosure, including but not limited to digital certificates, copy protection data and licensing information. For example, a digital right identifying the first media asset component and that is detectable by an image rendering device may include, but is not limited to, CGMS-A, HDCP, or other content protection data. In an embodiment where the information identifying the first media asset component comprises a digital right that is detectable by an image rendering device, the digital right may prevent display of the first media asset component on the image rendering device.

In receiving operation 1210, compensation is received for the transmitting step. In one embodiment, the compensation may take the form of monetary compensation. One skilled in the art will recognize that compensation may take many forms other than or in addition to monetary compensation. For example, compensation may include providing a link (e.g., a URL), a location, a file attribute, a virtual trinket (e.g., providing additional functionality to a user and/or providing a token, such as a audible, graphical or iconic “thank-you” indicator), incrementing a counter (e.g., a page count indicating a number of editing sessions by a plurality of remote computing devices), an attribution, a virtual badge (e.g., a graphical recognition), and/or an artificial currency (e.g., online credits). In another embodiment, receiving compensation may comprise receiving compensation from a user of the second remote computing device and compensating a user of the first remote computing device. For example, a user of a first remote computing device may be compensated for providing access to an asset that may be manipulated by another user of a second remote computing device. In another embodiment, receiving compensation may comprise receiving user information from one or more vendors. For example, where a vendor may be the user of the second remote computing device, information about the vendor (e.g., IP address, editing habits, etc.) may be provided as compensation. As another example, where advertising from one or more vendors is displayed to a user of the second remote computing device, the one or more vendors may provide statistical data summarizing or aggregating certain monitorable activities of one or more users. In still yet another embodiment, receiving compensation for the transmitting step may further comprise permitting access to the first component media asset at a first compensation level. For example, the first compensation level may permit free access to the first component media asset. In a related embodiment, receiving compensation for the transmitting step may further comprise permitting access to the composite media asset or the second component media asset at a second compensation level. As an example, while access to the first component media asset may be free, access to other components or composite media assets may incur a usage fee.

In another embodiment of receiving operation 1210, receiving compensation comprises compensating a user of the second remote computing device by a user of a third remote computing device, receiving compensation from the user of the second remote computing device and compensating a user of the first remote computing device. For example, a user who edits an existing edit specification may be compensated by a third party who makes further edits to or some other use of the edit specification. In this example, compensation would then be paid back to the user originating the existing edit specification.

In yet another embodiment, method 1200 may further comprise based upon the information identifying the first component media asset, identifying the second component media asset, and transmitting, to the second remote computing device, information identifying the second component media asset. For example, information from a composite media asset (e.g., defining the relationship of or identifying component media assets) may be transmitted to the second remote computing device.

FIG. 13 illustrates an embodiment of a method 1300 for inserting advertising into a composite media asset and receiving compensation therefor. In the method 1300, a composite media asset edit specification is received in receiving operation 1302 where the composite media asset edit specification identifies a first component media asset and a second component media asset. As discussed previously, composite media assets may be comprised of one or more component media assets (e.g., video files). For example, one or more video files may be combined together to create a video montage. As another example, different types of content (e.g., video, audio, images) may be combined within a standard format. For instance, images may be joined with video such that a slideshow is presented. In this example, the individual images and video clips may themselves remain as separate or distinguishable files sharing a common edit specification.

In receiving operation 1304, a request, by one or more vendors, to incorporate at least one advertisement with the composite media asset is received. As one example, by way of illustration and not of limitation, a coffee shop may request that an advertisement (e.g., as a video slate) advertising its coffee be inserted between a first and second component media asset comprising video files.

In incorporating operation 1306, the at least one advertisement is incorporated with the composite media asset. In one embodiment, advertisements may be incorporated with a composite media asset by inserting the at least one advertisement within the composite media asset. For example, advertisements may be spliced in between component media assets. As another example, advertisements may overlay (e.g., using watermarking technology, banner ads, tickers, etc.) a rendering of the composite media asset. In another embodiment, advertisements may be incorporated with a composite media asset by adding the at least one advertisement before or after the composite media asset. In receiving operation 1308, compensation from the one or more vendors for the incorporating step is received. For example, continuing with the example of the coffee shop, the vendor requesting advertising may compensate the insertion of advertising at the end of a composite media asset.

In yet another embodiment of incorporating operation 1306, incorporating the at least one advertisement with the composite media asset may comprise rendering at least one advertisement with the composite media asset. For example, advertising may be displayed, where rendering comprises displaying the composite media asset, alongside the rendered composite media asset. As another example, an advertisement may be heard on speakers separate from where a composite media asset comprising audio information is rendered.

In another embodiment of method 1300, the at least one advertisement is associated with contextual information that is derived from rendering the composite media asset edit specification. For example, where a composite media asset edit specification is accessible and rendered from an Internet page, the surrounding contextual information from the Internet page may be associated with the advertisement. In another embodiment, the at least one advertisement is associated with the composite media asset edit specification. For example, the advertisement may be associated with the composite media asset edit specification such that the advertisement is rendered alongside or in addition to rendering of the composite media asset edit specification.

FIG. 14 illustrates an embodiment, similar to the discussion of the method illustrated by FIG. 13, of another method 1400 for inserting advertising into a composite media asset and receiving compensation therefor. In a receiving operation 1402, a composite media asset edit specification, where the composite media asset edit specification identifies a first component media asset and a second component media asset, is received. In receiving operation 1404, a request, by one or more vendors, to incorporate at least one advertisement with the composite media asset is received. In incorporating operation 1406, the advertisement is incorporated with the composite media asset. In transmitting operation 1408, the composite media asset is transmitted to a remote computing device.

In receiving operation 1410, upon rendering of at least a portion of the composite media asset at the remote computing device, compensation from the one or more vendors is received. As discussed previously, and as known to those of ordinary skill in the art, rendering may take many forms. In one embodiment, receiving a composite media asset with incorporated advertising may not engender compensation by one or more vendors until some or all of the advertising is rendered (e.g., viewed on a display or output on a speaker) by a user of the remote computing device.

In another embodiment of method 1400, the at least one advertisement is rendered with the composite media asset. As discussed previously, the advertisement may be rendered alongside or in addition to the rendering of the composite media asset. In yet another embodiment, the at least one advertisement is associated with contextual information that is derived from rendering the composite media asset edit specification. In still yet another embodiment, the at least one advertisement is associated with the composite media asset edit specification.

FIG. 15 illustrates a method 1500 for reusing component media assets and receiving compensation therefor. In maintaining operation 1502, a library is maintained that includes at least a first edit specification corresponding to a first composite media asset, the first composite media asset including at least one component media asset. For example, a library may comprise a publicly-available repository for multiple edit specifications for a plurality of composite media assets. In receiving operation 1504, a request, from a remote computing device, to generate a second edit specification for a second composite media asset is received, where the second edit specification for the second composite media asset identifies the at least one component media asset and the second composite media asset including the at least one component media asset. In generating operation 1506, the second edit specification is generated. As one example, generating may further comprise rendering (e.g., as the second edit specification is generated in real-time) the second edit specification on an image rendering device. For example, a second edit specification may be generated by “remixing” or modifying a preexisting edit specification. In receiving operation 1508, compensation for the generating step is received. In one embodiment, the compensation may be in the form of monetary compensation. In another embodiment, as discussed previously, receiving compensation may comprise receiving compensation from a user of the remote computing device and compensating a user associated with the first edit specification corresponding to the first composite media asset and/or permitting access to the at least one component media asset. In yet another embodiment, as also discussed previously, receiving compensation may comprise receiving user information from one or more vendors and/or permitting access to the first composite media asset.

In a further embodiment, method 1500 may further comprise, upon generating the second edit specification, notifying a user of the first remote computing device about the existence of the second edit specification. For example, a user generating a first edit specification may be notified of child or derivative edit specifications based on his or her original (first) edit specification. In another embodiment, method 1500 may further comprise, upon generating the second edit specification, notifying a user of the of the first remote computing device of at least one difference between the first edit specification and the second edit specification. For example, a user of the first remote computing device may desire to become aware of the differences between their original or first edit specification and any subsequently generated edit specifications. In yet another embodiment, method 1500 may further comprise receiving compensation that comprises compensating a user of the second remote computing device by a user of a third remote computing device, receiving compensation from the user of the second remote computing device; and compensating a user of the first remote computing device.

Those skilled in the art will recognize that the methods and systems of the present disclosure may be implemented in many manners and as such are not to be limited by the foregoing exemplary embodiments and examples. In other words, functional elements being performed by a single or multiple components, in various combinations of hardware and software or firmware, and individual functions, can be distributed among software applications at either the client or server level or both. In this regard, any number of the features of the different embodiments described herein may be combined into single or multiple embodiments, and alternate embodiments having fewer than or more than all of the features herein described are possible. Functionality may also be, in whole or in part, distributed among multiple components, in manners now known or to become known. Thus, myriad software/hardware/firmware combinations are possible in achieving the functions, features, interfaces and preferences described herein. Moreover, the scope of the present disclosure covers conventionally known manners for carrying out the described features and functions and interfaces, and those variations and modifications that may be made to the hardware or software or firmware components described herein as would be understood by those skilled in the art now and hereafter.

While various embodiments have been described for purposes of this disclosure, various changes and modifications may be made which are well within the scope of the present invention. For example, an edit specification may further include instructions for layering multiple audio tracks together or splicing different audio samples together. As another example, the online reconfiguration of three-dimensional gaming environments (e.g., the editing of a 3D game environment) may be accomplished using the methods and systems described for generating low-resolution media assets corresponding to high-resolution media assets. As yet another example, the methods and systems described herein may permit interactive reconfiguration of Internet web pages.

Numerous other changes may be made which will readily suggest themselves to those skilled in the art and which are encompassed in the spirit of the invention disclosed and as defined in the appended claims.

Claims

1. A method for reusing component media assets and receiving compensation therefor, the method comprising:

maintaining a library including at least a first edit specification corresponding to a first composite media asset, the first composite media asset including at least one component media asset;
receiving a request, from a remote computing device, to generate a second edit specification for a second composite media asset, the second edit specification for the second composite media asset identifying the at least one component media asset and the second composite media asset including the at least one component media asset;
generating the second edit specification; and
receiving compensation for the generating step.

2. The method of claim 1 further comprising:

upon generating the second edit specification, notifying a user of the first remote computing device about the existence of the second edit specification.

3. The method of claim 1 further comprising:

upon generating the second edit specification, notifying a user of the of the first remote computing device of at least one difference between the first edit specification and the second edit specification.

4. The method of claim 1 wherein the compensation is in the form of monetary compensation.

5. The method of claim 1 wherein receiving compensation comprises:

receiving compensation from a user of the remote computing device; and
compensating a user associated with the first edit specification corresponding to the first composite media asset.

6. The method of claim 1 wherein receiving compensation comprises:

compensating a user of the second remote computing device by a user of a third remote computing device;
receiving compensation from the user of the second remote computing device; and
compensating a user of the first remote computing device.

7. The method of claim 1 wherein receiving compensation comprises:

receiving user information from one or more vendors.

8. The method of claim 1 wherein receiving compensation for the generating step further comprises:

permitting access to the first composite media asset.

9. The method of claim 1 wherein receiving compensation for the generating step further comprises:

permitting access to the at least one component media asset.

10. The method of claim 1 wherein the generating the second edit specification further comprises:

rendering the second edit specification on an image rendering device.
Patent History
Publication number: 20080215620
Type: Application
Filed: Jan 12, 2007
Publication Date: Sep 4, 2008
Applicant: Yahoo! Inc. (Sunnyvale, CA)
Inventors: Michael George Folgner (San Francisco, CA), Ryan Brice Cunningham (San Francisco, CA), Steven Waterhouse (San Francisco, CA)
Application Number: 11/622,968
Classifications
Current U.S. Class: 707/104.1; Information Processing Systems, E.g., Multimedia Systems, Etc. (epo) (707/E17.009)
International Classification: G06F 17/00 (20060101);