METHODS AND APPARATUS FOR INTEGRATING DISPARATE MEDIA FORMATS IN A NETWORKED MEDIA SYSTEM

A media system includes at least a source media device and a playback media device coupled through a network. The source media device presents media to the network. The media comprises at least one digital content file with a first format. A transcoder, also coupled to the network, converts the first file format of the digital content file to a second format. The playback device receives the digital content file, formatted in the second format, over the network, and processes the digital content file in the second format to generate processed signals. The processed signals drive the playback device to play the digital content file. In another embodiment, the transcoder operates in conjunction with one or more media servers. For this embodiment, media, stored on the media, stored on the media servers, is converted to one or more different file formats.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
BACKGROUND OF THE INVENTION

1. Field of the Invention

The present invention is directed toward the field of converging disparate types of media, and more particularly directed integrating media of different formats for playing across devices over a network.

2. Art Background

The widespread use of computers, digital cameras, and the Internet has resulted in the creation and use of digital media. Digital media has also largely replaced more traditional analog audio and video formats with the introduction and popular acceptance of audio compact discs (CDs) and digital video discs (DVDs). In general, digital media consists of various formats of data that stores audio, video, and images in binary files. These binary files are typically stored on a medium accessible to computer devices, such as CD-ROMs, hard drives, floppy disks and memory sticks.

The storage of digital media on commonly used computer medium allows for easy generation and transfer of digital media. For example, it has become popular to generate digital photos using a digital camera and then to transfer the digital photos onto computers. Computer software permits the user to manipulate the digital photos. The user may then transfer the digital photos to friends using e-mail, or post the digital photos on a web site accessible by the World Wide Web. These types of applications, which take advantage of the connectivity among different devices, have also contributed to the widespread popularity of digital media.

Digital media may be stored in a variety of formats. Special hardware or software compatible with the formats of the digital media is required to playback or view the digital media. For example, to listen to music stored in the popular MP3 format, a consumer must have a special MP3 player (i.e., either software running on a general purpose computer or a stand alone MP3 player). There are numerous formats for video, including high quality DVDs and various compression based MPEG and proprietary standards. To playback various formats of digital video, the consumer must use a device that reads the proper format of the digital media.

Because of the numerous different formats of digital media, the playback or viewing of numerous types of digital media today requires multiple types of devices. The playback of digital media stored in different formats is less problematic on a computer because the computer may play the digital media using software programs. However, a consumer may desire to play the media on other types of devices. For example, the consumer may desire to play digital audio files on a home stereo and view digital video on a television. Currently, stereos and televisions are not equipped to playback all formats of digital media. Accordingly, it is desirable to provide a media convergence platform that integrates various types of digital media into a single system.

SUMMARY OF THE INVENTION

A media system includes at least a source media device and a playback media device coupled through a network. The source media device presents media to the network. The media comprises at least one digital content file with a first format. A playback media device plays back digital content files formatted in a second format. A transcoder, also coupled to the network, converts the first format of the digital content file to a second format. In one embodiment, the transcoder comprises a device separate from the source media device and the playback media device. In another embodiment, the transcoder comprises a device integrated with the source media device. The playback device receives the digital content file, formatted in the second format, over the network, and processes the digital content file in the second format to generate processed signals. The processed signals drive the playback device to play the digital content file.

In one embodiment, the digital content file comprises a digital audio file, and the processed signals comprise audio signals. For example, the first format for the digital content file may comprise a Microsoft Windows Media Audio (“WMA”) file, a Moving Pictures Experts Group layer 3 (“MP3”) file or a Real Audio file. The second format may comprise, by way of example, a pulse code modulated (“PCM”) file or a Moving Pictures Experts Group layer 3 (“MP3”) file. In another embodiment, the digital content file comprises a digital video file, and the processed signals comprise video signals. For this embodiment, the first format may comprise, as example formats, a Moving Pictures Experts Group version 4 (“MPEG-4”) file, a Moving Joint Photographic Experts Group (“MJPEG”) file, a Quicktime file, a Microsoft Windows Media Video (“WMV”) file, or a Real Video file. The second video format may comprise a Moving Pictures Experts Group version 2 (“MPEG-2”) file.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 illustrates a media space configured in accordance with one embodiment of the present invention.

FIG. 2 illustrates one embodiment for integrating devices into a single media space.

FIG. 3 illustrates one embodiment for a media convergence platform configured in accordance with the present invention.

FIG. 4 illustrates an example media system.

FIGS. 5a and 5b illustrate one embodiment for using a transcoder device with media servers.

FIG. 6 is a flow diagram illustrating one embodiment for playback of media over a network.

FIG. 7 is a block diagram illustrating one embodiment for software components operating on a media convergence platform device.

FIG. 8 is a block diagram illustrating an example home network for the media convergence platform.

FIG. 9 is a flow diagram illustrating one embodiment for discovering devices in the media convergence system.

DETAILED DESCRIPTION Media Convergence Platform:

The media convergence platform of the present invention provides an efficient and easy way for one or more users to manage and playback media within a “media space.” As used herein, a “media space” connotes one or more media storage devices coupled to one or more media players for use by one or more users. The integration of media storage devices and media players into a single media space permits distributed management and control of content available within the media space.

FIG. 1 illustrates a media space configured in accordance with one embodiment of the present invention. As shown in FIG. 1, the media space 100 includes “n” media storage devices 110, where “n” is any integer value greater than or equal to one. The media storage devices 110 store any type of media. In one embodiment, the media storage devices 110 store digital media, such as digital audio, digital video (e.g., DVD, MPEG, etc.), and digital images. The media space 100 also includes “m” media players 120, where “m” is any integer value greater than or equal to one. In general, the media players 120 are devices suitable for playing and or viewing various types of media. For example, a media player may comprise a stereo system for playing music or a television for playing DVDs or viewing digital photos.

As shown in FIG. 1, the media storage devices 110 are coupled to the media players 120. The media storage devices 110 and the media players 120 are shown in FIG. 1 as separate devices to depict the separate functions of media storage and media playback; however, the media players may perform both the storage and playback functions. For example, a media player may comprise a DVD player that includes a hard drive for the storage and playback of digital video. In other embodiments, the storage of media and the playback/viewing of media are performed by separate devices. For this embodiment, the media players 120 playback content stored on the media storage devices 110. For example, a video clip stored on media storage device “1” may be played on any of the applicable “m” media players 120.

The storage devices 110 and media players 120 are controlled by management component 130. In general, management component 130 permits users to aggregate, organize, control (e.g., add, delete or modify), browse, and playback media available within the media space 100. The management component 130 may be implemented across multiple devices. The media space of FIG. 1 shows a plurality of users 140 to depict that more than one user may playback/view media through different media players. The system supports playback of different media through multiple media players (i.e., the system provides multiple streams of media simultaneously). The users 140, through management component 130, may also organize, control, and browse media available within the media space. The management component 130 provides a distributed means to manage and control all media within the media space. As described more fully below, the convergence media platform provides a common language to permit disparate devices to communicate (i.e., different devices utilize different network protocols).

FIG. 2 illustrates one embodiment for integrating devices into a single media space. For this embodiment, a media space 200 includes at least one media server 210 (e.g., a media space may include many media servers). The media server 210 stores media for distribution throughout the media space 200. In addition, the media server 210 stores system software to integrate the components of the media space, to distribute media through the media space, and to provide a user interface for the components of the media space. The media server 210 is coupled to different types of media players, including televisions 250 and 270, as well as an audio player 240 (e.g., stereo system). For this embodiment, the media server 210 is also coupled to a media manager 280 and to external content provider(s) 290.

For this embodiment, the media server 210 executes software to perform a variety of functions within the media space. Thus, in this configuration, the media server 210 operates as a “thick client.” A user accesses and controls the functions of the media convergence platform through a system user interface. The user interface utilizes the thick and thin clients, as well as some media players (e.g., televisions 250 & 270). In one embodiment, the user interface includes a plurality of interactive screens displayed on media player output devices to permit a user to access the functionality of the system. A screen of the user interface includes one or more items for selection by a user. The user navigates through the user interface using a remote control device (e.g., remote control 260). The user, through use of a remote control, controls the display of screens in the user interface and selects items displayed on the screens. A user interface displayed on a television permits the user, using a remote control, to perform a variety of functions pertaining to the media available in the media space.

The components of the media convergence platform are integrated through a network. For example, in the embodiment of FIG. 2, media server 210 communicates to thin audio client 230 and thin video client 220 through network 205. Network 205 may comprise any type of network, including wireless networks. For example, network 205 may comprise networks implemented in accordance with standards, such as Ethernet 10/100 on Category 5, HPNA, Home Plug, IEEE 802.11x, IEEE 1394, and USB 1.1/2.0.

For the embodiment of FIG. 2, one or more thin video clients are integrated into the media space. Specifically, a thin video client 220 is coupled to media server 210 to provide playback of digital media on television 270. The thin video client 220 does not store media. Instead, the thin video client 270 receives media from media server 210, and processes the media for display or playback on television 270 (e.g., a standard television). For example, media server 210 transmits a digital movie over network 205, and the thin video client processes the digital movie for display on television 270. In one embodiment, the thin video client 220 processes the digital movie “on the fly” to provide NTSC or PAL formatted video for playback on television 270. The thin video client 220 may be integrated into the television 270. In one embodiment, a user interface is implemented using media server 210 and thin video client 220 for display on television 270. For this embodiment, the user, using a remote control for television 270, selects items displayed on television 270 to command the system.

The media convergence platform system also optionally integrates one or more thin audio clients into the media space. For the embodiment of FIG. 2, a thin audio client 230 receives digital music (e.g., MP3 format) from media server 210 over network 205, and processes the digital music for playback on a standard audio system 240. In one embodiment, the thin audio client 210 includes a small display (e.g., liquid crystal display “LCD”) and buttons for use as a user interface. The media server 210 transmits items and identifiers for the items for display on the thin audio client 230. For example, the thin audio client 230 may display lists of tracks for playback on audio system 240. The user selects items displayed on the screen using the buttons to command the system. For example, the thin audio client screen may display a list of albums available in the media space, and the user, through use of the buttons, may command the user interface to display a list of tracks for a selected album. Then, the user may select a track displayed on the screen for playback on audio system 240.

The media manager 280 is an optional component for the media convergence platform system. In general, the media manager 280 permits the user to organize, download, and edit media in the personal computer “PC” environment. The media manager may store media for integration into the media space (i.e., store media for use by other components in the media space). In one embodiment, the media manager 280 permits the user to perform system functions on a PC that are less suitable for implementation on a television based user interface.

The media space may be extended to access media stored external to those components located in the same general physical proximity (e.g., a house). In one embodiment, the media convergence platform system integrates content from external sources into the media space. For example, as shown in FIG. 2, the media server 210 may access content external to the local network 205. The external content may include any type of media, such as digital music and video. The media convergence platform system may be coupled to external content 290 through a broadband connection (i.e., high bandwidth communications link) to permit downloading of media rich content. The external content may be delivered to the media convergence platform system through use of the Internet, or the external content may be delivered through use of private distribution networks. In other embodiments, the external content may be broadcasted. For example, the media server 210 may access external content 290 through a data casting service (i.e., data modulated and broadcast using RF, microwave, or satellite technology).

As used herein, a “device” connotes a home network client that supports a collection of services to operate a broader functionality. Also, as used herein, a “media server” is an entity on the home network that stores or presents media items to the network. Furthermore, a “node” connotes any entity on a home network, including a device and/or a media server.

The convergence media platform utilizes a “peer-to-peer” architecture. All client devices on the media platform have the ability to communicate with other devices, including multiple client devices and multiple servers. This architecture permits a device to obtain all media available on the network and to aggregate the media for presentation on that device.

A device, including a client device or a server device, may enter and/or exit the home network, at any time, and still maintain full functionality. Thus, when a device is powered off, other devices automatically recognize that the device is no longer available on the home network. When a new device is added or a portable device comes onto the network, the other nodes automatically recognize the new devices. The other nodes may utilize the services on the added device. A new media server may also automatically recognize new devices, as long as at least one other media server is currently on the network.

FIG. 3 illustrates one embodiment for a media convergence platform configured in accordance with the present invention. For this embodiment, a media network 309 couples a source device 305, a playback device 325, and a transcoder device 311. The source device 305, such as a media server, acquires media, in the form of a digital content file, formatted in “Media Format A.” In one embodiment, the media is stored in source device 305 on a permanent storage medium (e.g., hard disk drive) or on removable disk (e.g., CD or DVD). In another embodiment, the source device 305 acquires the media from an external source (e.g., Internet).

A transcoder device of the present invention operates as a translator to convert one or more input formats to one or more output formats. A transcoder may be configured to convert between any number and type of audio or video formats. Table 1 lists digital audio formats for conversion in accordance with one embodiment of a transcoder in a media convergence platform.

TABLE 1 Input Format Output Format Microsoft Windows Media Pulse Code Modulated (“PCM”) Audio (“WMA”) Microsoft Windows Media Moving Pictures Experts Group layer 3 Audio (“WMA”) (“MP3”) Moving Pictures Experts Pulse Code Modulated (“PCM”) Group layer 3 (“MP3”) Real Audio Pulse Code Modulated (“PCM”)

Table 1 lists example audio formats for conversion in one embodiment of a media convergence platform; however, a transcoder may convert between any two audio formats without deviating from the spirit or scope of the invention.

Table 2 lists digital video formats for conversion in accordance with one embodiment of a media convergence platform transcoder.

TABLE 2 Input Format Output Format Moving Pictures Experts Group Moving Pictures Experts Group layer 2 layer 4 (“MPEG-4”) (“MPEG-2”) Moving Joint Photographic Moving Pictures Experts Group layer 2 Experts Group (“MJPEG”) (“MPEG-2”) Quicktime Moving Pictures Experts Group layer 2 (“MPEG-2”) Microsoft Windows Media Moving Pictures Experts Group layer 2 Video (“WMV”) (“MPEG-2”) Real Video Moving Pictures Experts Group layer 2 (“MPEG-2”)

Table 2 identifies several example video formats that may be converted in a media convergence platform. However, a video transcoder may convert between any two video formats without deviating from the spirit or scope of the invention.

For the example of FIG. 3, transcoder device 311 receives media, in the format of “Media Format A”, from the source device 305 across network 309. The transcoder device 311 converts the media from “Media Format A” to “Media Format B.” The media file, formatted in the converted “Media Format B”, is transferred from transcoder 311 to playback device 325

As shown in FIG. 3, the playback device 325 is compatible with media formatted in “Media Format B.” As such, the playback device 325 only plays media formatted in Media Format B. The playback device 325 processes the digital content file in the second format to generate processed signals. The processed signals drive the playback device to play the digital content file. In one embodiment, the digital content file comprises a digital audio file, and the processed signals comprise audio signals. In another embodiment, the digital content file comprises a digital video file, and the processed signals comprise video signals.

In one embodiment, the media convergence platform implements the transcoder function in hardware. In other embodiments, the transcoder function is implemented in software. A software transcoder may be implemented using available open source code.

In one embodiment, the transcode function converts the input format of an audio format to a raw format, such as the pulse code modulated (“PCM”) format. For example, to conduct the WMA→PCM conversion, the transcoder decodes the Microsoft Windows Media Audio (“WMA”) format into the raw PCM format. For this embodiment, to perform the WMA→MP3 conversion, the transcoder decodes the WMA format to the raw PCM format, and then encodes the raw PCM format to the MP3 format. In other embodiments, the transcode function involves partial decoding and partial encoding (e.g., the conversion of the audio is performed fully in the frequency domain instead of converting the audio to a time domain format).

In one embodiment, video formats are also converted to “raw video formats” prior to conversion to other video formats. For example, in one embodiment, the raw video format consists of raw RGB frames. A raw RGB frame is a rasterized bitmap image of each frame in digital video represented as independent red, green and blue color channels for each pixel in each frame image. In other embodiment for generating a raw video format, the transcode function generates raw YUV 4:2:0 frames, raw YUV 4:2:2 frames, and/or raw YUV 4:4:4 frames. In general, the raw YUV 4:2:0, YUV 4:2:2 and YUV 4:4:4 frame formats represent rasterized bitmap images of each frame of digital video represented as independent Y (luminance) as well as U and V (chrominance) component color channels. Specifically, the raw YUV 4:2:0 frames format contains a single Y sample for each pixel, and one of each Cb and Cr chrominance samples spatially located in the center of each four (4) adjacent Y samples. The raw YUV 4:2:2 frames format also contains a single Y sample for each pixel, but contains one of each Cb and Cr chrominance samples spatially located in the center of each two (2) adjacent horizontal Y samples. The raw YUV 4:4:4 frames format contains a single Y sample and one of each Cb and Cr chrominance samples for each pixel.

FIG. 4 illustrates an example media system. For this example, three media servers (310, 320, and 330) are coupled to home network 340. As shown in FIG. 4, each media server stores several media items. The media items may be any type of media, including video, audio, photos, etc. For example, server 310 stores media items 1, 4, 6, 12, 22 and 33. For this example, playback device 350 is also on the home media network 340. The playback device 350 plays back media, including audio, video or photos. For example, playback device 350 may comprise a DVD player, and the media items, stored on media servers 310, 320 and 330, may comprise DVDs.

As shown in FIG. 4, media items, presented by media servers (310, 320 and 330), are associated with a format. For the example of FIG. 4, media server 330 presents media in formats A, B, D and E. For example, “Format A” may comprise a digital content file formatted as an MPEG 4 file, “Format B” may comprise a digital content file formatted as a Quicktime file, and “Format D” may comprise a digital content file formatted as a Microsoft Windows Media Video (“WMV”) file.

For the example of FIG. 4, the playback device 350 only plays media formatted in “Format A.” As such, format conversion is required to convert media of Formats B, C, D and E to Format A. The media system of FIG. 4 includes a transcoder device 360. For this embodiment, transcoder device 360 is capable of converting Formats B, C, D and E to Format A. The transcoder device 360 may convert media files “on-the-fly”, or it may convert media items for storage at playback device 350 or any other media server.

In one embodiment, after completing a discovery process, playback device 350 determines relevant media items stored on other devices (e.g., media servers) available on home network 340. Thus, playback device 350 aggregates all media, relevant to playback device 350, for use at playback device 350 (i.e., playback, control. etc.). For example, if playback device 350 plays digital video, playback device 350 aggregates a list of all digital video available on home media network 340. The aggregated list includes all media items that may be converted to Format A. As described more fully below, playback device 350 may invoke a service on transcoder device 360 to determine media formats that may be converted to Format A. As shown in FIG. 4, playback device 350 aggregates all media items stored on media servers 310, 320 and 330.

FIGS. 5a and 5b illustrate one embodiment for using a transcoder device with media servers. For the embodiment shown in FIG. 5a, media server 511 and media server 509 are coupled to transcoder device 507 through network 503. Each media server stores a number of media items, such as audio, video or photos. Each media item has an associated format (e.g., WMA, MP3, MPEG-2, MPEG-4, etc.). For the example shown in FIG. 5a, media server 509 stores media items Media_1_Format_B, Media_2_Format_E, Media_3_Format_C, Media_4—Format_C . . . Media_n_Format_X, wherein “n” is any integer value and X is any format type. Media server 511 stores, by way of example, Media_20_Format_D, Media_21_Format_F, Media_22_Format_B, Media_23_Format_C . . . Media_m_Format_y, wherein “m” is any integer value and y is any format type.

In one embodiment, an automated process is commenced to convert all media items stored on a media server to one or more formats. The process may run without initiation from a user and when system resources are available. Media items are transferred from a media server (509 or 511) to transcoder device 507. Transcoder 507 determines the current format of the media item, converts the media items to one or more different formats, and transfers the converted media items to the media server. The media server then stores the converted media items along with the media items in the original format. FIG. 5b illustrates an example for converting the media items of FIG. 5a to a specified format, “Format_A.” As shown in FIG. 5b, after the conversion process, media server 511 stores, in addition to the original media, all media items converted to Format_A (i.e., Media_20_Format_A, Media_21_Format_A, Media_22_Format_A, Media_23_Format_A and Media_m_Format_A). Similarly, all media items stored in media server 509 are converted to Format_A (i.e., Media_1_Format_A, Media_2_Format_A, Media_3_Format_A, Media_4_Format_A and Media_n_Format_A).

In one embodiment a program allows the user to configure the parameters for the conversion process. For example, a user may specify one or more formats for conversion based on the media type of the original media item. For example, if the original media item is a WMA file, then the user may specify conversion to MP3. The user may also specify the frequency to conduct the conversion. For example, the user may specify that the system update media server items once everyday (i.e., traverse media server items and convert any media items not already converted to one or more specified formats). Also, the user may specify one or more media servers to perform the conversion process.

The underlying protocols of the media system do not permit a client device to aggregate media items from devices on the home network. The protocols themselves have no requirement to support a distributed system. For this embodiment of the media convergence platform, aggregation logic creates a distributed system using non-distributed protocols. The aggregation logic uses multiple protocols to integrate devices on the home network.

FIG. 6 is a flow diagram illustrating one embodiment for playback of media over a network. Initially, a playback device discovers media servers and transcoder devices on the home network (block 410, FIG. 6). One embodiment for discovering devices on the home network is described more fully below. Based on information learned in the discovery process, the playback device constructs state information for the discovered media servers and transcoder devices (block 420, FIG. 6). The playback device aggregates a list of media items available through the network (block 425, FIG. 6). The aggregation logic for the playback device acquires media items from all media servers that contain compatible media items. For example, if the playback device plays music items, the playback device acquires all music items from all media servers available on the network.

The playback device receives a request to play media item(s) (block 430, FIG. 6). In response, the playback device connects to a selected media server and transcoder device using state information acquired during the discovery process (block 450, FIG. 6). Specifically, the playback device translates the request to play a media item to a protocol supported by the media server with the selected item, and forwards the request to the media server. Similarly, the playback device translates the request to transcode a media item to a protocol supported by the transcoder, and forwards the request to the transcoder. In one embodiment, the playback device invokes a service on the selected media server to obtain the media items and invokes a service on the transcoder to convert the file format. (block 460, FIG. 6). In one embodiment, the playback device invokes a content manager service on the media server to acquire the media item.

FIG. 7 is a block diagram illustrating one embodiment for software components operating on a media convergence platform device. Software components 500 include, at the highest level, application software 502. The application software 502 implements functionality for an underlying device. For example, application software 502 may implement functions for a DVD player. As shown in FIG. 7, underneath the application software 502 is aggregation logic 520. Aggregation logic 520 permits a client device to aggregate media items on the home network.

The software components 500 also include user interface (“UI”) rendering logic 510. UI rendering component 510 translates scene information to display information suitable for display on the client device. The UI rendering component 510 also renders the display data. For example, if the underlying client device includes a television display (e.g., CRT), then UI rendering engine 510 generates graphics data from scene information, and renders the graphics data on the television display. If the display on the client device is a LCD display, then UI rendering engine 510 generates lists from scene information, and displays the lists on the LCD display.

As shown in FIG. 7, the client device incorporates one or more network protocols and remote procedure calls (“RPC”) mechanisms. For example, FIG. 7 shows that the client device supports network protocolA (525), network protocolB (530), and network protocoln (535). For this example, client device software 500 supports RPC mechanismA (540), RPC mechanismB (545), and RPC mechanismn (550).

The client device software 500 supports one or more services. As shown in FIG. 7, one or more methods of a service are accessible through an interface. In general, the methods, when invoked, provide specific functionality for the underlying service. For this example, client device software 500 includes serviceA (555), serviceB (560), and servicen (565). Each service is associated with one or more methods (i.e., method(1)-method(n)).

In one embodiment, the media convergence platform supports a plurality of underlying protocols. In general, the protocols define commands, RPC mechanisms, and interfaces to services. In one embodiment, the media convergence platform supports an industry defined UPnP protocol. In general, the UPnP protocol defines discovery over IP networks, an RPC mechanism, and interfaces for activating services. UPnP services include: a content directory service, a connection manager service, an audio/video (“A/V”) transport service and an A/V control service.

In one embodiment, the media convergence platform also supports a proprietary protocol (i.e., non-industry standard protocol). For this embodiment, the proprietary protocol defines a network discovery process, an RPC mechanism, and an interface to services. The services include a content manager and a media player service. The content manager service allows a client device to interface to a database. Specifically, using the content manager service, the client device may extract information (e.g., URL to identify media, metadata, etc.) from a database on another network device. Thus, the content manager service provides a means for a device of the media convergence platform system to query a database. The media player service defines an interface to permit playback functionality (e.g., initiate and control media streams).

In one embodiment, the discovery process on the proprietary protocol implements asynchronous based messaging. The discovery protocol operates on any network that supports packet based messaging or on a serialized network. In one embodiment, the discovery protocol includes an “announce” command, a “discovery” command, and a “bye-bye” command. The announce command is used by a device to announce its presence on the home media network. A discovery command is a request for an announcement (i.e., queries whether any client devices are on the home network). The “bye-bye” command is used by a client device to announce that the client device is leaving the network. In one embodiment, there are two types of announcements and two types of “bye-bye” commands: one for devices and one for services.

In one embodiment, the RPC mechanism, supported by the proprietary protocol, uses a packet based protocol. The services include methods and an identification number to permit a device on the home network to construct RPC based packets with the appropriate arguments. In general, an RPC mechanism permits a device to control another device on the network. The protocol is effectuated through requests and responses. The RPC packets include a header. In one embodiment, the header contains: version information, a command class (maps to a particular service), the command (the method the device is requesting or the response coming from the method), an identification (identification of requests or identification of responses corresponding to a request), and a length. After the header, the RPC protocol format specifies data (i.e., arguments for requests and returns values for responses).

FIG. 8 is a block diagram illustrating an example home network for the media convergence platform. For this example, a home network includes transcoder 650, DVD player 640, and media server 630. A client device (610) enters the home network 620, and discovers, using a supporting protocol, three devices (e.g., transcoder 650, DVD player 640, and media server 630). As shown in FIG. 8, client device 610 stores state information for each of the devices discovered on home network 620. Specifically, client device 610 stores, for device1 (media server 630) a supporting network protocol (i.e., network protocolA) as well as a list of services supported by the devices (i.e., serviceA and serviceC). The network protocol also specifies an RPC mechanism to execute remote procedure calls on media server 630. Similarly, state information for device2 (i.e., DVD player 640) indicates that device2 supports network protocolA and implements serviceA and serviceB. Device3, transcoder 650, supports network protocolC, and implements serviceA and serviceB. For this example, serviceA supports the conversion between FormatA to FormatB, and serviceB supports the conversion from FormatC to FormatD.

As shown in FIG. 8, each service (e.g., serviceA, serviceB, and serviceC) supports an interface. The interface defines a specification to provide a means to access the methods or commands within a service. As such, the client device 610 utilizes services (serviceA or serviceB) on transcoder 650 through their respective interfaces. Note that media server 630, DVD player 640 and transcoder 650 all implement serviceA. Each interface for serviceA is the same to permit uniform accessibility to the service. However, the implementation of serviceA in each of the devices may be different.

In one embodiment, a media convergence platform implementation provides security. For this embodiment, the announcement command is open ended, such that the protocol only defines a minimum specification for communication. Thus, announcement protocols may support multiple network specifications, including TCP and secure sockets layer (“SSL”). The protocol supports implementation on TCP/IP networks. In addition, the protocol supports SSL operating on TCP/IP networks. SSL permits secure communications, including authentication, between two parties on a network.

The proprietary protocol also permits an implementation using partial security. For this embodiment, a service may include some methods that require secure communications and other methods that do not require secure communications. Thus, some methods utilize SSL technology to realize secure communications between two devices on the home network.

FIG. 9 is a flow diagram illustrating one embodiment for discovering devices in the media convergence system. A new device (i.e., a device not currently connected to the network) is connected to the home media network (710, FIG. 9). In order to communicate on the network, the new device obtains a network address (block 720, FIG. 9). For example, in an IP network, the client requires an IP address. If the underlying network supports TCP/IP, then the client device determines if there is a DHP server. If so, the DHP server assigns the IP address to the new device. If no DHP server is available to assign the new device an IP address, then the new device selects an IP address from a pool and determines whether any other device on the home network has that IP address. If no other device on the network has that IP address, then the client device uses this IP address. This process of auto IP addressing allows communication on a home network within a single subnet.

The new device transmits an “announcement” command over the network (block 730, FIG. 9). The format of the announcement command complies with a protocol supported by the devices. The new device may broadcast or multicast the announcement command over the network. For example, in an IP network, the new device may multicast the announcement in order to expand subnets if the home network includes a gateway. A multicast format is used to specify specific IP addresses (e.g., transmitting an announcement only to those devices on the network interested or compatible with the new device). As used herein, compatible devices are those devices that may be interested in communicating with the client device.

In response to the new device's announcement command, the new device constructs state information. In general, the state information provides details regarding devices available on the network. The state information includes protocols and services supported by those devices. When compatible devices on the network receive the announcement command, those compatible devices may add information, encapsulated in the announcement command, to a local cache.

If there are no compatible devices on the network or the new device does not desire to utilize a service on the network, then the process terminates. For example, if the new device is an MP3 player, then compatible devices include those media servers storing MP3 audio as well as other MP3 players. If there are other compatible devices on the network, those devices expose one or more services to the new device (block 750, FIG. 9). To discovery services on the network, the new device transmits a discovery command and waits for a response. For this example, a media server, which stores MP3 audio, exposes an interface to allow the new device to aggregate the MP3 audio stored on that media server. Similarly, a compatible MP3 player exposes a service, through a pre-defined interface, to permit the new device to play MP3 audio stored on the new device at the compatible MP3 player.

In response to the request (e.g., new device application logic), the new device connects to a compatible device via a supporting protocol (block 760, FIG. 9). Specifically, the device translates the protocol for the appropriate device using the state information. For example, if the compatible device supports an industry standard protocol, then the new device selects the industry standard protocol to communicate to that device. The new device utilizes the services on the compatible device (block 770, FIG. 9).

A media server entering a home network is one example of the discovery process. For this example, the media server, after obtaining a network address, transmits an announcement command over the network. The media server announces the services it supports (e.g., content manager, media player service), and exposes interfaces to network clients to permit access to those services. If a device enters the network, the device waits for an announcement from the server. When the client identifies the media server, the client connects to the media server via a protocol the server specified in the announcement command. This process allows the client device to navigate media on the media server. Using the supporting protocol, the client device connects to a playback device, either itself or another playback device, and instructs the playback device to play the item that a user selected from those media items available on the media server.

Although the present invention has been described in terms of specific exemplary embodiments, it will be appreciated that various modifications and alterations might be made by those skilled in the art without departing from the spirit and scope of the invention.

Claims

1. A method for integrating disparate media types in a networked media system, said method comprising:

presenting media to a network from a source media device, said media comprising at least one digital content file comprising a first format;
transmitting, over said network, said digital content file in said first format to a transcoder media device;
converting said first format of said media to a digital content file comprising a second format in said transcoder media device;
transmitting, over said network, said digital content file for said media in said second format from said transcoder media device to a playback device;
processing said digital content file in said second format at said playback media device, to generate processed signals; and
presenting said processed signals at said playback media device to play said digital content file.

2-36. (canceled)

Patent History
Publication number: 20120131218
Type: Application
Filed: Nov 16, 2011
Publication Date: May 24, 2012
Applicant: ROVI SOLUTIONS CORPORATION (Santa Clara, CA)
Inventors: Daniel Putterman (San Francisco, CA), Brad Dietrich (San Francisco, CA)
Application Number: 13/297,520
Classifications
Current U.S. Class: Computer-to-computer Data Streaming (709/231)
International Classification: G06F 15/16 (20060101);