Sharing gameplay in cloud gaming environments
A method for enabling multiplayer gameplay includes: executing, by a cloud gaming server, a first user's gameplay of a video game, wherein executing the first user's gameplay includes processing input received over a network from a first controller device that is associated with the first user to direct gameplay events occurring in the first user's gameplay; presenting, over the network, a live video feed of the first user's gameplay to a remote second user; processing a request for the second user to join the first user's gameplay; initiating gameplay by the second user in the first user's gameplay, wherein initiating gameplay by the second user includes processing input received over the network from a second controller device that is associated with the second user to direct at least some of the gameplay events occurring in the first user's gameplay.
Latest Sony Interactive Entertainment LLC Patents:
This application claims priority as a continuation of U.S. application Ser. No. 15/859,632, filed Dec. 31, 2017, entitled “Sharing Recorded Gameplay to a Social Graph,” which claims priority as a continuation of U.S. application Ser. No. 15/133,066, filed Apr. 19, 2016, entitled “Sharing Recorded Gameplay to a Social Graph,” which claims priority as a continuation of U.S. application Ser. No. 13/839,382, filed Mar. 15, 2013, entitled “Sharing Recorded Gameplay to a Social Graph.”
U.S. application Ser. No. 13/839,382 claims priority to U.S. Provisional Application No. 61/745,281, filed Dec. 21, 2012, entitled “Automatic Generation of Suggested Mini-Games for Cloud-Gaming Based on Recorded Gameplay.”
U.S. application Ser. No. 13/839,382 claims priority as a continuation-in-part of U.S. application Ser. No. 13/418,691, filed Mar. 13, 2012 (now issued as U.S. Pat. No. 8,672,765), entitled “System and Method for Capturing and Sharing Console Gaming Data.”
U.S. application Ser. No. 13/839,382 claims priority as a continuation-in-part of U.S. application Ser. No. 13/418,703, filed Mar. 13, 2012, entitled “System and Method for Capturing and Sharing Console Gaming Data.”
U.S. application Ser. No. 13/839,382 claims priority as a continuation-in-part of U.S. application Ser. No. 13/418,720, filed Mar. 13, 2012, entitled “System and Method for Capturing and Sharing Console Gaming Data.”
The disclosures of these applications are incorporated by reference herein in their entirety for all purposes.
Related ApplicationsThis application is related to U.S. application Ser. No. 13/779,730, filed Feb. 27, 2013, entitled “Systems and Methods for Tagging Content of Shared Cloud Executed Mini-games and Tag Sharing Controls.”
This application is related to U.S. application Ser. No. 13/779,735, filed Feb. 27, 2013, entitled “Systems and Methods for Ranking of Cloud Executed Mini-games Based on Tag Content and Social Network Content.”
This application is related to U.S. application Ser. No. 13/779,737, filed Feb. 27, 2013, entitled “Systems and Methods for Enabling Shadow Play for Video Games Based on Prior User Plays.”
This application is related to U.S. Provisional Application No. 61/746,478, filed Dec. 27, 2012, entitled “Systems and Methods for Sharing Cloud-Executed Mini-Games, Challenging Friends and Enabling Crowd Source Rating.”
This application is related to U.S. Provisional Application No. 61/745,543, filed Dec. 21, 2012, entitled “Cloud-Based Game Slice Generation and Frictionless Social Sharing with Instant Play.”
The disclosures of these applications are incorporated by reference herein in their entirety for all purposes.
BACKGROUND 1. Field of the InventionThe present invention relates to methods and systems for sharing recorded gameplay to a social graph.
2. Description of the Related ArtThe video game industry has seen many changes over the years. As computing power has expanded, developers of video games have likewise created game software that takes advantage of these increases in computing power. To this end, video game developers have been coding games that incorporate sophisticated operations and mathematics to produce a very realistic game experience.
Example gaming platforms, may be the Sony Playstation®, Sony Playstation2® (PS2), and Sony Playstation3® (PS3), each of which is sold in the form of a game console. As is well known, the game console is designed to connect to a monitor (usually a television) and enable user interaction through handheld controllers. The game console is designed with specialized processing hardware, including a CPU, a graphics synthesizer for processing intensive graphics operations, a vector unit for performing geometry transformations, and other glue hardware, firmware, and software. The game console is further designed with an optical disc tray for receiving game compact discs for local play through the game console. Online gaming is also possible, where a user can interactively play against or with other users over the Internet. As game complexity continues to intrigue players, game and hardware manufacturers have continued to innovate to enable additional interactivity and computer programs.
A growing trend in the computer gaming industry is to develop games that increase the interaction between user and the gaming system. One way of accomplishing a richer interactive experience is to use wireless game controllers whose movement is tracked by the gaming system in order to track the player's movements and use these movements as inputs for the game. Generally speaking, gesture input refers to having an electronic device such as a computing system, video game console, smart appliance, etc., react to some gesture made by the player and captured by the electronic device.
Another growing trend in the industry involves the development of cloud-based gaming systems. Such systems may include a remote processing server that executes a game application, and communicates with a local thin client that can be configured to receive input from users and render video on a display.
It is in this context that embodiments of the invention arise.
SUMMARYEmbodiments of the present invention provide methods and systems for sharing recorded gameplay to a social graph. It should be appreciated that the present invention can be implemented in numerous ways, such as a process, an apparatus, a system, a device or a method on a computer readable medium. Several inventive embodiments of the present invention are described below.
In one embodiment, a method for sharing recorded gameplay to a social graph is provided, including the following method operations: recording video of a user's gameplay during an active state of a gameplay session; receiving a command to initiate a sharing operation during the active state of the gameplay session; in response to receiving the command, entering a paused state of the gameplay session and presenting a sharing interface; processing input received via the sharing interface to determine a user-defined selection of the recorded video; sharing the user-defined selection of the recorded video to a social graph of the user; resuming the active state of the gameplay session; wherein the method is executed by a processor.
In one embodiment, recording video of the user's gameplay includes storing the video to a buffer during the active state of the gameplay session.
In one embodiment, presenting the sharing interface includes retrieving and presenting one or more portions of the video from the buffer.
In one embodiment, the sharing interface includes a duration selector for defining a duration of the user-defined selection.
In one embodiment, sharing the selection includes uploading the selection for availability on a social network service that defines the social graph of the user.
In one embodiment, uploading the selection includes compressing the selection and uploading the compressed selection.
In one embodiment, sharing the selection includes generating a notification to a member of the social graph of the user.
In one embodiment, the notification is defined by one or more of a private message, a chat message, an in-game notification, a post to a social news feed, an e-mail.
In one embodiment, sharing the selection includes presenting the selection on a profile page of the user.
In one embodiment, receiving the command to initiate the sharing operation is defined from a button press on a controller device.
In another embodiment, a method for sharing recorded gameplay to a social graph is provided, including the following method operations: recording video of a user's gameplay during an active state of a gameplay session; receiving a command to initiate a sharing operation during the active state of the gameplay session; in response to receiving the command, determining a user-defined selection of the recorded video; sharing the user-defined selection of the recorded video to a social graph of the user; wherein the method is executed by a processor.
In one embodiment, recording video of the user's gameplay includes storing the video to a buffer during the active state of the gameplay session.
In one embodiment, determining the user-defined selection of the recorded video includes retrieving and presenting the video from the buffer, and processing user input identifying a portion of the video.
In one embodiment, sharing the selection includes uploading the selection for availability on a social network service that defines the social graph of the user.
In one embodiment, receiving the command to initiate the sharing operation is defined from a button press on a controller device.
In another embodiment, a non-transitory computer readable medium having program instructions defined thereon for sharing recorded gameplay to a social graph is provided. The program instructions include: program instructions for recording video of a user's gameplay during an active state of a gameplay session; program instructions for receiving a command to initiate a sharing operation during the active state of the gameplay session; program instructions for, in response to receiving the command, determining a user-defined selection of the recorded video; program instructions for sharing the user-defined selection of the recorded video to a social graph of the user.
In one embodiment, recording video of the user's gameplay includes storing the video to a buffer during the active state of the gameplay session.
In one embodiment, determining the user-defined selection of the recorded video includes retrieving and presenting the video from the buffer, and processing user input identifying a portion of the video.
In one embodiment, sharing the selection includes uploading the selection for availability on a social network service that defines the social graph of the user.
In one embodiment, receiving the command to initiate the sharing operation is defined from a button press on a controller device.
Other aspects of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.
The invention may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
The following embodiments describe methods and apparatus for sharing recorded gameplay to a social graph.
It will be obvious, however, to one skilled in the art, that the present invention may be practiced without some or all of these specific details. In other instances, well known process operations have not been described in detail in order not to unnecessarily obscure the present invention.
In various embodiments, the degree of processing performed by the client may vary with respect to input and output processing. However, broadly speaking, the video game state is substantially maintained and executed on the cloud gaming servers 104, with the client primarily functioning to receive and communicate user inputs, and receive video/audio data for rendering. The client 101 may be a standalone device that is connected to the display 100 and provides video data for rendering on the display 100. In other embodiments, the client can be integrated into the display 100. In one embodiment, the display 100 is a networked display providing a platform operating system for applications or “apps” utilizing the network connectivity of the display. In such an embodiment, the client can be defined by an application executed on the platform provided by the display's operating system.
At each of the first, second, third, and fourth locations, at least one computing device is provided for processing input from the various users and rendering a cloud-based video game on their respective displays. It should be appreciated that the computing device can be integrated into a display, or may be a standalone device such as a personal computer, set top box, gaming console, or any other type of device having at least one processor and memory for processing and storing data. The computing device can execute or define a client, as has been described above. The computing devices are networked, and communicate over a network, such as the Internet 102, with cloud gaming servers 104.
The cloud gaming servers 104 execute the various video games which are being played by the users, defining a given video game's game state from moment to moment, and sending video data (including image data and audio data) to a computing device at a particular location. The computing device at a given location processes input from the user(s) playing the video game, and transmits input data to the cloud gaming server, which in turn processes the input data to affect the game state of the video game. It should be appreciated that cloud-based gaming facilitates multi-player gaming from players located at different locations by providing for execution of the video game at a remote server that is accessible by all players over a network. In this manner, execution of the video game is not dependent on any single player's hardware or network conductivity, though such will affect the user experience for that given player.
The cloud gaming service 210 includes resources for providing an environment in which a video game can be executed. Broadly speaking, resources can include various kinds of computer server hardware, including processors, storage devices, and networking equipment, which can be utilized to facilitate execution of a video game application. In the illustrated embodiment, a video game library 212 includes various game titles. Each game title defines executable code as well as associated data and asset libraries which are utilized to instantiate a video game. The host 214 can be a single computing device that defines a platform for instantiating virtual machines 216. In another embodiment, the host 214 can itself be a virtualized resource platform. In other words, the host 214 may operate over one or more server computing devices, handling the allocation and usage of the resources defined by the server computing devices, while presenting a unified platform upon which virtual machines 216 can be instantiated.
Each virtual machine 216 defines a resource environment which can support an operating system, upon which a video game application 218 can be run. In one embodiment, a virtual machine can be configured to emulate the hardware resource environment of a gaming console, with an operating system associated with the gaming console being run on the virtual machine to support the running of game titles which were developed for that gaming console. In another embodiment, the operating system can be configured to emulate a native operating system environment of a gaming console, though the underlying virtual machine may or may not be configured to emulate the hardware of the gaming console. In another embodiment, an emulator application is run on top of the operating system of a virtual machine, the emulator being configured to emulate the native operating system environment of a gaming console so as to support video games designed for that gaming console. It should be appreciated that a variety of current and legacy gaming consoles can be emulated in a cloud-based gaming system. In this manner, a user can access game titles from different gaming consoles via the cloud-gaming system.
When the user 200 requests to play a specific video game title, the video game title is retrieved from the library 212. If a compatible virtual machine has not been already instantiated or is not available for use, then a new compatible virtual machine is instantiated on the host 214. The retrieved video game title is then executed as an application 218 on the available or newly instantiated virtual machine 216. In one embodiment, this can entail determining the appropriate platform for the video game title (e.g. which gaming console or operating system the game requires to run) and assigning the video game title to an appropriate virtual machine for execution, e.g. one having an emulator application capable of handling execution of the video game title. The executing video game communicates with the game client 204 to provide an interactive gaming experience for the user 200. More specifically, the executing videogame application 218 receives input data from the client 204 over the network 208. The application 218 processes the input data to update the game state of the executing application. As the game state changes, the application 218 outputs video data that is sent to the client 204 for rendering on the display 206. Additionally, the application 218 may also output feedback data to the client 204 that is utilized to provide an additional feedback mechanism to the user. By way of example, the user's controller 202 may include a tactile vibration feedback mechanism that can be activated based on the output feedback data from the video game application.
In one embodiment, the cloud gaming system is configured to detect the type of client device associated with the user, and also a type of controller available for the user to provide input to the cloud-based video game. For example, in one embodiment, when a user logs in to the cloud gaming system, they may be presented with an option to designate the type of client device with which they are accessing the cloud gaming system. In one embodiment, a series of client device options are presented from which the user may select one corresponding to their client device. The user may also be presented with an option to designate the type of controller device they will use to play a video game. In one embodiment, a series of controller options can be presented to the user, from which the user may select to designate a controller type corresponding to their controller hardware. In other embodiments, the cloud gaming system can be configured to automatically detect the client device type and/or the controller device type.
For example, at the time of login, the client device may send information to the cloud gaming server identifying itself as well as a connected controller device (e.g. in response to a request from the cloud gaming server). Based on this information, the cloud gaming server may determine an appropriate video game output configuration and input parameter configuration to provide a gaming experience optimized for the user's client device and controller device. In one embodiment, a look-up table is employed to determine video game configuration and input parameter configuration based on a detected client device and a detected controller device.
It should be appreciated that a given video game may be developed for a specific platform and a specific associated controller device. However, when such a game is made available via a cloud gaming system as presented herein, the user may be accessing the video game with a different controller device. For example, a game might have been developed for a game console and its associated controller, whereas the user might be accessing a cloud-based version of the game from a personal computer utilizing a keyboard and mouse. In such a scenario, the input parameter configuration can define a mapping from inputs which can be generated by the user's available controller device (in this case, a keyboard and mouse) to inputs which are acceptable for the execution of the video game.
In another example, a user may access the cloud gaming system via a tablet computing device, a touchscreen smartphone, or other touchscreen driven device. In this case, the client device and the controller device are integrated together in the same device, with inputs being provided by way of detected touchscreen inputs/gestures. For such a device, the input parameter configuration may define particular touchscreen inputs corresponding to game inputs for the video game. For example, buttons, a directional pad, or other types of input elements might be displayed or overlayed during running of the video game to indicate locations on the touchscreen that the user can touch to generate a game input. Gestures such as swipes in particular directions or specific touch motions may also be detected as game inputs. In one embodiment, a tutorial can be provided to the user indicating how to provide input via the touchscreen for gameplay, e.g. prior to beginning gameplay of the video game, so as to acclimate the user to the operation of the controls on the touchscreen.
In some embodiments, the client device serves as the connection point for a controller device. That is, the controller device communicates via a wireless or wired connection with the client device to transmit inputs from the controller device to the client device. The client device may in turn process these inputs and then transmit input data to the cloud gaming server via a network (e.g. accessed via a local networking device such as a router). However, in other embodiments, the controller can itself be a networked device, with the ability to communicate inputs directly via the network to the cloud gaming server, without being required to communicate such inputs through the client device first. For example, the controller might connect to a local networking device (such as the aforementioned router) to send to and receive data from the cloud gaming server. Thus, while the client device may still be required to receive video output from the cloud-based video game and render it on a local display, input latency can be reduced by allowing the controller to send inputs directly over the network to the cloud gaming server, bypassing the client device.
In one embodiment, a networked controller and client device can be configured to send certain types of inputs directly from the controller to the cloud gaming server, and other types of inputs via the client device. For example, inputs whose detection does not depend on any additional hardware or processing apart from the controller itself can be sent directly from the controller to the cloud gaming server via the network, bypassing the client device. Such inputs may include button inputs, joystick inputs, embedded motion detection inputs (e.g. accelerometer, magnetometer, gyroscope), etc. However, inputs that utilize additional hardware or require processing by the client device can be sent by the client device to the cloud gaming server. These might include captured video or audio from the game environment that may be processed by the client device before sending to the cloud gaming server. Additionally, inputs from motion detection hardware of the controller might be processed by the client device in conjunction with captured video to detect the position and motion of the controller, which would subsequently be communicated by the client device to the cloud gaming server. It should be appreciated that the controller device in accordance with various embodiments may also receive data (e.g. feedback data) from the client device or directly from the cloud gaming server.
In some embodiments the selection markers may not be continuously adjustable along the gameplay timeline, but may be configured instead to snap to predefined time points along the gameplay timeline. For example, predefined time points may be defined to correspond with specific events occurring in the gameplay timeline. The specific events of a given gameplay timeline for which predefined time points will be assigned can be generated based on analysis of the users gameplay, and will depend on the specific architecture of the videogame gameplay. In one embodiment, predefined time points can be assigned based on geographical location of a character within a virtual world defined by the videogame. For example, predefined time points can be assigned to the specific times at which a character moved from one geographical locale to another geographical locale, e.g. movement from one scene location to another scene location, movement from one city to another city, entering a structure, entering a room within a structure, entering the vehicle, entering a different type of environment, or any other kind of geographical transition of significance. In another embodiment, predefined time points can be assigned based on development of a user's character or entity which is controlled in the videogame. For example, predefined time points can be assigned when a character or entity controlled by the user accomplishes a task, acquires a skill, acquires an object, passes a level or otherwise completes a portion of the videogame, or performs or achieves any other kind of significant activity in the video game.
After a user has selected a portion of gameplay from which to create a minigame, embodiments of the present invention provide for systems and methods to create a playable minigame based on the selected portion of gameplay. More specifically, the minigame enables another user to play substantially the same portion of the video game that the original user played, and possibly under substantially the same conditions and parameters. In this sense, the minigame is more than simply a replay video of the original user's gameplay (though a replay video clip of the original user's gameplay may be presented in conjunction with the minigame), but is a playable portion of the video game itself that has been designated based on user selection of their own gameplay. A secondary user may thus experience a substantially similar gameplay experience to that of the original user.
Though in the currently described embodiment, an option to generate a game slice is presented to a user following completion of a level of the videogame, it should be appreciated that in other embodiments the user can generate a game slice at any other time during or outside of gameplay of the videogame, provided that recorded gameplay of the videogame exists from which a selection may be defined by the user for a game slice. For example, in one embodiment an interface can be presented which provides access to various recorded gameplay from various video games that are associated with the user. The user may select gameplay of a specific videogame and generate a game slice by selecting a portion of the recorded gameplay from which to generate the game slice in accordance with embodiments described herein.
As has been described, a user interface can graphically depict the user's recorded gameplay to facilitate selection by the user of a portion of the user's gameplay from which to generate a game slice. In the illustrated embodiment, the user has defined a selection 610 from their recorded gameplay. This selection of the user's recorded gameplay is utilized by a game slice generator to generate game slice code 622 which defines a limited game based on the selected portion of the user's gameplay. The game slice generator 612 includes a game state analyzer 614 which analyzes the game state of the recorded selection 610. Based on the analysis of the game state of the recorded selection, a game breakpoint processor determines appropriate breakpoints to define the beginning and ending of the game slice. Breakpoints can be defined based on geography, time, task or goal accomplishment, scene boundaries (physical or temporal), or any other aspect of a video game according to which the video game's gameplay can be segmented to generate a game slice. A brief description of some illustrative embodiments will serve to highlight certain possibilities for breakpoint determination.
For example, some video games entail control of a character that can be moved from one geographic scene or locale to another scene or locale. The selected portion of the user's gameplay may be determined to be generated from gameplay at a particular scene. In such an embodiment, the boundaries of the particular scene can define the geographic breakpoint for the game slice, selecting the scene to the exclusion of other scenes, which can entail exclusion of other adjacent or adjoining scenes, as well as scenes which are non-adjacent or non-adjoining or otherwise less related or unrelated to the particular scene. It should be appreciated that the recorded gameplay selection 610 may entail gameplay from multiple scenes, in which case, the game breakpoint processor 616 may be configured to define breakpoints according to the boundaries of the multiple scenes which are utilized for the recorded gameplay selection.
It should be noted that a scene can be both geographic and temporal in nature. That is, the scene may not only define a geographic region within a virtual space defined by the video game, but may also be configured to exist during a certain time or at a particular chronological point within the larger context of the video game. Such a scene may have defined objectives or goals that are to be accomplished by the player. Thus, game breakpoints can be defined based on chronology or other temporal aspects as defined by the video game.
Furthermore, a given scene may have associated objects or features which are presented as part of the scene during gameplay. These objects or features may be analyzed to define further breakpoints according to their inclusion. For example, the objects in the scene may be taken from a subset of an asset library, in which case the subset of the asset library can be defined for the game slice by the game breakpoint processor 616, to the exclusion of other objects in the asset library which are not utilized in the scene(s) associated with recorded gameplay selection. It should be understood that objects and features can be dynamic elements of a given scene, with associated mechanisms defining their change in response to events occurring in the video game. For example, an object might have a damage modeling module that determines and adjusts the appearance of the object when it is damaged (e.g. when struck by a weapon). Or a feature could be a vehicle that is made available during the scene, with the vehicle having associated logic which defines its appearance during gameplay as well as its operation and response to user input. Such logic or damage modeling can further define game breakpoints for generation of the game slice.
Various aspects of a video game which define or are otherwise utilized for a selected portion of a video game can be the basis for defining a game breakpoint. The presently described examples are provided by way of example only and not by way of limitation. It should be appreciated that in other embodiments, other aspects of a video game can form the basis for defining breakpoints to generate a game slice.
In one embodiment, a video game may be organized into various scenes which must be completed in a linear fashion, such that a later scene cannot be attempted until its preceding scene has first been completed. Each scene can include a number of objectives or goals, some of which may be required for completion of the scene, and some of which may be optional for completion of the scene. Objectives can include navigating from a starting location to a predefined ending location within the scene, surviving for a predefined length of time, destroying a predefined number of enemies, acquiring a certain number of points, defeating a particular enemy, or any other activity which can define an objective within the game. A scene may have various predefined completion points, that is, points wherein the user, once having achieved the completion point, is able to return to that point if the user becomes unable to continue gameplay for some reason (e.g. user quits the game, the user's game character dies or runs out of lives or health, user's vehicle crashes, etc.). At predefined completion points, a video game may be configured to automatically save the user's progress, or present an option for the user to save their progress.
In one embodiment, the game breakpoint processor 616 is configured to define a game breakpoint at predefined completion points. In one embodiment, this is accomplished by finding the nearest completion points to the selected start and end points of the user's recorded gameplay selection, and utilizing these nearest completion points to define the game breakpoints for the game slice. In another embodiment, the nearest completion point occurring before the selected start point of the recorded gameplay selection is utilized to define a starting breakpoint, whereas a nearest completion point occurring after the selected end point of the recorded gameplay selection is utilized to define an ending breakpoint for the creation of the game slice. In still another embodiment, if a completion point lies within a predefined radius of (e.g. either before or after) either of the start or end points of the user's recorded gameplay selection, then that completion point is utilized to define a corresponding start or end game breakpoint for the game slice. Whereas if no completion point lies within the predefined radius, then a game breakpoint is defined that more closely matches the user's selected start or end point for the recorded selection. In other embodiments, the predefined radius for the start and end points may differ for purposes of determining whether to utilize an existing completion point to define a game breakpoint.
As has been discussed, the game breakpoint processor 616 determines appropriate breakpoints applicable to various aspects of the videogame based on analysis of the users recorded gameplay selection. The breakpoints defined by the processor 616 serve to define the limited scope of the game slice that will be produced based on users recorded gameplay selection. In one embodiment, an overlay processor 617 is provided for generating overlays that may contribute to an improved user experience when playing the game slice generated by the game slice generator 612. For example, in one embodiment the overlay processor 617 defines pre-game slice data which defines video or gameplay or additional information that can be provided as an introduction to the game slice prior to actual gameplay of the game slice. One example of pre-game slice data is an introductory video which can provide context to a user who initiates gameplay of became slice. In another embodiment, pregame slice data can define introductory gameplay for the game slice, that may provide a user an opportunity to learn skills that may be useful or required for playing game the slice. In another embodiment, pregame slice data may define a series of one or more informational screens or images which provide information about the game slice to the user. Such information may include controller configuration, story background information, objectives or goals, maps, or any other type of information relating to the game slice which may be useful for the user or otherwise improve the user's experience of playing the game slice.
The overlay processor 617 can also be configured to define postgame slice data. In some embodiments, the postgame slice data can define video or images to be shown following completion of game play of the game slice. For example, a congratulatory video might be shown after a user completes the game slice. Such a video may be customized based on the user's gameplay of the game slice, for example, by showing information or images that are based on the user's gameplay. In one embodiment, the postgame slice data can define a playback mechanism to play recorded portions of the user's gameplay of the game slice following its completion. In another embodiment, the postgame slice data can be configured to display statistics about the users gameplay of the game slice, and may indicate a comparison of the users gameplay to that of other users or that of the original creator of the game slice. In still other embodiments, the postgame slice data can define additional interactive elements to be presented to the user upon completion of the game slice. These may include options to purchase in part or in whole the video game upon which the game slice is based, redirect options to additional sources of information relating to the videogame, etc.
In some embodiments, the overlay processor 617 can be configured to define elements which are overlaid into the game slice. These may include elements that may be customized by a user playing the game slice, such as customization of characters, objects, properties, and other types of customization options. In some embodiments, the overlay processor 617 may be configured to define simplified elements for a game slice so as to reduce the complexity of the game slice code and the amount of resources required to execute the game slice. By way of example, many video games include artificial intelligence (AI) entities such as characters, vehicles, enemies, etc. These AI entities may in the full videogame be governed by artificial intelligence models that define the reaction and activity of the AI entities based on events occurring in the videogame. However, in the context of a game slice which is of limited scope, it may be acceptable to simply define the activity of an AI entity through hardcoded definition or simplified extrapolations, rather than fully modeling the activity of the AI entity as would be the case in the full videogame.
For example, if in the recorded gameplay selection of the full videogame a given AI character moves in a certain manner according to its AI model that is unlikely to change in the game slice, then it may be more efficient to define an approximation of the AI character's movement for the game slice. Such an approximation would not require the full AI model to be included as part of the game slice code, yet would provide to the user playing the game slice a substantially similar experience with respect to that AI character to that of the original user's gameplay from which the game slice was generated. The resource savings realized through approximation of AI entities activity can be even more significant when multiple AI entities are present and interacting in the user's recorded gameplay selection. For each of the AI entities may have AI models that depend upon the output of those of the other AI entities. However, when gameplay has been recorded the activity of each of these AI entities is known, and therefore can be reproduced in the game slice through simplified mechanisms such as direct encoding of their controlling variables and approximation of their activity.
With continued reference to
It should be appreciated that in one embodiment, the game slice code 622 is fully self-contained, including all code portions which are required to execute the game slice. However, in other embodiments, the game slice code 622 may incorporate references or pointers to existing code portions in the main game code of the full video game. Furthermore, the game slice code 622 may include reference or utilize existing assets in asset libraries of the main game code of the full video game. However, in other embodiments, new asset libraries can be generated for the game slice code.
It will be appreciated that the portion of a virtual space defined for a game slice or minigame can be defined by boundaries which are determined based on the user's recorded gameplay. The boundaries will define a sub-region of the larger virtual space, and include a subset of the features which are available in the larger virtual space. In some embodiments, virtual space boundaries can be determined by determining locations in the virtual space defined by the user's gameplay, and then determining predefined boundaries associated with the virtual space that are nearest to those locations and arranged so as to encompass them. For example, a user's gameplay may define a path traversed by a user's video game character. This path can be analyzed, and based on the path's location in the virtual space, a set of predefined boundaries can be selected so as to define a region encompassing the path. In some embodiments, predefined boundaries can be defined by specific features which inherently define portions of the virtual space, e.g. doors, windows, walls, rooms, hallways, fences, roads, intersections, hallways, etc.
In one embodiment, the method can include method operation 816, wherein suggested game slice video selections are generated based on the recorded gameplay video and data of the user. Suggested selections of the users gameplay can be determined based on analysis of the recorded gameplay of the user. For example, portions of the gameplay wherein a high level of activity (e.g. a level of activity exceeding a predefined threshold) is detected might be suggested as possible game slice video selections. At method operation 818, the aforementioned user interface for game slice selection can present the suggested game slice video selections to the user. In one embodiment, a screenshot representative of each suggested game slice video selection can be presented to the user. At method operation 820, a selection by the user of one of the suggested game slice video selections is received. Based on the user selection, the game slice code can be generated and stored as has previously been described.
The game slice information page can further include a selectable button 912 for initiating gameplay of the game slice. The game slice information page can also include a details section 914, which can feature various details and statistical information about the game slice, such as the number of plays, the average completion rate, etc. The game slice information page may also include a comments section 916, featuring comments left by users. A sort button 918 can be provided to select various options for sorting the comments (e.g. chronological order, reverse chronological order, most popular, by rating, by relationship to the current user (e.g. comments by friends of the user in a social graph are prioritized), etc.)
In one embodiment, a live active user section 1006 provides live views of the gameplay of friends who are currently online and may be playing a video game. In one embodiment, each user has an option to define whether or not to make their live gameplay available for live viewing by other users. In such an embodiment, live views are presented only from those users who have the option designated to allow their live gameplay to be viewable by other users. In the illustrated embodiment, the live active user section 1006 includes a live view 1008 of the current gameplay of a friend A, as well as a live view 1010 of the current gameplay of a friend B. In one embodiment, the current user can navigate or browse the friends of the friends listing 1004 and/or browse the live views which are available in the live active user section 1006. In one embodiment, a live view can be highlighted when the current user navigates to it, and may be rendered in a manner distinct from other live views. For example, live views might be displayed in a desaturated color scheme by default, but be displayed in a fully saturated color scheme when highlighted or selected. Live views might also be displayed at a lower resolution, framerate, or size by default, but when selected be displayed at a higher resolution, framerate, or size. In this manner, bandwidth can be allocated to a specific live view based on user selection, to present a live view that the user is interested in viewing with greater fidelity than other live views that may be active simultaneously. It should be appreciated that the live view may display not just specific gameplay of a video game, but also other activity of a friend on the cloud gaming system, such as their navigation through menus or other types of activity related to their cloud gaming session.
In one embodiment, live views are available only for those users who are currently actively engaged in gameplay of a video game. In other words, no live view of a given user is available when that user is either offline, or online but not actively engaged in gameplay of a video game. Thus, when a user is performing other non-gameplay activity on the cloud gaming system (e.g. navigating a graphical interface of the cloud gaming system while logged in to the system), such activity is not made available in a live feed for others to view. In another embodiment, the live view can include all activity of a user who is logged in to the cloud gaming system, including both gameplay of the user as well as other non-gameplay activity.
In one embodiment, the live view of a given user can be filtered so as not to expose potentially sensitive or personal information for view by others. For example, the cloud gaming system may support a chat function. As some users may wish for their chats to remain private, an option can be provided to exclude chat logs when presenting a live view. It should be appreciated that a chat function can be implemented during gameplay activity as well as during non-gameplay activity, and may be filtered out of live views of either or both of these circumstances. In another embodiment, aspects of a video game might be filtered out of a live view. For example, a user may wish to keep certain user-defined settings secret as they may confer an advantage to that user during gameplay. Thus, an option may be provided for activity related to settings to not be displayed as part of the live feed (e.g. when a user accesses a settings interface of the video game). In another embodiment, the live view can be configured to prevent viewing of personal information (e.g. preventing viewing when a user accesses a personal information page, enters payment information, enters a password, etc.).
In one embodiment, the interface provides an option to the primary user to request to join the gaming session of a secondary user who is currently online. For example, the primary user may view the secondary user's live gameplay feed and wish to join the secondary user's gameplay. In one embodiment, activation of the option sends a request to the secondary user notifying the secondary user that the primary user wishes to join their session. If the secondary accepts the request, then a multiplayer mode of the video game is initiated, facilitating multiplayer gameplay by the first and second users. In another embodiment, two or more secondary users may already be engaged in multiplayer gameplay. In such an embodiment, the primary user can send a request to join the multiplayer gameplay. Upon acceptance of the request by one of the secondary users (e.g. a designated host of the gameplay session), the primary user is able to join the multiplayer session of the video game. It should be appreciated that the aforementioned interface showing live gameplay feeds of secondary users facilitates the primary user joining the gameplay of secondary users after being able to see their gameplay.
In one embodiment, the option to request to join the gameplay of the secondary user is predicated upon determination of an ownership status of the video game by the primary user. If the primary user does not own the video game, then no option may be presented, whereas if the primary user owns the same video game as the secondary user, then the option to request to join the gameplay of the secondary user is made available as part of the interface. In one embodiment, when the primary user is determined to not be an owner of the video game, then the primary user may still join the gameplay of the secondary user in a multiplayer mode, but in a limited capacity, such as being limited in terms of duration of gameplay, available scenes/levels/stages/etc., customization options, abilities, skills, weapons, characters, vehicles, or any other aspect of the video game that may be limited. In one embodiment, following gameplay of the limited version of the video game, the primary user is provided an option to purchase the video game in full. In another embodiment, the primary user is provided an option to purchase an additional portion of the video game.
With continued reference to
It will be appreciated that many methods and configurations for presenting a cloud gaming interface are possible in accordance with various embodiments of the invention. In one such embodiment, a method for displaying a current gaming status of users of a cloud gaming system is provided, including the following method operations: presenting a cloud gaming interface of a primary user; determining one or more secondary users that are friends of the primary user; determining a current status of each of the secondary users, the current status being one of online or offline; for at least one online secondary user, presenting a live feed of a current gaming session of the online secondary user in the cloud gaming interface of the primary user, the live feed includes providing an option for the primary user to join the current gaming session of the online secondary user; in response to receiving a request to activate the option for the primary user to join the current gaming session of the online secondary user, initiating a multiplayer mode of the current gaming session of the online secondary user, the multiplayer mode providing for gameplay of the primary user in the current gaming session of the online secondary user; wherein initiating the multiplayer mode includes determining an ownership status of the primary user with respect to a video game defining the current gaming session of the online secondary user; wherein when the primary user is determined to own the video game, then the multiplayer mode provides for gameplay of a full version of the video game; wherein when the primary user is determined to not own the video game, then the multiplayer mode provides for gameplay of a limited version of the video game.
In one embodiment, the limited version of the video game defines a reduction, as compared to the full version of the video game, in one or more of available levels, available scenes, available features, a time limit, a virtual space, a campaign length, a number of lives, or a number of replays.
In one embodiment, the method further includes a method operation of presenting, when the video game is not owned by the primary user, an option for the primary user to purchase at least a portion of the video game.
In one embodiment, the live feed of the online secondary user is presented at a first resolution; and selection of the live feed triggers presentation of the live feed of the online secondary user at a second resolution higher than the first resolution.
In one embodiment, the live feed of the online secondary user is presented in a desaturated color mode; and selection of the live feed triggers presentation of the live feed of the online secondary user in a saturated color mode.
In one embodiment, presenting the cloud gaming interface includes presenting a library of game titles associated with each of the secondary users.
In one embodiment, determining the one or more secondary users includes accessing a social graph associated with the primary user. In one embodiment, accessing the social graph includes accessing an API of a social network.
In one embodiment, presenting the cloud gaming interface of the primary user includes presenting a listing of each of the secondary users in a priority order, the priority order based on one or more of current status, recency of login, or common ownership of games with the primary user.
With continued reference to
In one embodiment, the user A opts to share from his gameplay session A to his social graph. User A's session A communicates via the API 1224 to activate the notification module 1226 of the social network 1216 to send an appropriate notification to friends in the social graph of the user A. When a social network user 1230 who is in the social graph of the user A accesses the social network via a device 1228, they may see a message or posting from user A about user A's session. If set up, a user might receive notifications such as an e-mail indicating that user A has shared something on the social network. It should be appreciated that user A may share about various video game related activity, such as achievements in a video game, invitations to play a video game, comments about a video game, an invitation to view or play a game slice the user A has created, a video clip of user A's gameplay, etc.
With reference again to the embodiment of
In various embodiments, a region of interest of the user's recorded gameplay may be automatically determined based on threshold detection of any one or more of the following: one or more user inputs, rate of user input, frequency of input, repeats of types of inputs, occurrences of input patterns, combination inputs (e.g. combo keys), motion vectors, pressure exerted on a controller, excitation of a user (e.g. detected based on captured image or audio data of the user).
In one embodiment, a user can insert additional material, such as a user defined video, message, images, or any other type of information, before or following a game slice. In the illustrated embodiment, an introduction 1512 is provided for introducing a player to the minigame 1510 and perhaps also introducing the player to the first game slice 1514. Additionally, a message 1516 is inserted in between the first game slice 1514 and the second game slice 1518, while another message 1520 is inserted in between the second game slice 1518 and the third game slice 1522. In one embodiment, the inserted material can include recorded gameplay video of the original user's gameplay from which the game slice was generated. If the recorded gameplay is shown before playing the game slice, the player initiating the game slice can better understand the game slice and its objectives before beginning gameplay, whereas is shown after gameplay of the game slice, the player can understand how their gameplay of the game slice compares to that of the original user.
In the illustrated embodiment, it will be appreciated that the game slices are taken from different video games. This allows the user great flexibility to mix together game slices from across different game titles, genres, and even console platform generations. Merely by way of example, a user might create a minigame having game slices from each of several game titles in a singular video game series. In this manner, a player of the minigame is able to experience and appreciate the evolution of the video game series in a continuous gameplay experience.
Embodiments of the invention have generally been described with reference to cloud-based gaming systems. However, it should be appreciated by those skilled in the art that similar concepts and principles as have been described herein can be applied to traditional console-based videogame systems, possibly in conjunction with cloud-based gaming systems. For example, a user can play a console-based videogame and have the user's gameplay input data and game stayed metadata recorded during gameplay. Based on the user's gameplay input data and game state metadata, the actual gameplay output of the videogame can be regenerated at a later time. Therefore, the recorded input data and game state metadata can be utilized in conjunction with the videogame code to provide an interface for selection of a portion of the user's gameplay from which to generate a minigame, as has been described. The minigame code can be generated at the console and uploaded to a cloud system and made available for download by other users. In another embodiment, the minigame code is generated by the cloud system after receiving the selected portion of the users gameplay input data and game state metadata. The cloud system processes the selected portion of the users gameplay input data and game state metadata to generate the minigame code based on videogame code stored in the cloud system. Once generated, the minigame can be made available for cloud-based gaming wherein execution of the minigame occurs in the cloud-based system, but can also be made available for download to traditional console-based systems for execution on the console to facilitate console-based gameplay of the minigame. In this manner, minigame's can be created and played by both users of console-based video gaming systems and users of cloud-based video gaming systems.
Embodiments of the invention have generally been described with reference to minigames or playable user-defined portions of video games. However, it will be appreciated by those skilled in the art that many of the principles elucidated herein also readily apply to the generation and sharing of recorded gameplay, including sharing recorded gameplay video, screenshots, and live streaming of active gameplay. In some embodiments, providing access to a minigame (e.g. in response to receiving a notification) can include a presentation of recorded video of the gameplay of the original user which formed the basis for the minigame. In still other embodiments, methods, systems, and interfaces are contemplated for facilitating sharing of a user's gameplay to the user's social graph.
In one embodiment, a method for storing gameplay is contemplated. Gameplay can be executed by the operating system of a game console in response to a user request, which can come in the form of a standard file operation with respect to a set of data associated with the desired gameplay. The request can be transmitted from an application associated with a game. The gameplay can comprise, for example, video content, audio content and/or static visual content, including wallpapers, themes, code “add-on” content, or any other type of content associated with a game. It is contemplated that such content can be user- or developer-generated, free or paid, full or trial, and/or for sale or for rent.
A portion of the gameplay can be buffered, i.e., stored temporarily. For example, the previous 15 seconds, the previously completed level, for the previous action within the gameplay can be stored temporarily, as described further herein. The term “portion” used herein can correspond to any part of the gameplay that is divisible into any related or arbitrary groups of single or multiple bits or bytes of data. For example, “portions” of gameplay may correspond to levels, chapters, scenes, acts, characters, backgrounds, textures, courses, actions, songs, themes, durations, sizes, files, parts thereof, and combinations thereof. Further, portions of gameplay can comprise screenshots or prescribed durations of video capture.
In one embodiment, portions of the gameplay can be stored locally on the game console in either temporary or permanent storage. Alternatively or additionally, portions of the gameplay can be transmitted over a network stored remotely. For example, portions of the gameplay can be transmitted over a wireless or wired network to another computing device, to another game console, or to a remote server. Such remote servers may include social media servers.
Optionally, portions of the gameplay not retrieved from the buffer or portions of the gameplay outside a particular gaming interval (e.g., a particular duration, level, chapter, course, etc.) can be removed from the buffer. This removal process can be completed using standard file operations on the operating system.
The portions of the gameplay can be displayed on any of a number of display devices having access to the stored gameplay. For example, the stored gameplay can be displayed on a television set connected to the game console from which the gameplay was captured. In another example the stored gameplay can be displayed on a computer to which the stored gameplay was transmitted. The stored gameplay can be displayed alone or in conjunction with other information, such as on a social media website.
In one embodiment, portions of the gameplay are displayed by another game console associated with the user other than the user that buffered or captured the gameplay. According to this embodiment, the portions of the gameplay may show a ball being thrown from a first user to a second user, from the point of view of the first user. The portions of gameplay can then be transmitted to the game console of the second user. Thus, the second user can then view the gameplay from the point of view of the first user. The second user can also have portions of gameplay stored showing the ball being thrown by the first user and caught a second user, from the point of view of the second user. In this embodiment the second user can review the gameplay from both point of view the first user and the point of view of the second user. Still further, the portions of the gameplay stored by the second user can be transmitted to the game console of the first user, so that the first user may review the gameplay from two points of view. This embodiment can apply to any number of users having any number of points of view, so the gameplay can be reviewed from any number of different perspectives.
With respect to storage, transmission and/or display of the portions of the gameplay as described herein, it is contemplated that portions of the gameplay can be stored, transmitted and displayed as image or video data. In another embodiment, however, portions of the gameplay can be stored and transmitted as telemetry or metadata representative of the image or video data, and can be re-created as images or video by a game console or other device prior to display.
In some embodiments, the portion of the gameplay has a predetermined relationship with the executed gameplay. For example, the portion of the gameplay can correspond to a certain amount of gameplay prior to the currently executing gameplay, such as the previous 10 seconds of gameplay. In another embodiment, a first portion of the gameplay has a predetermined relationship with a second portion of the gameplay. For example, the first portion of the gameplay can correspond to a certain amount of gameplay prior to receipt of a request to capture a second portion of gameplay, such as the 10 seconds of gameplay prior to selection of a capture button. In each of these embodiments, the amount of gameplay buffered prior to the current gameplay or the requested gameplay can be configured and adjusted by the user according to his or her particular preferences.
In other embodiments, the buffer is “smart” or “elastic,” such that it captures gameplay according to variables without regard to time. In one such embodiment, the first portion of the gameplay has a predetermined relationship with an event related to the gameplay. For example the first portion of the gameplay may be buffered to include a statistical anomaly, such as a high score being reached, the gathering of a large number of points in a short amount of time, the multiple selections of buttons on a controller, and other rare events. Such statistical anomalies can be determined by comparing gameplay metrics to average metrics for a particular game or scene or for all games generally. Such average metrics can be stored locally or remotely for comparison. For example, the game console can track global high scores for a particular game, and buffer gameplay in which a user approaches and surpasses that high score. In another example, a remote server can track global high scores for a particular game, and can communicate that information to the game console, which buffers gameplay in which the user approaches and surpasses that high score.
In another example, the portion of the gameplay can be buffered to include an achievement, such as a trophy being attained or other landmark being reached. Such trophies or landmarks memorialized any goal or gaming achievement, such as a certain number of points being attained, a certain level being reached, and the like. For example, gameplay can be buffered to include the awarding of a trophy for reaching level 10, for reaching 100,000 points, etc.
Similarly, progress toward reaching an event, in addition to the actual attainment of the trophy or statistical anomaly, can be buffered to be included in the portion of the gameplay. For example, a screenshot can be taken at each of levels one through 10, creating a photo album to memorialize the receipt of a trophy for reaching level 10. Another example, a video can be taken of the user winning a race for the first through fifth times, where a trophy is awarded for five wins.
Thus, according to embodiments of the invention, at least a portion of executed gameplay can always be kept in a running buffer. In other words, when a request to share a portion of the gameplay is received, a portion of the prior gameplay can already be captured to include previous footage. For example, if a request to share gameplay is received after a user crosses the finish line in a racing game, the buffered gameplay can include footage of the user crossing the finish line. In other words, a user will be able to capture moments occurring before a request is made to share gameplay.
The icon 1702 can be selected to initiate uploading of a screenshot from the user's gameplay. In one embodiment, the region 1704 of the icon 1702 is populated with a representative screenshot from the user's recent gameplay, thereby providing a small-scale preview of the screenshot that may be shared. In one embodiment, selection of the icon 1702 may provide access to an additional screenshot selection interface which allows the user to select a particular screenshot from the user's recorded gameplay for sharing to the user's social graph. This may take the form of a navigable gameplay video timeline, which can be traversed or navigated to identify a particular time point within the gameplay and its corresponding screenshot.
The icon 1706 can be selected to initiate uploading of a video of the user's gameplay. The region 1708 of the icon 1706 can be configured to display a representative video clip from the user's recent gameplay, e.g. the last 5 seconds of the user's gameplay, a recent achievement, etc., thereby displaying a small-scale preview of a video clip from the user's gameplay which can be shared with others.
The icon 1710 can be selected to initiate live video streaming of the user's active gameplay. In one embodiment, selection of the icon 1710 will trigger resumption of the user's gameplay while initiating live video streaming of the user's gameplay. In another embodiment, selection of the icon 1710 provides access to an interface to determine settings for the streaming video broadcast, such as who to share the streaming video with, whether to include a video stream from a local image capture device intended to show the user himself/herself during the gameplay, whether to enable comments, etc.
It will be appreciated that the user may share gameplay (e.g. a selected screenshot, video, or live gameplay stream) to one or more specifically selected friends, to their entire social graph, or to any user of the social network. The social network can be a gaming social network associated with the platform on which the video game is run, or a third-party social network that exists separate from the video game or its platform. The social network can be accessed through an API defined to allow interfacing with the social network. Users to whom gameplay has been shared may receive a notification informing them of the shared gameplay. Such a notification may take the form of a posting to a social news feed, a private message through the social network, an in-game notification, an e-mail, a chat notification, etc. Sharing gameplay to the social network may entail making the gameplay available to other other subsets of users of the social network who may or may not be part of the sharing user's social graph. For example, for a given video game, gameplay may be shared or made available to any user of the social network who also owns the video game and therefore is granted access to shared gameplay of the video game. Such shared gameplay may be accessed through online forums, chat rooms, or other online channels that are available only to players of the video game. In one embodiment, a video game may have a dedicated page or site on the social network. Shared gameplay can be made available to users accessing the page or site of the video game game. Of course, it will be appreciated that from the perspective of the sharing user, options can be provided to allow the user to specify and tailor who and what forum to which their gameplay will be shared.
In one embodiment, buttons 1818 and 1820 can be selected to reduce or increase, respectively, the duration of the selected video portion. A preview option 1822 can be selected to trigger playback of a fullscreen preview of the currently selected video clip. A trimming option 1824 can be selected to access additional video trimming features.
An option 1912 allows the viewing user to join the gameplay of the streaming user. And an option 1914 enables the viewing user to purchase the in-progress video game.
Furthermore, the interface can include a comments section 1916 which displays comments from users viewing the live video stream, indicating the time of each comment. A comment input field 1918 is provided for the viewing user to enter text for a comment to be posted. And the send button 1920 is pressed to upload the comment for viewing by the live streaming user and others viewing the live video stream.
Though embodiments of the invention have been described with reference to accessing various interfaces for sharing from a dedicated button press, it will be appreciated that in other embodiments, some or all of these interfaces may not be required to facilitate sharing of gameplay to a user social graph. For example, in one embodiment, a controller button can be configured to capture a screenshot of the users gameplay when pressed. The captured screenshot can then be automatically uploaded and shared to the user social graph.
In another embodiment, pressing a specific button on the controller initiatiates recording of gameplay video. When the specific button is pressed a second time, recording of the gameplay video is stopped, and the video clip can be uploaded and shared to the user social graph. In one embodiment, the uploading and sharing of the video clip to the user social graph may take place automatically following termination of the video recording operation. However, in another embodiment, when the specific button is pressed a second time to stop recording, an interface is presented to allow the user to customize various options such as trimming the video, selecting a representative screenshot for the video, determine specific users with whom to share the video with, add a caption or title, etc. After customization by the user, the video can be shared with others or otherwise made available for viewing.
In one embodiment, a specific button on the controller can be configured to share a predefined duration of gameplay video on a social network. For example, a user might specify that when the button is pressed, the previous 10 seconds of gameplay video will be shared to the user social graph. In another embodiment, it may be specified that when the button is pressed, the next 10 seconds of gameplay video will be recorded and shared to the social graph. It should be appreciated that options for trimming the video and performing other types of customization may be applied to the recorded gameplay video. Furthermore, recorded gameplay video of a predefined duration following the button trigger can be combined with previously buffered gameplay video as has been described.
In yet another embodiment, a specific button on the controller device can be configured to initiate live video streaming of the users active gameplay. A live video streaming can be predefined to be made available only to members of the user social graph, or to other smaller or larger groups of users, such as a specific subset of the user social graph, all users who own or otherwise have access to the same video game, any user of the gaming platform, etc.
The I/O bridge 2034 also connects to six Universal Serial Bus (USB) 2.0 ports 2024; a gigabit Ethernet port 2022; an IEEE 802.11b/g wireless network (Wi-Fi) port 2020; and a Bluetooth® wireless link port 2018 capable of supporting up to seven Bluetooth connections.
In operation, the I/O bridge 2034 handles all wireless, USB and Ethernet data, including data from one or more game controllers 2002-2003. For example when a user is playing a game, the I/O bridge 2034 receives data from the game controller 2002-2003 via a Bluetooth link and directs it to the Cell processor 2028, which updates the current state of the game accordingly.
The wireless, USB and Ethernet ports also provide connectivity for other peripheral devices in addition to game controllers 2002-2003, such as: a remote control 2004; a keyboard 2006; a mouse 2008; a portable entertainment device 2010 such as a Sony Playstation Portable® entertainment device; a video camera such as an EyeToy® video camera 2012; a microphone headset 2014; and a microphone 2015. Such peripheral devices may therefore in principle be connected to the system unit 2000 wirelessly; for example the portable entertainment device 2010 may communicate via a Wi-Fi ad-hoc connection, whilst the microphone headset 2014 may communicate via a Bluetooth link.
The provision of these interfaces means that the Playstation 3 device is also potentially compatible with other peripheral devices such as digital video recorders (DVRs), set-top boxes, digital cameras, portable media players, Voice over IP telephones, mobile telephones, printers and scanners.
In addition, a legacy memory card reader 2016 may be connected to the system unit via a USB port 2024, enabling the reading of memory cards 2048 of the kind used by the Playstation® or Playstation 2® devices.
The game controllers 2002-2003 are operable to communicate wirelessly with the system unit 2000 via the Bluetooth link, or to be connected to a USB port, thereby also providing power by which to charge the battery of the game controllers 2002-2003. Game controllers 2002-2003 can also include memory, a processor, a memory card reader, permanent memory such as flash memory, light emitters such as an illuminated spherical section, LEDs, or infrared lights, microphone and speaker for ultrasound communications, an acoustic chamber, a digital camera, an internal clock, a recognizable shape such as the spherical section facing the game console, and wireless communications using protocols such as Bluetooth®, WiFi™, etc.
Game controller 2002 is a controller designed to be used with two hands, and game controller 2003 is a single-hand controller with an attachment. In addition to one or more analog joysticks and conventional control buttons, the game controller is susceptible to three-dimensional location determination. Consequently gestures and movements by the user of the game controller may be translated as inputs to a game in addition to or instead of conventional button or joystick commands. Optionally, other wirelessly enabled peripheral devices such as the Playstation™ Portable device may be used as a controller. In the case of the Playstation™ Portable device, additional game or control information (for example, control instructions or number of lives) may be provided on the screen of the device. Other alternative or supplementary control devices may also be used, such as a dance mat (not shown), a light gun (not shown), a steering wheel and pedals (not shown) or bespoke controllers, such as a single or several large buttons for a rapid-response quiz game (also not shown).
The remote control 2004 is also operable to communicate wirelessly with the system unit 2000 via a Bluetooth link. The remote control 704 comprises controls suitable for the operation of the Blu Ray™ Disk BD-ROM reader 2040 and for the navigation of disk content.
The Blu Ray™ Disk BD-ROM reader 2040 is operable to read CD-ROMs compatible with the Playstation and PlayStation 2 devices, in addition to conventional pre-recorded and recordable CDs, and so-called Super Audio CDs. The reader 2040 is also operable to read DVD-ROMs compatible with the Playstation 2 and PlayStation 3 devices, in addition to conventional pre-recorded and recordable DVDs. The reader 2040 is further operable to read BD-ROMs compatible with the Playstation 3 device, as well as conventional pre-recorded and recordable Blu-Ray Disks.
The system unit 2000 is operable to supply audio and video, either generated or decoded by the Playstation 3 device via the Reality Synthesizer graphics unit 2030, through audio and video connectors to a display and sound output device 2042 such as a monitor or television set having a display 2044 and one or more loudspeakers 2046. The audio connectors 2050 may include conventional analogue and digital outputs whilst the video connectors 2052 may variously include component video, S-video, composite video and one or more High Definition Multimedia Interface (HDMI) outputs. Consequently, video output may be in formats such as PAL or NTSC, or in 720p, 1080i or 1080p high definition.
Audio processing (generation, decoding and so on) is performed by the Cell processor 2028. The Playstation 3 device's operating system supports Dolby® 5.1 surround sound, Dolby® Theatre Surround (DTS), and the decoding of 7.1 surround sound from Blu-Ray® disks.
In the present embodiment, the video camera 2012 comprises a single charge coupled device (CCD), an LED indicator, and hardware-based real-time data compression and encoding apparatus so that compressed video data may be transmitted in an appropriate format such as an intra-image based MPEG (motion picture expert group) standard for decoding by the system unit 2000. The camera LED indicator is arranged to illuminate in response to appropriate control data from the system unit 2000, for example to signify adverse lighting conditions. Embodiments of the video camera 2012 may variously connect to the system unit 700 via a USB, Bluetooth or Wi-Fi communication port. Embodiments of the video camera may include one or more associated microphones and also be capable of transmitting audio data. In embodiments of the video camera, the CCD may have a resolution suitable for high-definition video capture. In use, images captured by the video camera may for example be incorporated within a game or interpreted as game control inputs. In another embodiment the camera is an infrared camera suitable for detecting infrared light.
In general, in order for successful data communication to occur with a peripheral device such as a video camera or remote control via one of the communication ports of the system unit 2000, an appropriate piece of software such as a device driver should be provided. Device driver technology is well-known and will not be described in detail here, except to say that the skilled man will be aware that a device driver or similar software interface may be required in the present embodiment described.
The Power Processing Element (PPE) 2150 is based upon a two-way simultaneous multithreading Power 570 compliant PowerPC core (PPU) 855 running with an internal clock of 3.2 GHz. It comprises a 512 kB level 2 (L2) cache and a 32 kB level 1 (L1) cache. The PPE 2150 is capable of eight single position operations per clock cycle, translating to 25.6 GFLOPs at 3.2 GHz. The primary role of the PPE 2150 is to act as a controller for the Synergistic Processing Elements 2110A-H, which handle most of the computational workload. In operation the PPE 2150 maintains a job queue, scheduling jobs for the Synergistic Processing Elements 2110A-H and monitoring their progress. Consequently each Synergistic Processing Element 2110A-H runs a kernel whose role is to fetch a job, execute it and synchronized with the PPE 2150.
Each Synergistic Processing Element (SPE) 2110A-H comprises a respective Synergistic Processing Unit (SPU) 2120A-H, and a respective Memory Flow Controller (MFC) 2140A-H comprising in turn a respective Dynamic Memory Access Controller (DMAC) 2142A-H, a respective Memory Management Unit (MMU) 2144A-H and a bus interface (not shown). Each SPU 2120A-H is a RISC processor clocked at 3.2 GHz and comprising 256 kB local RAM 2130A-H, expandable in principle to 4 GB. Each SPE gives a theoretical 25.6 GFLOPS of single precision performance. An SPU can operate on 4 single precision floating point members, 4 32-bit numbers, 8 16-bit integers, or 16 8-bit integers in a single clock cycle. In the same clock cycle it can also perform a memory operation. The SPU 2120A-H does not directly access the system memory XDRAM 2026; the 64-bit addresses formed by the SPU 2120A-H are passed to the MFC 2140A-H which instructs its DMA controller 2142A-H to access memory via the Element Interconnect Bus 2180 and the memory controller 2160.
The Element Interconnect Bus (EIB) 2180 is a logically circular communication bus internal to the Cell processor 2028 which connects the above processor elements, namely the PPE 2150, the memory controller 2160, the dual bus interface 2170A,B and the 8 SPEs 2110A-H, totaling 12 participants. Participants can simultaneously read and write to the bus at a rate of 8 bytes per clock cycle. As noted previously, each SPE 2110A-H comprises a DMAC 2142A-H for scheduling longer read or write sequences. The EIB comprises four channels, two each in clockwise and anti-clockwise directions. Consequently for twelve participants, the longest step-wise data-flow between any two participants is six steps in the appropriate direction. The theoretical peak instantaneous EIB bandwidth for 12 slots is therefore 96B per clock, in the event of full utilization through arbitration between participants. This equates to a theoretical peak bandwidth of 307.2 GB/s (gigabytes per second) at a clock rate of 3.2 GHz.
The memory controller 2160 comprises an XDRAM interface 2162, developed by Rambus Incorporated. The memory controller interfaces with the Rambus XDRAM 2026 with a theoretical peak bandwidth of 25.6 GB/s.
The dual bus interface 2170A,B comprises a Rambus FlexIO® system interface 2172A,B. The interface is organized into 12 channels each being 8 bits wide, with five paths being inbound and seven outbound. This provides a theoretical peak bandwidth of 62.4 GB/s (36.4 GB/s outbound, 26 GB/s inbound) between the Cell processor and the I/O Bridge 2034 via controller 2170A and the Reality Simulator graphics unit 2130 via controller 2170B.
Data sent by the Cell processor 2128 to the Reality Simulator graphics unit 2130 will typically comprise display lists, being a sequence of commands to draw vertices, apply textures to polygons, specify lighting conditions, and so on.
A user interacts with the game client via controller. In some embodiments the controller is a game client specific controller while in other embodiments, the controller can be a keyboard and mouse combination. In one embodiment, the game client is a standalone device capable of outputting audio and video signals to create a multimedia environment through a monitor/television and associated audio equipment. For example, the game client can be, but is not limited to a thin client, an internal PCI-express card, an external PCI-express device, an ExpressCard device, an internal, external, or wireless USB device, or a Firewire device, etc. In other embodiments, the game client is integrated with a television or other multimedia device such as a DVR, Blu-Ray player, DVD player or multi-channel receiver.
Within scene A of
When a game client 2202 connects to a server processing module, user session control may be used to authenticate the user. An authenticated user can have associated virtualized distributed storage and virtualized network processing. Examples items that can be stored as part of a user's virtualized distributed storage include purchased media such as, but not limited to games, videos and music etc. Additionally, distributed storage can be used to save game status for multiple games, customized settings for individual games, and general settings for the game client. In one embodiment, the user geo-location module of the server processing is used to determine the geographic location of a user and their respective game client. The user's geographic location can be used by both the sharing/communication logic and the load balance processing service to optimize performance based on geographic location and processing demands of multiple server processing modules. Virtualizing either or both network processing and network storage would allow processing tasks from game clients to be dynamically shifted to underutilized server processing module(s). Thus, load balancing can be used to minimize latency associated with both recall from storage and with data transmission between server processing modules and game clients.
The server processing module has instances of server application A and server application B. The server processing module is able to support multiple server applications as indicated by server application X1 and server application X2. In one embodiment, server processing is based on cluster computing architecture that allows multiple processors within a cluster to process server applications. In another embodiment, a different type of multi-computer processing scheme is applied to process the server applications. This allows the server processing to be scaled in order to accommodate a larger number of game clients executing multiple client applications and corresponding server applications. Alternatively, server processing can be scaled to accommodate increased computing demands necessitated by more demanding graphics processing or game, video compression, or application complexity. In one embodiment, the server processing module performs the majority of the processing via the server application. This allows relatively expensive components such as graphics processors, RAM, and general processors to be centrally located and reduces to the cost of the game client. Processed server application data is sent back to the corresponding game client via the internet to be displayed on a monitor.
Scene C illustrates an exemplary application that can be executed by the game client and server processing module. For example, in one embodiment game client 2202C allows user C to create and view a buddy list 2220 that includes user A, user B, user D and user E. As shown, in scene C, user C is able to see either real time images or avatars of the respective user on monitor 2204C. Server processing executes the respective applications of game client 2202C and with the respective game clients 2202 of users A, user B, user D and user E. Because the server processing is aware of the applications being executed by game client B, the buddy list for user A can indicate which game user B is playing. Further still, in one embodiment, user A can view actual in game video directly from user B. This is enabled by merely sending processed server application data for user B to game client A in addition to game client B.
In addition to being able to view video from buddies, the communication application can allow real-time communications between buddies. As applied to the previous example, this allows user A to provide encouragement or hints while watching real-time video of user B. In one embodiment two-way real time voice communication is established through a client/server application. In another embodiment, a client/server application enables text chat. In still another embodiment, a client/server application converts speech to text for display on a buddy's screen.
Scene D and scene E illustrate respective user D and user E interacting with game consoles 2210D and 2210E respectively. Each game console 2210D and 2210E are connected to the server processing module and illustrate a network where the server processing modules coordinates game play for both game consoles and game clients.
ISP 2370 includes Application Service Provider (ASP) 2372, which provides computer-based services to customers over a network. Software offered using an ASP model is also sometimes called on-demand software or software as a service (SaaS). A simple form of providing access to a particular application program (such as customer relationship management) is by using a standard protocol such as HTTP. The application software resides on the vendor's system and is accessed by users through a web browser using HTML, by special purpose client software provided by the vendor, or other remote interface such as a thin client.
Services delivered over a wide geographical area often use cloud computing. Cloud computing is a style of computing in which dynamically scalable and often virtualized resources are provided as a service over the Internet. Users do not need to be an expert in the technology infrastructure in the “cloud” that supports them. Cloud computing can be divided in different services, such as Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). Cloud computing services often provide common business applications online that are accessed from a web browser, while the software and data are stored on the servers. The term cloud is used as a metaphor for the Internet, based on how the Internet is depicted in computer network diagrams and is an abstraction for the complex infrastructure it conceals.
Further, ISP 2370 includes a Game Processing Server (GPS) 2374 which is used by game clients to play single and multiplayer video games. Most video games played over the Internet operate via a connection to a game server. Typically, games use a dedicated server application that collects data from players and distributes it to other players. This is more efficient and effective than a peer-to-peer arrangement, but it requires a separate server to host the server application. In another embodiment, the GPS establishes communication between the players and their respective game-playing devices exchange information without relying on the centralized GPS.
Dedicated GPSs are servers which run independently of the client. Such servers are usually run on dedicated hardware located in data centers, providing more bandwidth and dedicated processing power. Dedicated servers are the preferred method of hosting game servers for most PC-based multiplayer games. Massively multiplayer online games run on dedicated servers usually hosted by the software company that owns the game title, allowing them to control and update content.
Broadcast Processing Server (BPS) 2376 distributes audio or video signals to an audience. Broadcasting to a very narrow range of audience is sometimes called narrowcasting. The final leg of broadcast distribution is how the signal gets to the listener or viewer, and it may come over the air as with a radio station or TV station to an antenna and receiver, or may come through cable TV or cable radio (or “wireless cable”) via the station or directly from a network. The Internet may also bring either radio or TV to the recipient, especially with multicasting allowing the signal and bandwidth to be shared. Historically, broadcasts have been delimited by a geographic region, such as national broadcasts or regional broadcast. However, with the proliferation of fast internet, broadcasts are not defined by geographies as the content can reach almost any country in the world.
Storage Service Provider (SSP) 2378 provides computer storage space and related management services. SSPs also offer periodic backup and archiving. By offering storage as a service, users can order more storage as required. Another major advantage is that SSPs include backup services and users will not lose all their data if their computers' hard drives fail. Further, a plurality of SSPs can have total or partial copies of the user data, allowing users to access data in an efficient way independently of where the user is located or the device being used to access the data. For example, a user can access personal files in the home computer, as well as in a mobile phone while the user is on the move.
Communications Provider 380 provides connectivity to the users. One kind of Communications Provider is an Internet Service Provider (ISP) which offers access to the Internet. The ISP connects its customers using a data transmission technology appropriate for delivering Internet Protocol datagrams, such as dial-up, DSL, cable modem, wireless or dedicated high-speed interconnects. The Communications Provider can also provide messaging services, such as e-mail, instant messaging, and SMS texting. Another type of Communications Provider is the Network Service provider (NSP) which sells bandwidth or network access by providing direct backbone access to the Internet. Network service providers may consist of telecommunications companies, data carriers, wireless communications providers, Internet service providers, cable television operators offering high-speed Internet access, etc.
Data Exchange 2388 interconnects the several modules inside ISP 2370 and connects these modules to users 2382 via network 2386. Data Exchange 2388 can cover a small area where all the modules of ISP 2370 are in close proximity, or can cover a large geographic area when the different modules are geographically dispersed. For example, Data Exchange 2388 can include a fast Gigabit Ethernet (or faster) within a cabinet of a data center, or an intercontinental virtual area network (VLAN).
Users 2382 access the remote services with client device 2384, which includes at least a CPU, a display and I/O. The client device can be a PC, a mobile phone, a netbook, a PDA, etc. In one embodiment, ISP 2370 recognizes the type of device used by the client and adjusts the communication method employed. In other cases, client devices use a standard communications method, such as html, to access ISP 2370.
Embodiments of the present invention may be practiced with various computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like. The invention can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a wire-based or wireless network.
With the above embodiments in mind, it should be understood that the invention can employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Any of the operations described herein that form part of the invention are useful machine operations. The invention also relates to a device or an apparatus for performing these operations. The apparatus can be specially constructed for the required purpose, or the apparatus can be a general-purpose computer selectively activated or configured by a computer program stored in the computer. In particular, various general-purpose machines can be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
The invention can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data, which can thereafter be read by a computer system. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes and other optical and non-optical data storage devices. The computer readable medium can include computer readable tangible medium distributed over a network-coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
Although the method operations were described in a specific order, it should be understood that other housekeeping operations may be performed in between operations, or operations may be adjusted so that they occur at slightly different times, or may be distributed in a system which allows the occurrence of the processing operations at various intervals associated with the processing, as long as the processing of the overlay operations are performed in the desired way.
Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications can be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.
Claims
1. A method for enabling multiplayer gameplay, comprising:
- executing, by a cloud gaming server, a first user's gameplay of a video game, wherein executing the first user's gameplay includes processing input received over a network from a first controller device that is associated with the first user to direct gameplay events occurring in the first user's gameplay;
- presenting, over the network, a live video feed of the first user's gameplay to a remote second user;
- processing a request for the second user to join the first user's gameplay;
- initiating gameplay by the second user in the first user's gameplay, wherein initiating gameplay by the second user includes processing input received over the network from a second controller device that is associated with the second user to direct at least some of the gameplay events occurring in the first user's gameplay.
2. The method of claim 1, wherein initiating gameplay by the second user in the first user's gameplay includes initiating a multiplayer mode of the video game that facilitates multiplayer gameplay by the first user and the second user.
3. The method of claim 2, wherein the multiplayer mode includes continuing the processing of the input received from the first controller device while actively processing the input received from the second controller device.
4. The method of claim 1, wherein the multiplayer gameplay is configured to enable one or more additional players to join based on availability or priority ordering of said one or more additional players.
5. The method of claim 1, wherein one or more additional players are enabled to join the first user's gameplay or discontinue said joining.
6. The method of claim 1, wherein said first player's gameplay is associated with a gaming process, and said gaming process is configured to run to enable said second user or other users to join or discontinue said joining over time.
7. The method of claim 6, wherein an interface is provided by the cloud gaming server to enable access to one or more other video games or game functions, said cloud gaming server enabling navigation to the interface and said video game while the gaming process remains available to continue gameplay by said first user, second user, or one or more users that join the multiplayer gameplay of the video game.
8. The method of claim 6, wherein during said gaming process control of the first user's gameplay is transitioned from the first controller device to the second controller device, or from the second controller device to the first controller device.
9. A non-transitory computer readable medium having program instructions embodied thereon that, when executed by at least one computing device, cause said at least one computing device to perform a method for enabling multiplayer gameplay, said method including the following method operations:
- executing, by a cloud gaming server, a first user's gameplay of a video game, wherein executing the first user's gameplay includes processing input received over a network from a first controller device that is associated with the first user to direct gameplay events occurring in the first user's gameplay;
- presenting, over the network, a live video feed of the first user's gameplay to a remote second user;
- processing a request for the second user to join the first user's gameplay;
- initiating gameplay by the second user in the first user's gameplay, wherein initiating gameplay by the second user includes processing input received over the network from a second controller device that is associated with the second user to direct at least some of the gameplay events occurring in the first user's gameplay.
10. The non-transitory computer readable medium of claim 9, wherein initiating gameplay by the second user in the first user's gameplay includes initiating a multiplayer mode of the video game that facilitates multiplayer gameplay by the first user and the second user.
11. The non-transitory computer readable medium of claim 10, wherein the multiplayer mode includes continuing the processing of the input received from the first controller device while actively processing the input received from the second controller device.
12. The non-transitory computer readable medium of claim 9, wherein the multiplayer gameplay is configured to enable one or more additional players to join based on availability or priority ordering of said one or more additional players.
13. The non-transitory computer readable medium of claim 9, wherein one or more additional players are enabled to join the first user's gameplay or discontinue said joining.
14. The non-transitory computer readable medium of claim 9, wherein said first player's gameplay is associated with a gaming process, and said gaming process is configured to run to enable said second user or other users to join or discontinue said joining over time.
15. The non-transitory computer readable medium of claim 14, wherein an interface is provided by the cloud gaming server to enable access to one or more other video games or game functions, said cloud gaming server enabling navigation to the interface and said video game while the gaming process remains available to continue gameplay by said first user, second user, or one or more users that join the multiplayer gameplay of the video game.
16. The non-transitory computer readable medium of claim 14, wherein during said gaming process control of the first user's gameplay is transitioned from the first controller device to the second controller device, or from the second controller device to the first controller device.
17. A system comprising at least one computing device, said at least one computing device configured to perform a method including the following method operations:
- executing, by a cloud gaming server, a first user's gameplay of a video game, wherein executing the first user's gameplay includes processing input received over a network from a first controller device that is associated with the first user to direct gameplay events occurring in the first user's gameplay;
- presenting, over the network, a live video feed of the first user's gameplay to a remote second user;
- processing a request for the second user to join the first user's gameplay;
- initiating gameplay by the second user in the first user's gameplay, wherein initiating gameplay by the second user includes processing input received over the network from a second controller device that is associated with the second user to direct at least some of the gameplay events occurring in the first user's gameplay.
18. The system of claim 17, wherein initiating gameplay by the second user in the first user's gameplay includes initiating a multiplayer mode of the video game that facilitates multiplayer gameplay by the first user and the second user.
19. The system of claim 18, wherein the multiplayer mode includes continuing the processing of the input received from the first controller device while actively processing the input received from the second controller device.
20. The system of claim 17, wherein the multiplayer gameplay is configured to enable one or more additional players to join based on availability or priority ordering of said one or more additional players.
21. The system of claim 17, wherein one or more additional players are enabled to join the first user's gameplay or discontinue said joining.
22. The system of claim 17, wherein said first player's gameplay is associated with a gaming process, and said gaming process is configured to run to enable said second user or other users to join or discontinue said joining over time.
23. The system of claim 22, wherein an interface is provided by the cloud gaming server to enable access to one or more other video games or game functions, said cloud gaming server enabling navigation to the interface and said video game while the gaming process remains available to continue gameplay by said first user, second user, or one or more users that join the multiplayer gameplay of the video game.
24. The system of claim 22, wherein during said gaming process control of the first user's gameplay is transitioned from the first controller device to the second controller device, or from the second controller device to the first controller device.
5269687 | December 14, 1993 | Mott et al. |
5354202 | October 11, 1994 | Moncrief et al. |
6488505 | December 3, 2002 | Hightower |
6699127 | March 2, 2004 | Lobb et al. |
8043156 | October 25, 2011 | Ackley et al. |
8069125 | November 29, 2011 | Jung et al. |
8221220 | July 17, 2012 | Ackley et al. |
8475269 | July 2, 2013 | Okada |
8672744 | March 18, 2014 | Gronkowski et al. |
8672765 | March 18, 2014 | Gary |
8843557 | September 23, 2014 | Ranade |
9111080 | August 18, 2015 | Dawson |
9352226 | May 31, 2016 | Miura et al. |
20020019258 | February 14, 2002 | Kim et al. |
20020124053 | September 5, 2002 | Adams |
20040148221 | July 29, 2004 | Chu |
20040224740 | November 11, 2004 | Ball et al. |
20040224741 | November 11, 2004 | Jen et al. |
20040225386 | November 11, 2004 | Thompson et al. |
20050277455 | December 15, 2005 | Chudley et al. |
20060038890 | February 23, 2006 | Macintosh et al. |
20060121971 | June 8, 2006 | Slomiany et al. |
20060135236 | June 22, 2006 | Sato et al. |
20060287029 | December 21, 2006 | Yoshinobu et al. |
20070060359 | March 15, 2007 | Smith |
20070078989 | April 5, 2007 | van Datta et al. |
20070117635 | May 24, 2007 | Spanton et al. |
20070294089 | December 20, 2007 | Garbow et al. |
20080004115 | January 3, 2008 | Chatani |
20080039203 | February 14, 2008 | Ackley et al. |
20080039204 | February 14, 2008 | Ackley et al. |
20080076546 | March 27, 2008 | Moyle et al. |
20080119286 | May 22, 2008 | Brunstetter |
20080125226 | May 29, 2008 | Emmerson |
20080139301 | June 12, 2008 | Holthe |
20080167122 | July 10, 2008 | Maeda et al. |
20080193099 | August 14, 2008 | Nakai et al. |
20090005141 | January 1, 2009 | Lehtiniemi et al. |
20090036220 | February 5, 2009 | Nagashima |
20090089830 | April 2, 2009 | Chandratillake et al. |
20090131177 | May 21, 2009 | Pearce |
20090149246 | June 11, 2009 | Opaluch |
20090165140 | June 25, 2009 | Robinson et al. |
20090208181 | August 20, 2009 | Cottrell |
20090209335 | August 20, 2009 | Pearce |
20090235150 | September 17, 2009 | Berry |
20090305788 | December 10, 2009 | Bronstein |
20100056269 | March 4, 2010 | Zalewski |
20100160038 | June 24, 2010 | Youm et al. |
20100203933 | August 12, 2010 | Eyzaguirre et al. |
20100235329 | September 16, 2010 | Koren et al. |
20100240459 | September 23, 2010 | Shelton |
20100292011 | November 18, 2010 | Kira et al. |
20110041153 | February 17, 2011 | Simon et al. |
20110098108 | April 28, 2011 | Kuper et al. |
20110126110 | May 26, 2011 | Vilke |
20110250971 | October 13, 2011 | Van Os et al. |
20110269548 | November 3, 2011 | Barclay et al. |
20120004040 | January 5, 2012 | Pereira et al. |
20120059699 | March 8, 2012 | Zhou et al. |
20120134651 | May 31, 2012 | Cottrell |
20120185066 | July 19, 2012 | Kern et al. |
20120220370 | August 30, 2012 | Layne, IV et al. |
20120262463 | October 18, 2012 | Bakalash et al. |
20120270647 | October 25, 2012 | Zalewski |
20120278439 | November 1, 2012 | Ahiska et al. |
20120283011 | November 8, 2012 | Van Luchene |
20120295718 | November 22, 2012 | Paquet et al. |
20120302351 | November 29, 2012 | Murphy et al. |
20120323898 | December 20, 2012 | Kumar et al. |
20130023343 | January 24, 2013 | Schmidt |
20130040730 | February 14, 2013 | Barclay et al. |
20130072299 | March 21, 2013 | Lee |
20130072308 | March 21, 2013 | Peck et al. |
20130084969 | April 4, 2013 | Knoles et al. |
20130176481 | July 11, 2013 | Holmes et al. |
20130190096 | July 25, 2013 | Ronen et al. |
20130225290 | August 29, 2013 | Zhou et al. |
20130244785 | September 19, 2013 | Gary |
20130244789 | September 19, 2013 | Gary |
20130244790 | September 19, 2013 | Gary |
20130260896 | October 3, 2013 | Miura et al. |
20140038721 | February 6, 2014 | Archer |
20140066177 | March 6, 2014 | Zalewski |
20140066191 | March 6, 2014 | Yang |
20140094314 | April 3, 2014 | Watson et al. |
20140179426 | June 26, 2014 | Perry et al. |
20140187323 | July 3, 2014 | Perry |
20140229866 | August 14, 2014 | Gottlieb |
101661529 | March 2010 | CN |
101847175 | September 2010 | CN |
101981589 | February 2011 | CN |
102171996 | August 2011 | CN |
102209120 | October 2011 | CN |
102450031 | May 2012 | CN |
102571735 | July 2012 | CN |
102687146 | September 2012 | CN |
103310098 | September 2013 | CN |
103336883 | October 2013 | CN |
103366083 | October 2013 | CN |
2278552 | January 2011 | EP |
2004167172 | June 2004 | JP |
2008194314 | August 2008 | JP |
2010178010 | August 2010 | JP |
2011072735 | April 2011 | JP |
2012038042 | February 2012 | JP |
20060100696 | September 2006 | KR |
20110116782 | October 2011 | KR |
20130105469 | September 2013 | KR |
20130105470 | September 2013 | KR |
20130105471 | September 2013 | KR |
201402183 | January 2014 | TW |
201404443 | February 2014 | TW |
201406431 | February 2014 | TW |
I547302 | September 2016 | TW |
WO 2006098606 | September 2006 | WO |
WO 2009020869 | February 2009 | WO |
WO 2011051091 | May 2011 | WO |
WO 2013138163 | September 2013 | WO |
WO 2013138165 | September 2013 | WO |
WO 2013138166 | September 2013 | WO |
2009073832 | June 2019 | WO |
- PCT International Search Report and Written Opinion in Application No. PCT/US2013/029704, dated Mar. 29, 2013 (5 total pages).
- PCT International Search Report and Written Opinion in Application No. PCT/US2013/029727, dated Mar. 29, 2013 (5 total paegs).
- PCT International Search Report and Written Opinion for PCT Application No. PCT/US2013/29719 dated May 13, 2013 (8 total pages).
- PCT Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority from International Application No. PCT/US2013/077320, dated Jul. 14, 2014.
- Gamasutra, “Hard Drivin', Hard Bargainin': Investigating Midway's ‘Ghost Racer’ Patent,” http://www.gamasutra.com/view/feature/129914/hard_drivin_hard_bargainin_.php, Feb. 27, 2015 (3 pages).
- PCT Notification Concerning Transmittal of International Preliminary Report on Patentability and Written Opinion in Application No. PCT/US2013/077320, dated Jul. 2, 2015 (8 total pages).
- Extended European Search Report issued in European Patent Application No. 13199209.1, dated Feb. 23, 2017 (15 pages).
- Office Action issued in related Chinese Patent App. No. 201310717878.2, dated Aug. 28, 2017 (23 total pages).
- First Office Action issued in related Chinese Patent App. No. 201310717261.0, dated Oct. 11, 2017 (19 total pages).
- First Office Action issued in Chinese Patent App. No. 201310717877.8, dated Oct. 20, 2017 (35 total pages).
- Fourth Substantive Examination Requirement issued in related Mexican Patent App. No. MX/a/2014/000227, dated Oct. 24, 2017 (4 total pages.
- Communication pursuant to Article 94(3) EPC, issued in related European Patent App. No. 13199209.1, dated Oct. 30, 2017 (5 total pages).
- Notification of Reason(s) for Refusal issued in related Japanese Patent App. No. 2017-157465, dated Oct. 31, 2017.
- Second Office Action issued in Chinese Patent App. No. 201310717878.2, dated Apr. 17, 2018 (20 total pages).
- Final Notification of Reason(s) for Refusal issued in related Japanese Patent Application No. 2017-157465, dated May 15, 2018 (7 total pages).
- Second Office Action issued in related Chinese Patent App. No. 201310717261.0, dated Jul. 9, 2018, with Eng. translation (22 total pages).
- First Office Action issued in related Korean Patent Application No. 10-2017-7014398, dated Sep. 12, 2018 (11 total pages).
- Third Office Action issued in related Chinese Patent Application No. 201310717877.8, dated Jan. 4, 2019 (7 total pages).
- Third Office Action issued in related Chinese Patent Application No. 201310717878.2, dated Jan. 17, 2019 (8 total pages).
- Rejection Decision issued in related Chinese Patent Application No. 201310717261.0, dated Mar. 1, 2019 (18 total pages).
- Final Office Action issued in related Korean Patent Application No. 10-2017-7014398, dated Mar. 21, 2019 (8 total pages).
- Extended European Search Report issued in related European Patent Application No. 19 15 0410.9, dated May 27, 2019 (8 pages).
- Office Action issued in related Korean Patent Application No. 10-2019-7021762, dated Sep. 17, 2019 (5 pgs).
- Office Action issued in related Korean Patent Application No. 10-2019-7014565, dated Sep. 17, 2019 (4 pgs).
- Anonymous: “Civilization IV—Wikipedia,” Dec. 18, 2012, XP055728021, Retrieved from the Internet: URL:https://en.wikipedia.org/w/index.php?title=civilization_IV&oldid=528568320 (12 pages).
- Decision to Refuse a European Patent Application issued in European Divisional Patent Application No. 19150410.9-1210, dated Oct. 5, 2020 (42 pages).
Type: Grant
Filed: Nov 26, 2019
Date of Patent: May 25, 2021
Patent Publication Number: 20200094150
Assignee: Sony Interactive Entertainment LLC (San Mateo, CA)
Inventors: Victor Octav Suba Miura (Foster City, CA), David Perry (Monarch Beach, CA), Philippe Dias (Mission Viejo, CA), Kelvin Yong (Irvine, CA), Nathan Gary (Santa Monica, CA)
Primary Examiner: Michael A Cuff
Application Number: 16/696,964
International Classification: A63F 9/24 (20060101); A63F 13/86 (20140101); H04N 21/478 (20110101); A63F 13/497 (20140101); A63F 13/533 (20140101); A63F 13/87 (20140101); A63F 13/795 (20140101); H04N 21/4788 (20110101); A63F 13/47 (20140101); A63F 13/35 (20140101);