Patent classifications
A63F13/358
Methods and systems for rendering and encoding content for online interactive gaming sessions
This application is directed to a method of managing processing capability of a server system having one or more processing cores that further include multiple processing slices. Upon receiving requests to initiate online gaming sessions, the server system allocates each processing slice of the processing cores to a subset of the online gaming sessions to be executed thereon. A first processing slice is allocated to a first subset of the online gaming sessions including a first gaming session and a second gaming session. At the first processing slice, a time-sharing processing schedule is determined for the first subset of the online gaming sessions. In accordance with the time-sharing processing schedule, the first and second gaming sessions share a duty cycle of the first processing slice, and are executed dynamically and in parallel according to real-time data processing need of the first and second gaming sessions.
Methods and systems for rendering and encoding content for online interactive gaming sessions
This application is directed to a method of managing processing capability of a server system having one or more processing cores that further include multiple processing slices. Upon receiving requests to initiate online gaming sessions, the server system allocates each processing slice of the processing cores to a subset of the online gaming sessions to be executed thereon. A first processing slice is allocated to a first subset of the online gaming sessions including a first gaming session and a second gaming session. At the first processing slice, a time-sharing processing schedule is determined for the first subset of the online gaming sessions. In accordance with the time-sharing processing schedule, the first and second gaming sessions share a duty cycle of the first processing slice, and are executed dynamically and in parallel according to real-time data processing need of the first and second gaming sessions.
Dynamic client buffering and usage of received video frames for cloud gaming
A method is disclosed including setting, at a server, a server VSYNC signal to a server VSYNC frequency defining a plurality of frame periods. The server VSYNC signal corresponding to generation of a plurality of video frames at the server during the plurality of frame periods. The method including setting, at a client, a client VSYNC signal to a client VSYNC frequency. The method including sending a plurality of compressed video frames based on the plurality of video frames from the server to the client over a network using the server VSYNC signal. The method including decoding and displaying, at the client, the plurality of compressed video frames. The method including analyzing the timing of one or more client operations to set the amount of frame buffering used by the client, as the client receives the plurality of compressed video frames.
Dynamic client buffering and usage of received video frames for cloud gaming
A method is disclosed including setting, at a server, a server VSYNC signal to a server VSYNC frequency defining a plurality of frame periods. The server VSYNC signal corresponding to generation of a plurality of video frames at the server during the plurality of frame periods. The method including setting, at a client, a client VSYNC signal to a client VSYNC frequency. The method including sending a plurality of compressed video frames based on the plurality of video frames from the server to the client over a network using the server VSYNC signal. The method including decoding and displaying, at the client, the plurality of compressed video frames. The method including analyzing the timing of one or more client operations to set the amount of frame buffering used by the client, as the client receives the plurality of compressed video frames.
REDUCING LATENCY IN ANTICHEAT DATAFLOW
Systems and methods for generating and storing metrics are described herein. In particular, a game server may receive game activity data from one or more client devices connected to the game server via a network. The game server sends the activity data to an anticheat server that uses one or more nodes to each calculate a portion of the activity data to generate one or more metrics. The metrics may indicate whether the data is indicative of cheating behavior within a parallel reality game. The nodes add their respective generated metrics to a relational database using prepared insert statements. The anticheat server may instruct the game server to take action with respect to one or more client devices if the metrics indicate that cheating behavior was exhibited within the data.
REDUCING LATENCY IN ANTICHEAT DATAFLOW
Systems and methods for generating and storing metrics are described herein. In particular, a game server may receive game activity data from one or more client devices connected to the game server via a network. The game server sends the activity data to an anticheat server that uses one or more nodes to each calculate a portion of the activity data to generate one or more metrics. The metrics may indicate whether the data is indicative of cheating behavior within a parallel reality game. The nodes add their respective generated metrics to a relational database using prepared insert statements. The anticheat server may instruct the game server to take action with respect to one or more client devices if the metrics indicate that cheating behavior was exhibited within the data.
SERVER APPARATUS, TERMINAL APPARATUS, INFORMATION PROCESSING SYSTEM, AND INFORMATION PROCESSING METHOD
There is provided a technology capable of reducing the processing load on a server apparatus side in cloud rendering. A server apparatus according to the present technology includes a controller. The controller groups terminal apparatuses each having a viewing position within an identical segment on the basis of viewing position information of each terminal apparatus within a viewing region including a plurality of segments, and transmits common video information to each of the grouped terminal apparatuses by multicasting.
Systems and methods for reducing hops associated with a head mounted system
Systems and methods for reducing hops associated with a head mounted display are described. The head mounted display includes a communications circuit for receiving and transmitting interactive media associated with a game program via a network. The interactive media is processed by the game cloud system and streamed directly to the communications circuit of the head mounted display. The head mounted display further includes a user input circuit for receiving an action from a user to generate an input, which includes position and motion detected by the user input circuit. The head mounted display includes a game processing circuit for decoding the interactive media received from the network. The game processing circuit drives a portion of interactivity associated with the game program. The portion of interactivity is generated based on the input.
PREDICTIVE EXECUTION OF DISTRIBUTED GAME ENGINES
Systems described herein may automatically and dynamically adjust the amount and type of computing resources usable to execute, process, or perform various tasks associated with a video game. Using one or more machine learning algorithms, a prediction model can be generated that uses the historical and/or current user interaction data obtained by monitoring the users playing the video game. Based on the historical and/or current user interaction data, future user interactions likely to be performed in the future can be predicted. Using the predictions of the users' future interactions, the amount and type of computing resources maintained in the systems can be adjusted such that a proper balance between reducing the consumption of computing resources and reducing the latency experienced by the users of the video game is achieved and maintained.
PREDICTIVE EXECUTION OF DISTRIBUTED GAME ENGINES
Systems described herein may automatically and dynamically adjust the amount and type of computing resources usable to execute, process, or perform various tasks associated with a video game. Using one or more machine learning algorithms, a prediction model can be generated that uses the historical and/or current user interaction data obtained by monitoring the users playing the video game. Based on the historical and/or current user interaction data, future user interactions likely to be performed in the future can be predicted. Using the predictions of the users' future interactions, the amount and type of computing resources maintained in the systems can be adjusted such that a proper balance between reducing the consumption of computing resources and reducing the latency experienced by the users of the video game is achieved and maintained.