SYSTEM AND METHODS THEREOF FOR DISPLAYING VIDEO CONTENT

A system is configured to display a video clip on a display of a user device. The system receives a request from the user device to display the video clip. The system then queries an advertiser server to identify the video clip. The system fetches the identified video clip. The system then extracts from the fetched video clip its video content. The video content is a plurality of frames. The system then determines a type of the user device. Then, the system determines a number of frames per second for the user device. The system generates a plurality of images from the plurality of frames. The system further generates a timing sequence for displaying each of the plurality of images of the imagized video clip on the display of the user device. The system then sends the imagized video clip and the timing sequence to the user device.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation-in-part of and claims the benefit of PCT Application No. PCT/IL2014/050380 filed on Apr. 24, 2014 that claims the benefit of U.S. Provisional Application No. 61/833,025 filed on Jun. 10, 2013, the contents of which are herein incorporated by reference for all that it contains.

This application is a continuation-in-part of and claims the benefit of PCT Application No. PCT/IL2014/051053 filed on Dec. 4, 2014 that claims the benefit of U.S. Provisional Application No. 62/006,179 filed on Jun. 1, 2014, the contents of which are herein incorporated by reference for all that it contains.

This application claims the benefit of U.S. Provisional Application No. 62/157,470 filed on May 6, 2015, the contents of which are herein incorporated by reference for all that it contains.

TECHNICAL FIELD

The disclosure generally relates to systems for playing video content, and more specifically to systems and methods for converting video content to imagized video content and displaying the imagized video content on user devices.

BACKGROUND

The Internet, also referred to as the worldwide web (WWW), has become a mass media where the content presentation is largely supported by paid advertisements that are added to web-page content. Typically, advertisements displayed in a web-page contain video elements that are intended for display on the user's display device.

Mobile devices such as smartphones are equipped with mobile browsers through which users access the web. Such mobile browsers typically cannot display auto-played video clips on mobile web pages as the mobile HTML5 video component does not allow autoplay and requires user interaction such as clicking on the page in order to start the video play. The term autoplay refers to starting playing a video on an HTML page when the page is loaded without requiring a user interaction such as clicking on the page. Furthermore, there are multiple video formats supported by different phone manufactures which makes it difficult for the advertisers to know which phone the user has, and what video format to broadcast it with.

BRIEF DESCRIPTION OF THE DRAWINGS

The subject matter disclosed herein is particularly pointed out and distinctly claimed in the claims at the conclusion of the specification. The foregoing and other objects, features and advantages of the disclosure will be apparent from the following detailed description taken in conjunction with the accompanying drawings.

FIG. 1—is a system for generating imagized video clip streams respective of video content according to an embodiment.

FIG. 2—is a flowchart of the operation of a system for generating imagized video clips respective of video content according to an embodiment.

FIG. 3—is a flowchart of the operation of a system for displaying imagized video content according to an embodiment.

FIG. 4—is a schematic diagram of the operation of a system for displaying imagized video content according to an embodiment.

DETAILED DESCRIPTION

It is important to note that the embodiments disclosed herein are only examples of the many advantageous uses of the innovative teachings herein. In general, statements made in the specification of the present application do not necessarily limit any of the various claimed disclosures. Moreover, some statements may apply to some inventive features but not to others. In general, unless otherwise indicated, singular elements may be in plural and vice versa with no loss of generality. In the drawings, like numerals refer to like parts through several views.

The embodiments disclosed herein are only examples of the many possible advantageous uses and implementations of the innovative teachings presented herein.

A system is configured to generate an imagized video clip respective of a video clip (see FIG. 4, element 415). The system receives electronically (see FIG. 4, element 420) at least one video clip. The system analyzes the video clip (see FIG. 4, generation of frames 425 that comprise the video clip) and generates (see FIG. 4, element 430) a plurality of images (see FIG. 4, images 435) respective thereto as well as a timing sequence file (see FIG. 4, element 437). The system is then configured to generate an imagized video clip respective of the plurality of images and the timing sequence file. The system is further configured to display the imagized video clip in a sequence at certain timing, thereby providing an output of a video clip respective of the imagized video clip.

In order to enable a rich interactive in-stream ad experience, a universal specification was developed for interaction between ad units and video players focused on the technology named video player ad-serving interface definition (VPAID)VPAID was designed to work across multiple platforms, many of which include various mobile devices. VPAID is designed to work with a standard HTMLS video player and as a result, VPAID interactions are limited in mobile devices where HTML5 video player functionality is limited. HTML5 video player limitations include for example: lack of auto-play of video content where a user must click on the video content in order to play it, video play in IOS® operating system is done through the native video player and does not allow play through a web page, HTML5 video player is not designed to enable autoplay of video content, and more. By converting the video content (e.g., video frames) to an imagized video clip as further described herein below, and by providing VPAID interface with an emulated video player instead of the original HTML5 video player, such features are enabled and the difficulties associated with displaying VPAID on mobile devices is overcome.

In order to initialize the VPAID interface with an emulated video player that supports such features, an emulated video player is instantiated in a way that it includes all required HTML5 video player methods, properties and events and therefore emulates all HTML5 video player functionalities.

When initializing VPAID on a user device, it is required to send a request to, for example, an initAd( ) method from a data source 140. In ‘environmentVars’ object, a reference to ‘videoSlot’ representing the HTML5 video player should be initiated. According to this disclosure, instead of passing an instance of the HTML5 video object, in this embodiment an initiation of the emulated video player is made by the server 130 where the emulated video player supports all required HTML5 methods, properties and events thereby allowing VPAID interface to render the video through the emulated video player instead of the HTML5 video player. It should be clear for one of ordinary skill in the art that other methods, properties and/or events may be used for initialization as described hereinabove.

In other embodiments described herein, other standards may be used in order to play the video content. However, in all such cases HTML5 video elements are used in order to play the video. In some cases, the HTML5 video object is passed as a parameter and in such cases the emulated video player is launched instead. In other cases, the HTML5 video element may be created dynamically, for example by calling document createElement (“video”) or by other methods known to a person of ordinary skill in the art. In these cases the function used to create or locate the HTML5 video element is emulated and when identifying the call to create the element, the emulated video player is created and returned instead of the HTML5 video element.

FIG. 1 depicts an exemplary and non-limiting diagram of a system 100 for generating imagized video clips respective of video content according to an embodiment. The system 100 comprises a network 110 the enables communications between various portions of the system 100. The network may comprise the likes of busses, local area network (LAN), wide area network (WAN), metro area network (MAN), the worldwide web (WWW), the Internet, as well as a variety of other communication networks, whether wired or wireless, and in any combination, that enable the transfer of data between the different elements of the system 100. The system 100 further comprises a user device 120 connected to the network 110. The user device 120 may be, for example but without limitations, a smart phone, a mobile phone, a laptop, a tablet computer, a wearable computing device, a personal computer (PC), a smart television and the like. The user device 120 comprises a display unit 125 such as a screen, a touch screen, etc.

A server 130 is further connected to the network 110. The server 130 is configured to receive a request to display at least one video content item on the display unit 125 of the user device 120. The video content item comprises a plurality of frames that create the video content item. The request is received over the network 110. According to one embodiment, the request may be received as a user's gesture over the display unit 125 of the user device 120. According to another embodiment, the request may be received through an application program installed on the user device 120.

The system 100 further comprises one or more data sources 140-1 through 140-N (collectively referred hereinafter as data sources 140 or individually as a data source 140, merely for simplicity purposes), where N is an integer equal to or greater than 1. The data sources 140 are accessible through the network 110 and configured to store video content. The data sources 140 may be operative by one or more advertiser servers 160.

Respective of the request for the at least one video content, the server 130 is configured to query an advertiser server 160 to identify the video content. Respective of an answer to the query received from the advertiser server 160, the at least one video content item is fetched. The fetching may be made from a data source, for example the data source 140-1. The query may include additional metadata that assists in the identification of the at least one video content and thereby assists in the fetching of the at least one video content from the data source 140-1. The server 130 then extracts a plurality of frames from the video content. The amount of frames is determined by the server 130 based on, for example, determination of a type of the user device 120 requesting the video content. As an example, in case the user device 120 is an iPhone® smartphone, a certain amount of frames may be extracted while in case the user device 120 is an Android® smartphone, the amount of frames extracted may be different. According to another embodiment, the amount of frames extracted is determined respective of an analysis of the video content. The number of frames per second may be smaller than the number of frames per second of the video clip. The analysis enables the server 130 to determine a minimum amount of frames required for an optimal performance of display.

The server 130 is then configured to generate a plurality of images respective of the extracted frames wherein at least one of the plurality of images is created from at least two different frames of the extracted plurality of frames. Prior art solutions introduced the idea of Animated GIF files where a number of “frames” are put together in a single file that could be played back at an interframe duration set in a file. Additionally, each frame could be used to update only a portion of the whole image to help in compression. This format, however, does not allow downloading images at a rate or bit encoding based on the observed network. Generating, as described herein, is the process of extracting existing frames and then saving them in a different format. Generating an image respective of a video content as further described herein, typically involves two or more frames composited to a single image. This generated image was never a part of the original video content, as opposed to frame images which are by definition a part of the original video content, regardless of the format they are stored in.

According to another embodiment, the server 130 is further configured to cluster the images to one or more streams of images for further processing by the server 130.

The server 130 is then configured to generate a timing sequence for displaying the plurality of images or streams of images as an imagized video clip on the display unit 125 of the user device 120. The timing sequence enables the display of the imagized video clip as a video content. According to one embodiment, the timing sequence is generated respective of the type of the user device 120 and/or the display unit 125. As a non-limiting example, upon determination by the server 130 that the user device 120 is a smart phone a timing sequence of ten images per second is determined while upon determination that the user device 120 is a PC, a timing sequence of 20 images per second is determined. In other embodiment the image sequence includes 15 images per 2 seconds.

The server 130 is then configured to send the imagized video clip and the timing sequence to the user device 120 for display on the display unit 125, in compliance with VPAID. Optionally, the system 100 further comprises a database 150. The database is configured to store data related to requests received from the user device 120, imagized video clips and the video content from which the imagized video clips were generated, etc.

FIG. 2 is an exemplary and non-limiting flowchart 200 of the operation of a system for generating image streams respective of video content according to an embodiment. In S210, the method starts when a request to display video content is received from a user device, for example, the user device 120.

In S220, the requested video content is fetched from a data source 140-1 through the network 110. In S230, a plurality of images are generated respective of the video content as further described hereinabove with respect of FIG. 1. In S240, a timing sequence for displaying the plurality of images as an imagized video clip on the display unit 125 of the user device 120 is generated. In S250, the imagized video clip and the timing sequence are sent for display on the display unit 125 of the user device 120. In S260, it is checked whether additional requests for video content are received from the user device 120 and if so, execution continues with S210; otherwise, execution terminates.

FIG. 3 is an exemplary and non-limiting flowchart 300 of the operation of a system for displaying video content according to an embodiment. In S310, the operation starts when a request to display a video content on a user device, for example, the user device 120 is received. The request may be received from, for example, an advertiser server 160 over the network 110. In S320, the advertiser server 160 is queried by the server 130 in order to identify the video content. In S330, the server 130 fetches the video content. The video content may be fetched from a data source of the one or more data sources 140 over the network 110. In S340, a plurality of frames are extracted respective of the video content. In S350, a type of the user device 120 is determined by the server 130. In S360, a plurality of images are generated respective of the extracted frames. In S370, a timing sequence for displaying each of the plurality of images as an imagized video clip on the display of the user device is generated. In S380, the imagized video clip is sent for display on a display of the user device 120. In S390, it is checked whether there additional requests and if so, execution continues with S310; otherwise, execution terminates.

FIG. 4 is an exemplary and non-limiting schematic diagram 400 of the operation of a system for displaying imagized video content according to an embodiment. An advertiser server 410 comprises at least one video clip 415 thereon. According to an embodiment, the system 100 receives 420 electronically a request from the user device 120 to display the video clip 415 on the display of the user device 120. The system 100 then analyzes the video clip 415 and generates a plurality of frames 425-1 through 425-n that comprise the video clip 415. The system 100 then generates 430 a plurality of images 435-1 through 435-m respective of the plurality of frames 425. Respective of the plurality of frames 425, the system is further configured to generate a timing sequence 437. Respective thereof, the system 100 can send over the network 110 the imagized video clip 435 and the timing sequence 437 to the user device 120.

It should be clear to one of ordinary skill in the art that the imagized video clip which is basically an emulated video tag of the video content, enables features which are typically unavailable while displaying video content on mobile user devices, such as, for example, automatic play (autoplay) of the video content, responsiveness (e.g., clickability), and more.

The principles of the disclosure, wherever applicable, are implemented as hardware, firmware, software or any combination thereof. Moreover, the software is preferably implemented as an application program tangibly embodied on a program storage unit or computer readable medium. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPUs”), a memory, and input/output interfaces. The computer platform may also include an operating system and microinstruction code. The various processes and functions described herein may be either part of the microinstruction code or part of the application program embodied in non-transitory computer readable medium, or any combination thereof, which may be executed by a CPU, whether or not such computer or processor is explicitly shown, to cause performance of operations described in this disclosure. Implementations may further include full or partial implementation as a cloud-based solution. In some embodiments certain portions of a system may use mobile devices of a variety of kinds. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.

All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the principles of the disclosure and the concepts contributed by the inventor to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions. Moreover, all statements herein reciting principles, aspects, and embodiments of the disclosure, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.

Claims

1. A computerized method for displaying a video clip on a display of a user device, the method comprising:

identifying a request over a network from the user device to display the video clip on the display;
fetching over the network the identified video clip;
extracting from the fetched video clip its video content, wherein the video content is a plurality of frames;
determining the type of the user device;
determining a number of frames per second for the user device, wherein the determination is made respective of the type of the user device;
generating a plurality of images from the plurality of frames in order to form an imagized video clip, wherein the plurality of images fits the number of frames per second determined for the user device;
generating a timing sequence for displaying each of the plurality of images of the imagized video clip on the display of the user device; and,
sending over the network the imagized video clip and the timing sequence to the user device.

2. The computerized method of claim 1, wherein the request to display the video clip on a display includes requests for at least one of: a VPAID request, a request for a software development kit (SDK), a request to display the video clip on the display of the user device, and a request to display the video clip of the user device using an emulated video player.

3. The computerized method of claim 1, wherein the number of frames per second for the user device is smaller than the number of frames per second of the video clip.

4. The computerized method of claim 1, wherein at least one of the plurality of images is created from at least two different frames of the extracted plurality of frames.

5. The computerized method of claim 1, further comprising:

displaying the imagized video clip in an emulated video player on the user device.

6. The computerized method of claim 5, wherein the display of the video clip is in compliance with at least one of: Video Player Ad-Serving Interface Definition (VPAID), and Mobile Rich Media Ad Interface Definitions (MRAID).

7. The computerized method of claim 1, further comprising:

enabling auto-play of the imagized video clip on the user device when a web page is loaded by the user device without user interaction.

8. The computerized method of claim 1, wherein determining the number of frames per second for the user device includes determining that the user device is of a first type of user device that is assigned a first number of frames per second, wherein a second type of user device is assigned a different, second number of frames per second to be extracted from video clips.

9. The computerized method of claim 1, wherein the following operations are performed by a server that is remote from the user device:

(i) the identifying the request,
(ii) the fetching the identified video clip,
(iii) the extracting the video content,
(iv) the determining the type of the user device,
(v) the determining the number of frames per second for the user device,
(vi) the generating the plurality of images,
(vii) the generating the timing sequence, and
(viii) the sending the imagized video clip and the timing sequence.

10. A computer system, comprising:

one or more processors; and
non-transitory computer readable medium including instructions that, when executed by the one or more processors, cause performance of operations that include: identifying a request over a network from the user device to display the video clip on the display;
fetching over the network the identified video clip;
extracting from the fetched video clip its video content, wherein the video content is a plurality of frames;
determining the type of the user device;
determining a number of frames per second for the user device, wherein the determination is made respective of the type of the user device;
generating a plurality of images from the plurality of frames in order to form an imagized video clip, wherein the plurality of images fits the number of frames per second determined for the user device;
generating a timing sequence for displaying each of the plurality of images of the imagized video clip on the display of the user device; and,
sending over the network the imagized video clip and the timing sequence to the user device.

11. The computerized method of claim 10, wherein the request to display the video clip on a display includes requests for at least one of: a VPAID request, a request for a software development kit (SDK), a request to display the video clip on the display of the user device, and a request to display the video clip of the user device using an emulated video player.

12. The computer system of claim 10, wherein the number of frames per second for the user device is smaller than the number of frames per second of the video clip.

13. The computer system of claim 10, wherein at least one of the plurality of images is created from at least two different frames of the extracted plurality of frames.

14. The computer system of claim 10, wherein the operations further comprise:

displaying the imagized video clip in an emulated video player on the user device.

15. The computer system of claim 14, wherein the display of the video clip is in compliance with at least one of: Video Player Ad-Serving Interface Definition (VPAID), and Mobile Rich Media Ad Interface Definitions (MRAID).

16. The computer system of claim 10, wherein the operations further comprise:

enabling auto-play of the imagized video clip on the user device when a web page is loaded by the user device without user interaction.

17. The computer system of claim 10, wherein determining the number of frames per second for the user device includes determining that the user device is of a first type of user device that is assigned a first number of frames per second, wherein a second type of user device is assigned a different, second number of frames per second to be extracted from video clips.

18. The computer system of claim 10, wherein the following operations are performed by a server that is remote from the user device:

(i) the identifying the request,
(ii) the fetching the identified video clip,
(iii) the extracting the video content,
(iv) the determining the type of the user device,
(v) the determining the number of frames per second for the user device,
(vi) the generating the plurality of images,
(vii) the generating the timing sequence, and
(viii) the sending the imagized video clip and the timing sequence.
Patent History
Publication number: 20180192121
Type: Application
Filed: Feb 27, 2018
Publication Date: Jul 5, 2018
Inventor: Tal Melenboim (Ashdod)
Application Number: 15/905,923
Classifications
International Classification: H04N 21/435 (20060101); H04N 21/81 (20060101); H04N 21/239 (20060101); H04N 21/2343 (20060101); H04N 21/258 (20060101); H04N 21/472 (20060101);