Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
METHOD AND SYSTEM FOR DISPLAYING INTERACTIVE QUESTIONS DURING STREAMING OF REAL-TIME AND ADAPTIVELY ASSEMBLED VIDEO
Document Type and Number:
WIPO Patent Application WO/2018/011688
Kind Code:
A1
Abstract:
The present disclosure provides a system and method for enabling display of interactive questions during streaming of a real time, dynamic, adaptive and non-sequentially assembled video. The method includes reception of a set of preference data associated with a user and a set of user authentication data. In addition, the method includes serving of the assembled video based on the received set of preference data to the user in real time. Moreover, the method includes fetching of a pre-defined set of interactive questions. Further, the method includes posting of the pre-defined set of interactive questions during the serving of the assembled video. Furthermore, the method includes collection of the set of user feedbacks for a posted pre-defined set of interactive questions from the user.

Inventors:
VENKATRAMAN N DILIP (IN)
DILIP SAVITRI (IN)
Application Number:
PCT/IB2017/054119
Publication Date:
January 18, 2018
Filing Date:
July 07, 2017
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
VIDEOTAP PTE LTD (SG)
International Classes:
G06F17/30
Foreign References:
US20080214162A12008-09-04
US20130144871A12013-06-06
US20150324379A12015-11-12
Other References:
"Networking Bible", 15 September 2009, Indianapolis, Indiana, USA, ISBN: 978-0-470-43131-3, article BARRIE SOSINSKY: "Streaming Media", pages: 639 - 660, XP055406663
"YouTube for Dummies", 30 July 2007, WILEY PUBLISHING, INC., Indianapolis, IN, US, ISBN: 978-0-470-14925-6, article DOUG SAHLIN ET AL: "Searching the tube for your One-in-a-Million Video", pages: 71 - 92, XP055406661
Download PDF:
Claims:
CLAIMS

What is claimed is:

1. A method for enabling display of interactive questions during streaming of a real time, dynamic, adaptive and non-sequentially assembled video, the method comprising:

receiving a set of preference data associated with a user from a pre-defined selection criteria corresponding to a digitally processed repository of videos and a set of user authentication data;

serving an assembled video corresponding to the received set of preference data to the user in real time; fetching a pre-defined set of interactive questions, wherein the pre-defined set of interactive question being fetched for obtaining a set of user feedbacks;

posting the pre-defined set of interactive questions during the serving of the assembled video, wherein the pre-defined set of interactive questions being posted in a first pre-defined order of preference, wherein the pre-defined set of interactive questions being posted based on a semantic context information mined from the set of preference data of the user and one or more tagged videos, a real time viewing and selection behavior of the user, a request of the user for a series of questions, dynamic requests of posting questions from one or more publishers and wherein each interactive question of the pre-defined set of interactive questions being posted in at least one of one or more pre-defined regions; and

collecting the set of user feedbacks corresponding to the posted pre-defined set of interactive questions from the user, wherein the set of user feedbacks being collected in the real time.

2. The method as recited in claim 1, further comprising pushing the pre-defined set of interactive questions from at least one of the one or more publishers, one or more system administrators, one or more third party content managers and one or more sponsors, wherein the pre-defined set of interactive questions being pushed in the real time.

3. The method as recited in claim 1, further comprising extracting the one or more tagged videos related to the set of preference data of the user from the digitally processed repository of videos, wherein the one or more tagged videos being fetched based on a correlation of a set of tags associated with each video of the one or more tagged videos with the set of preference data associated with the user.

4. The method as recited in claim 1 , further comprising fragmenting each tagged video of the one or more tagged videos into one or more tagged fragments, wherein each tagged video being fragmented into the one or more tagged fragments, wherein each tagged fragment being characterized by a pre-determined interval of time and wherein each tagged video being fragmented based on segmentation of the tagged video for each predetermined interval of time.

5. The method as recited in claim 1 , further comprising segregating one or more mapped fragments of the one or more tagged fragments into one or more logical sets of mapped fragments, wherein the one or more mapped fragments being segregated based on a positive mapping of keywords from the set of preference data with the set of tags associated with each tagged fragment of the one or more tagged fragments.

6. The method as recited in claim 1, further comprising mining the semantic context information from each mapped fragment of one or more mapped fragments and each logical set of mapped fragments of the one or more logical sets of mapped fragments, wherein the semantic context information comprises an object specific context information and scene specific context information of each mapped fragment and each logical set of mapped fragments.

7. The method as recited in claim 1, further comprising clustering the one or more logical sets of mapped fragments into corresponding one or more logical clusters of mapped fragments.

8. The method as recited in claim 1, further comprising assembling at least one of the one or more logical clusters of mapped fragments in a second pre-defined order of preference to obtain the assembled video, wherein each logical cluster of mapped fragments being clustered based on analysis of the set of preference data of the user and the semantic context information, wherein the second pre-defined order of preference being derived from the set of preference data and a user profile.

9. The method as recited in claim 1, further comprising creating the user profile based on the set of user authentication data and the set of preference data, wherein the user profile comprises the set of preference data segregated on a basis of the pre-defined selection criteria, the set of user authentication data, a past set of preference data, a physical location of the user and a bio data of the user and wherein the set of user authentication data comprises an email address, an authentication key, a physical location and a time of request of video.

10. The method as recited in claim 1, further comprising transcoding the assembled video into a predefined video format by utilizing a codec, wherein the assembled video being transcoded to enable adaptive bitrate streaming based on one or more device parameters and one or more network parameters, wherein the one or more device parameters comprises screen size, screen resolution and pixel density and wherein the one or more network parameters comprises an IP address, network bandwidth, maximum bitrate support over network, throughput, connection strength and location of requesting server.

Description:
METHOD AND SYSTEM FOR DISPLAYING INTERACTIVE QUESTIONS DURING STREAMING OF REAL-TIME AND ADAPTIVELY ASSEMBLED VIDEO

TECHNICAL FIELD

[0001] The present disclosure relates to a field of online video serving. More specifically, the present disclosure relates to a method and system for enabling display of interactive question during streaming of assembled video.

BACKGROUND

[0002] With the advent of online multimedia revolution along with sudden rise in network bandwidth in recent years, the popularity of online video on demand platforms has gained momentum. These video on demand platforms provide a plethora of online streaming services. These services include television news, sports shows, television shows, non-televised shows, interviews, location specific events, national events, international events, movies and the like. The videos are arranged in different categories with different tags for complete video. Nowadays, there are many platforms that provide video streaming services on demand. These platforms stream videos based on mapping user selection or search query with tags on a complete video. These video on demand streaming platforms do not efficiently address the dynamically changing user interests.

SUMMARY

[0003] In one aspect, the present disclosure provides a method for enabling display of interactive questions during streaming of a real time, dynamic, adaptive and non-sequentially assembled video. The method includes a step of receiving a set of preference data associated with a digitally processed repository of videos and a set of user authentication data. The set of preference data is selected from pre-defined selection criteria. In addition, the method includes another step of serving an assembled video related to the received set of preference data to the user in real time. The method includes yet another step of fetching a pre-defined set of interactive questions. The method includes yet another step of posting the pre-defined set of interactive questions during the serving of the assembled video. Furthermore, the method includes yet another step of collecting the set of user feedbacks of a posted pre-defined set of interactive questions from the user. The pre-defined set of interactive question is fetched to obtain a set of user feedbacks. The pre-defined set of interactive questions is posted in a first predefined order of preference. The pre-defined set of interactive questions is posted based on the semantic context information mined from the set of preference data of the user and the one or more tagged videos. In addition, the pre-defined set of interactive questions may be posted based on a real time viewing and selection behavior of the user, a request of the user for a series of questions and dynamic requests of posting questions from one or more publishers. In addition, each interactive question of the pre-defined set of interactive questions is posted in at least one of one or more pre-defined regions of a user interface. The collection of the set of user feedbacks is done in the real time.

BRIEF DESCRIPTION OF THE FIGURES

[0004] FIG. 1A illustrates an interaction of a user and one or more publishers with an interactive video system, in accordance with an embodiments of the present disclosure;

[0005] FIG. IB illustrates the interaction of the user with the interactive video system, in accordance with another embodiment of the present disclosure;

[0006] FIG. 1C illustrates the interaction of the one or more publishers with the interactive video system, in accordance with yet another embodiment of the present disclosure;

[0007] FIG. ID illustrates the interaction between the user and the one or more publishers associated with the interactive video system, in accordance with yet another embodiment of the present disclosure;

[0008] FIG. IE illustrates the interaction between a live media server and the interactive video system, in accordance with yet another embodiment of the present disclosure;

[0009] FIG. IF illustrates the interaction between a question management system, a question database and the user, in accordance with yet another embodiment of the present disclosure;

[0010] FIG. 2A illustrates an example of a web platform for posting interactive questions during streaming of a real time, dynamic, adaptive and non-sequentially assembled video;

[0011] FIG. 2B illustrates the example of a real time, dynamic, adaptive and non-sequential assembling of one or more tagged clips corresponding to one or more videos;

[0012] FIG. 3 illustrates a block diagram of the interactive video system, in accordance with various embodiments of the present disclosure;

[0013] FIG. 4 illustrates a flow chart for enabling display of the interactive questions during the streaming of the real time, dynamic, adaptive and non-sequentially assembled video, in accordance with various embodiments of the present disclosure; and

[0014] FIG. 5 illustrates a block diagram of a communication device, in accordance with various embodiments of the present disclosure.

DETAILED DESCRIPTION

[0015] FIG. 1A illustrates an interaction of a user 102a and one or more publishers 106 with an interactive video system 112, in accordance with an embodiments of the present disclosure. The interactive video system 112 enables a pre-defined set of interactive questions posted during streaming of an assembled video. The interactive video system 112 interacts by invoking a set of user feedbacks from the user 102a in real time. The interactive video system 112 collects the set of user feedbacks to improve and add more user centric customization in video on demand service. Each interactive question is framed to collect statistics associated with video on demand service. In addition, the pre-defined set of interactive questions help in analysis of user interest behavior and user selection behavior. The interactive video system 112 may use machine learning to inform administrators about real time requirements of the user 102a. Accordingly, the interactive video system 112 may customize user interface and video assembling based on extrapolation of the set of user feedbacks. In an embodiment of the present disclosure, the user 102a and each of one or more publishers 106 interact with the interactive video system 112 based on a pay per view model. In another embodiment of the present disclosure, the user 102a and each of the one or more publishers interact with the interactive video system 112 based on a subscription based model. The above interaction of the user 102a and the one or more publishers 106 is part of an interactive environment. The interactive environment includes a communication device 102, a communication network 104, one or more communication devices 102, the one or more publishers 106 and a main server 110.

[0016] The user 102a is associated with the one or more communication devices 102. Each of the one or more communication devices 102 may be any suitable device with at least a display, a storage unit and network connectivity. In an embodiment of the present disclosure, each of the one or more communication devices 102 is a portable communication device. Example of the portable communication device includes a laptop, a smart phone, a tablet and the like. For example, the smartphone may be an Apple smartphone, an Android smartphone, a Windows smartphone and the like. In another embodiment of the present disclosure, each of the one or more communication devices 102 is a fixed communication device. Examples of the fixed communication device include a desktop, a workstation PC and the like. Each of the one or more communication devices 102 runs on an operating system. In general, the operating system provides an interface for the user 102a to interact with hardware of each of the one or more commurjication devices 102 and other connected devices. In an example, the operating system installed in the one or more communication devices 102 is a Windows based operating system. In another example, the operating system installed in the one or more communication devices 102 is a Mac based operating system. In yet another embodiment of the present disclosure, the operating system installed in the one or more communication devices 102 is a Linux based operating system. In yet another example, the operating system installed in the one or more communication devices 102 is a mobile operating system. Example of the mobile operating system includes but may not be limited to Android operating system, Apple iOS, Symbian based operating system, BADA operating system and blackberry operating system. [0017] The one or more communication devices 102 are connected to the main server 110 through the communication network 104. Each communication device of the one or more communication devices 102 is registered with the main server 110. In general, the communication network 104 is a part of a network layer responsible for connection of two or more communication devices. Further, the communication network 104 may be any type of network. In an embodiment of the present disclosure, the type of communication network 104 is a wireless mobile network. In another embodiment of the present disclosure, the type of communication network 104 is a wired network with a finite bandwidth. In yet another embodiment of the present disclosure, the type of communication network 104 is a combination of the wireless and the wired network for the optimum throughput of data transmission. In yet another embodiment of the present disclosure, the type of communication network 104 is an optical fiber high bandwidth network that enables a high data rate with negligible connection drops.

[0018] The communication network 104 includes a set of channels. Each channel of the set of channels supports a finite bandwidth. The finite bandwidth of each channel of the set of channels is based on capacity of the communication network 104. Further, the one or more communication devices 102 possesses a unique machine address (hereinafter "MAC"). The MAC uniquely identifies each of the one or more communication devices 102 over the communication network 104. In addition, the communication network 104 provides a unique identity to each of the one or more communication devices 102. The unique identity is often referred to as an internet protocol (hereinafter "IP") address. In general, an IP address is a unique string of numbers separated by full stops that identify the one or more communication devices 102 using IP to communicate over the communication network 104. The IP address is characterized by IP versions. In an embodiment of the present disclosure, the IP address assigned to the one or more communication devices 102 is an IPv4 address. In another embodiment of the present disclosure, the IP address assigned to the one or more communication devices 102 is an IPv6 address.

[0019] The one or more communication devices 102 accesses data over the communication network 104 by utilizing one or more applications. The one or more applications include but may not be limited to a web browser, a mobile application, a widget and web applets. In general, each of the one or more applications have a graphical user interface (hereinafter "GUT') that is designed to display and fetch data from the main server 110. In addition, each of the one or more applications on any of the one or more communication devices associated with the user 102a may provide an interface for real time streaming, uploading and downloading of video files and audio files. The web browser installed in the one or more communication devices 102 may be any web browser. Example of the web browsers includes Google Chrome, Mozilla Firefox, Opera, UC Web, Safari, Internet Explorer and the like. In addition, the mobile application installed in at least one of the one or more communication devices 102 may be based on any mobile platform. Examples of the mobile platform include but may not be limited to Android, iOS Mobile, Blackberry and Bada.

[0020] In addition, the communication network 104 provides a unique identity to the communication device 102. The unique identity is often referred to as an internet protocol (hereinafter "IP") address. In general, an IP address is a unique string of numbers separated by full stops that identify the communication device 102 using IP to communicate over the communication network 104. The IP address is characterized by IP versions. In an embodiment of the present disclosure, the IP address assigned to the communication device 102 is an IPv4 address. In another embodiment of the present disclosure, the IP address assigned to the communication device 102 is an IPv6 address.

[0021] Each of the one or more communication devices 102 and the one or more publishers 106 are connected to the main server 110. In an embodiment of the present disclosure, the main server 110 interacts with requests from the one or more communication devices 102 through the communication network 104 (as shown in FIG. IB). In another embodiment of the present disclosure, the main server 110 interacts with each of the one or more publishers 106 through the communication network 104 (as shown in FIG. 1C). In yet another embodiment of the present disclosure, the main server 110 interacts with requests of each of the one or more communication devices 102 through the one or more publishers 106 (as shown in FIG. ID).

[0022] In addition, the user 102a and each of the one or more publishers 106 are a requestor of service from the main server 110. Each publisher of the one or more publishers 106 may be any website, web application, mobile application, third party applications and the like. Each publisher may be managed by a media content provider. In an example, XYZ is a news network and a broadcaster of news on television and online platform The publisher of XYZ news may be a web based platform, mobile app based platform or any individual content provider of media content. In another example, the publisher may be an individual or group providing videos to the interactive video system 112. Each of the one or more publishers 106 may be associated with a publisher database of the one or more publisher databases 108. Each publisher database of the one or more publisher databases 108 is a database of a digitally processed repository of videos. Each publisher of the one or more publishers 106 is registered on the main server 110.

[0023] The main server 110 provides a platform for video assembling and serving of interactive questions to the user 102a and each of the one or more publishers 106. The platform may be a web platform, mobile application platform, mobile web platform and the like. The main server 110 includes the interactive video system 112, a first database 114 and a second database 116. The interactive video system 112 services the request of the user 102a and each of the one or more publishers 106 in the real time. Further, the first database 114 is a proprietary database. The first database 114 includes a set of user authentication data and a user profile associated with the user 102a. Also, the first database 114 includes a set of publisher authentication data and a publisher profile associated with each publisher of the one or more publishers 106. The user 102a is identified uniquely by the set of user authentication data. The set of user authentication data includes an email address of the user 102a, a bio-data of the user 102a, an authentication key, a physical location and a standard time and time zone of login. The bio data of the user 102a may include full name, nickname, chronological age, gender and the like. In an embodiment of the present disclosure, the first database 114 is an encrypted database. In another embodiment of the present disclosure, the first database 114 is an unencrypted database. In addition, the first database 114 includes the set of feedbacks from each registered user 102a against each of the pre-defined set of interactive questions.

[0024] Further, the second database 116 is a database of digital processed repository of videos. The second database 116 stores one or more tagged videos. Each tagged video is virtually divisible into one or more tagged fragments. Each tagged video in the second database 116 is associated with a genre and a title. Examples of the genre include but may not be limited to sports, comedy, horror, drama, adventure, science fiction and autobiography. Also, each video may be associated with a popularity index and a number of views. In addition, each video is characterized by a set of technical specifications and non-technical specifications. The set of technical specifications include encoding format, frame rate, bit rate, frame height, frame width, pixel density, video resolution, size of video and the like. Each video may have different set of technical specifications. Each video in the second database 116 may have any encoding format. In an embodiment of the present disclosure, the encoding format is MPEG-4. In another embodiment of the present disclosure, the encoding format is FLV. In yet another embodiment of the present disclosure, the encoding format is AVI. In yet another embodiment of the present disclosure, the encoding format is 3GP. In yet another embodiment of the present disclosure, the encoding format is derived from proprietary codec. Moreover, the set of non-technical specifications include duration of video, a time reference associated with each video, the genre of video and the like.

[0025] Each video is tagged with one or more tags of a set of tags. The set of tags may correspond to a context of video, location reference in video, famous persons, events, genres, date, time and the like. In an example, a video of Moto GP race event is divisible into a lap of one or more laps. Each lap corresponds to a relative position of each racer in race chart. Each section may be tagged with the top racer of each lap. In another example, a video of interview of Mike Tyson is divisible into personal life, social life, professional life, struggles, success, events, etc. Each section of the interview of Mike Tyson can be tagged based on context of discussion. In an embodiment of the present disclosure, the second database 116 is updated with the one or more tagged videos from the one or more publishers 106. In another embodiment of the present disclosure, each publisher of the one or more publishers 106 updates the second database 116 with one or more untagged videos. Each video may be tagged with the set of tags and uploaded to the second database 116. Each video may be uploaded to the second database and tagged with the set of tags. The one or more untagged videos may be tagged manually by one or more administrators associated with the interactive video system 112.

[0026] The digital repository of videos in the second database 116 is updated with the one or more tagged videos from one or more sources. The one or more sources may include third party video content providers, the one or more publishers 106, the one or more advertisers, one or more sponsors and the like. Each publisher is a platform that uploads tagged videos to the digital repository of videos in the main server 110. The platform of each publisher may include a web based platform, a mobile application based platform, a web application based platform and the like. Additionally, the digital repository of videos may be updated and managed by the platform administrators. In an embodiment of the present disclosure, each video is manually tagged by the one or more administrators. In another embodiment of the present disclosure, the one or more administrators associated with operations of the main server 110 tag each video based on voice instructions. In yet another embodiment of the present disclosure, each video may be tagged based on speech rendering and analysis. In yet another embodiment of the present disclosure, each video is automatically tagged by the interactive video system 112. The automatic tagging of each video is done based on context mining and supervised digital fingerprinting of a set of frames. In yet another embodiment of the present disclosure, each video may be tagged by proprietary software and algorithms. In yet another embodiment of the present disclosure, each video may be tagged by the user 102a registered on the main server 110 and the publisher of the one or more publishers 106. In addition, each video may be tagged by media agency, advertiser, creative agency and the like. Each tag of the set of tags may be rated for ranking each tag and improving search efficiency.

[0027] Going further, the set of tags for each video may be updated based on real time determination of frequently used tags, frequently searched tags and less used tags. In addition, the set of tags for each video may be updated based on dynamic meta-tagging. The set of tags for each video may be updated based on incremental machine learning on the set of tags and the metadata for each tagged video. In an embodiment of the present disclosure, the metadata and meta-tagging for each tagged video may performed according to MPEG 7 standard. The MPEG 7 standard is also called as Multimedia Content Description Interface. For example, a video on Sachin may be tagged with Sachin, Master blaster, legend, god of cricket, and the like. The interactive video system 112 may determine the most used keyword to refer to content on Sachin. Let us suppose, in due course of 1 year, the interactive video system 112 determines that Sachin is frequently searched with "King of Cricket" tag. The interactive video system 112 updates the database of the set of tags associated with Sachin. In addition, the tags will be associated with any other video currently discussed in the public domain. If the name of Sachin surfaces in any new content related to any award show, then the tags will be automatically attached with the award show video too. The interactive video system 112 may present a Gantt chart of set of tags that are temporally classified based on occurrences within search queries and preferences of the users.

[0028] The updated set of tags may be determined based on feature detection and correlation in a specific quadrant of one or more frames of the tagged videos. For example, a 10 minute tagged video having a frame rate of 30 fps may be processed by selecting 1 key frame per second and performing feature detection. The feature detection may be based on incremental machine learning. Examples of the feature detection includes but may not be limited to face detection, object detection, motion detection, text detection, moving object detection and the like.

[0029] Also, the main server 110 provides the platform for real time display of the pre-defined set of interactive questions during streaming of the assembled video. The main server 110 dynamically handles requests of video assembling and posting of the pre-defined set of interactive questions. The platform may correspond to any one of the website, mobile application, web application, mobile browser based platform. In an embodiment of the present disclosure, the platform is a subscription based paid platform. In another embodiment of the present disclosure, the platform is a pay per view based paid platform. In yet another embodiment of the present disclosure, the platform is a free access, single registration and login based platform. The platform provides a video on demand service. Further, the platform includes but may not be limited to a media player, a list of thumbnails of the one or more tagged videos, recommendation panel, account panel, search panel, preference panel. The pre-defined selection criteria includes but may not be limited to a set of intervals of video broadcast, a physical location of the user 102a, an identified name of celebrity and categories of video. The pre-defined selection criteria are based on dates, time zones, days, seasons, physical locations, occasions, identified names, video genres and the like. The set of intervals of video broadcast corresponds to a time reference in the video. For example, the user 102a may be provided with criteria to view all the news aired between 4:00 PM to 4: 15 PM of a specific day In an example, the physical location may be used to narrow down content relevant to the physical location. The user 102a may like to watch videos relevant to the physical location. The physical location may be derived through many techniques. In an embodiment of the present disclosure, the physical location is derived from the global positioning system (hereinafter "GPS") module present in at least one of the one or more communication devices 102 associated with the user 102a. In another embodiment of the present disclosure, the physical location is derived from manual selection of the physical location from a pre-defined list of locations by the user 102a. In yet another embodiment of the present disclosure, the physical location is derived from internet service provider's server's location.

[0030] The user 102a enters the set of authentication data through the one or more communication devices 102. The user 102a provides a set of preference data to the main server 110 through the communication network 104. The set of preference data is a subset of the pre-defined selection criteria provided to the user 102a. In an example, the user 102a may select "Barack Obama", "elections" and "civil rights" from a politics category listed in the pre-defined selection criteria. In another example, the user may select "Tennis", "Serena Williams" and "Wimbledon" from sports category listed in the pre-defined selection criteria. The interactive video system 112 is configured to receive the set of preference data and extract the one or more videos corresponding to the set of preference data of the user 102a from the second database 116. For example, the interactive video system 112 may extract all the speeches of Barack Obama's at the time of presidential elections. In addition, the interactive video system 112 may filter any video from the speeches of Barack Obama without any sectional tag of civil rights. In addition, the interactive video system 112 is configured to virtually fragment and segregate the one or more videos. The interactive video system 112 is configured to virtually cluster and assemble logical clusters of mapped fragments, transcode and render the assembled video. The interactive video system 112 is configured to serve the assembled video, post or push the pre-defined set of interactive questions and collect the set of user feedbacks in the real time.

[0031] The interactive video system 112 receives the set of preference data associated with the user 102a from the pre-defined selection criteria. The set of preference data corresponds to the digitally processed repository of videos. In addition, the interactive video system 112 receives the set of user authentication data. The interactive video system 112 compares the set of authentication data corresponding to the user with corresponding set of authentication data in the first database 114. The interactive video system 112 allows for login based on a positive comparison of received set of authentication data with the set of the user authentication data present in the first database 114. In an embodiment of the present disclosure, the interactive video system 112 automatically handles content management associated with the set of preference data and the set of user authentication data. In another embodiment of the present disclosure, the content management associated with the set of preference data of the user 102a and the set of user authentication data is manually handled by one or more administrators. Each of the one or more administrators handles the content management by utilizing a content management tool. The content management corresponds to management of the user profile, streaming of the assembled video, editing and updating pre-defined selection criteria, editing pages in the user interface and the like.

[0032] The interactive video system 112 creates the user profile based on a received set of user authentication data, the set of preference data and the real time user viewing and selection behavior. The real time viewing and selection behavior corresponds to dynamic variation in the preferences of the user during due course of one or more active sessions of user on the video assembling and interactive video platform. In an embodiment of the present disclosure, the user profile is created based on a request from the user 102a. In another embodiment of the present disclosure, the user profile is created automatically by the interactive video system 112. The set of authentication data of the user 102a is stored in the user profile present in the first database 114. Furthermore, the interactive video system 112 extracts the one or more tagged videos related to the set of preference data of the user 102a from the digitally processed repository of videos present in the second database 116. The interactive video system 112 extracts the one or more tagged videos based on a correlation of the set of tags with the set of preference data associated with the user 102a. The set of tags are associated with each tagged video of the one or more tagged videos.

[0033] The interactive video system 112 virtually fragments each tagged video of the one or more tagged videos into the one or more tagged fragments. Each tagged video is fragmented into the one or more tagged fragments and each tagged fragment is characterized by length measured in a pre-determined interval of time. For example, the pre-determined interval of time is 5 seconds for each tagged fragment of a 300 seconds video. In an embodiment of the present disclosure, the pre-determined interval of time for each tagged fragment may be manually adjusted by the one or more administrators. In another embodiment of the present disclosure, the pre-determined interval of time for each tagged fragment may be automatically adjusted by the interactive video system 112 based on proprietary algorithms. Each tagged video is fragmented based on segmentation of the tagged video for each pre-determined interval of time. Also, the fragmentation of each tagged video is a virtual fragmentation in temporary memory of the main server 110.

[0034] The interactive video system 112 virtually segregates one or more mapped fragments of the one or more tagged fragments into one or more logical sets of mapped fragments. In an embodiment of the present disclosure, the one or more mapped fragments are segregated based on a positive mapping of keywords from the set of preference data with the set of tags. The set of tags are associated with each tagged fragment of the one or more tagged fragments. In addition, each tagged videos of the one or more tagged videos in the second database 116 is associated with a set of metadata. In another embodiment of the present disclosure, the one or more mapped fragments are segregated based on the positive mapping of the keywords from the set of preference data with the set of metadata. Each logical set of mapped fragments may correspond to a common tag from each tagged video of the one or more tagged videos.

[0035] For example, a user, say ABC provides preferences like Comedy, Jim Carrey and funny to the interactive video system 112. The interactive video system 112 fetches one or more tagged videos related to Jim Carrey, Comedy and funny preferences. The interactive video system fragments each of the one or more videos into tagged fragments. Each tagged fragment may be of 5 second duration. The interactive video system 112 may segregate the mapped fragments from the tagged fragments based on a positive mapping with the set of preference data of the user ABC.

[0036] The interactive video system 112 mines semantic context information from each mapped fragment of the one or more mapped fragments. In addition, the interactive video system 112 mines semantic context information from each logical set of mapped fragments of the one or more logical sets of mapped fragments. The semantic context information includes object specific context information and scene specific context information of each mapped fragment and each logical set of mapped fragments. For example, the one or more mapped fragments may be associated with common tags of comedy, movie, Hollywood and Jim Carrey. The interactive video system 112 mines semantic context information that includes dialogues, music, location, faces and the like. The interactive video system 112 may mine sentiments of characters in each mapped fragment from feature analysis of audio and faces. The interactive video system 112 may mine features that include geometrical shapes, color saturation, motion of objects, scene changes, number of scenes, animations and the like.

[0037] Going further, the interactive video system 112 virtually clusters the one or more logical sets of mapped fragments into one or more logical clusters of mapped fragments. Each logical cluster of mapped fragments is derived from at least one of the one or more logical sets of mapped fragments. For example, the interactive video system 112 fetches three tagged comedy videos of Jim Carrey. The interactive video system 112 fragments each of the three tagged comedy videos of Jim Carrey. The mapped fragments out of tagged fragments for each tagged video may be segregated into the logical set of mapped fragments. The mapped fragments for action and comedy tags in the three videos may be segregated to obtain the logical set of mapped fragments. The logical set of mapped fragments for comedy and action tags for each tagged video may be clustered in the logical cluster.

[0038] The interactive video system 112 performs auto volume leveling on each audio segment associated with the one or more mapped fragments or logical clusters of the mapped fragments. For example, the first logical cluster may contain fragments having different volume levels. The interactive video system 112 may dynamically normalize volume levels on a uniform scale. In addition, the interactive video system 112 may perform image normalization on each frame of the mapped fragments.

[0039] In an embodiment of the present disclosure, the interactive video system 112 virtually assembles at least one of the one or more logical clusters of mapped fragments in a pre-defined order of preference to obtain an assembled video. Each logical cluster of mapped fragments is assembled based on an analysis of the set of preference data of the user and the semantic context information. For example, the user may provide preferences like adventure, Nicholas Cage, movie and fighting scenes. The one or more tagged video with tags of adventure and Nicholas Cage and movie may be tagged with specific fighting scenes. The interactive video system 112 mines semantic context information from each tagged video through searching for fights related keywords from rendered speeches, scene detection, object movement, music, speech analysis, tone analysis and the like. The semantic context information may be used to automatically tag, fragment, cluster or assemble videos on demand.

[0040] In an embodiment of the present disclosure, the pre-defined order of preference is derived from the set of preference data, the user profile and the semantic context information mined from the activities of user 102a. In another embodiment of the present disclosure, the pre-defined order of preference is derived from preferences of users with similar user profiles and situations. In another embodiment of the present disclosure, the interactive video system 112 virtually assembles at least one of the one or more logical clusters of mapped fragments in a dynamically generated pre-defined order of preference. The dynamically generated pre-defined order of preference is based on a real time viewing and selection behavior of the user 102a. In an embodiment of the present disclosure, the pre-defined order of preference corresponds to a linear and non-sequential assembling of the one or more logical set of mapped fragments. In another embodiment of the present disclosure, the pre-defined order of preference corresponds to a non-linear and non-sequential assembling of the one or more logical set of mapped fragments. Each logical set of mapped fragments is a virtually clustered in the temporary memory of the main server 110. The interactive video system 112 presents a personalized video solution for each user 102a.

[0041] The interactive video system 112 removes duplicate tags from set of tags of the real time and dynamically assembled video in the temporary memory of the main server 110. The duplicate tags along the set of metadata of the assembled video are flushed in the disk for faster transmission and caching of the assembled video on different communication devices.

[0042] In an example, a person, say X selects tags corresponding to sports. The person (X) selects tags corresponding to Mike Tyson and boxing. In addition, the person selects the knockout tag from the pre-defined selection criteria. The knockout moment is often an ending portion of a boxing match. The interactive video system 112 fetches the one or more tagged videos associated to matches of Mike Tyson. The interactive video system 112 searches for a knockout tag in at least one of the one or more pre-defined sections of each tagged video. The interactive video system 112 fragments each tagged video of Mike Tyson into tagged fragments and segregates logical set of mapped fragments for knockout by Mike Tyson tag from other tagged clips of Mike Tyson. The interactive video system 112 may clusters each logical set of mapped fragments to obtain logical clusters of mapped fragments. The logical clusters may be assembled in the real time to obtain the assembled video. In addition, the interactive video system 112 may assemble each logical cluster or mapped fragments for the knockout by Mike Tyson based on number of views. The interactive video system 112 dynamically serves a reassembled video to the user 102a in the real time upon a click on any video recommendations. The interactive video system 112 dynamically reassembles the one or more mapped fragments or logical clusters of mapped fragments in the real time.

[0043] The user 102a may request to stream the assembled video that includes specific segments of 360° videos (or immersive videos), the tagged set of videos and a live video (as shown in FIG. IE). The main server 110 is associated with a live media server 118. The live media server 118 is a high bandwidth media server that is configured to stream live videos to each communication device of the one or more communication devices 102. The interactive video system 112 virtually fetches and segregates the one or more mapped fragments of the 360° videos and the one or more tagged videos. The mapped fragments of 360° videos and mapped fragments of tagged videos are derived from comparison of the keywords from the set of preference data with tags of the 360° videos and traditional videos. In addition, the interactive video system 112 requests a live media server 118 for live streaming of the live video. The interactive video system 112 virtually assembles the mapped fragments of 360° videos and mapped fragments of videos. The interactive video system 112 streams the virtually assembled mapped fragments of 360° videos and the mapped fragments of videos. In addition, the interactive video system 112 switches from the assembled content to the live video received from the live media server in the real time.

[0044] The interactive video system 112 transcodes the assembled video into a pre-defined video format by utilizing a codec. The assembled video is transcoded to enable adaptive bitrate streaming on each communication device of the one or more communication devices 102. The assembled video is transcoded based on one or more device parameters and one or more network parameters. The one or more device parameters include screen size, screen resolution, pixel density and the like. Further, the one or more network parameters include an IP address, network bandwidth, maximum bitrate support over network, throughput, connection strength, location of requesting server and the like. In an example, the user 102a may be using a laptop with a limited bandwidth insufficient for high definition streaming of videos. Accordingly, the interactive video system 112 transcodes the assembled video in a format up-loadable from the main server 110. In another example, the user 102a may be using a smartphone with a low bandwidth and a lower display resolution. Accordingly, the interactive video system 112 transcodes the assembled video in the format viewable for the lower display resolution screens. Further, the interactive video system 112 utilizes salt stack to scale up and down transcoding requirements. The salt stack utilizes shell scripts to execute FFMPEG in the main server 110.

[0045] In an embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in 144p quality. In another embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in 240p quality. In yet another embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in 360p quality. In yet another embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in 480p quality. In yet another embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in 720p quality. In yet another embodiment of the present disclosure, the interactive video system 112 transcodes the video in 1080p quality. In yet another embodiment of the present disclosure, the interactive video system 112 transcodes the assembled video in any standard quality.

[0046] In addition, the interactive video system 112 trans-rates and trans-sizes the assembled video to enable adaptive streaming for each communication device of the one or more communication devices 102. The interactive video system 112 transcodes the assembled in any standard video coding format, container and audio coding format. Examples of the video coding format includes but may not be limited to MPEG-2 Part 2, MPEG-4 Part 2, H.264 (MPEG-4 Part 10), HEVC, Theora, Real Video RV40, VP9, and AVI. Examples of the container includes but may not be limited to Matroska, FLV, MPEG-4 part 12, VOB, HTML and real media. Example of the audio coding format includes but may not be limited to MP3, AAC, Vorbis, FLAC, and Opus. In an embodiment of the present disclosure, the assembled video is in the MP4 file format. In another embodiment of the present disclosure, the assembled video in the matroska file format. In yet another embodiment of the present disclosure, the assembled video is in the AVI file format. In yet another embodiment of the present disclosure, the assembled video is in the FLV file format. In yet another embodiment of the present disclosure, the assembled video is in the 3GP file format.

[0047] The assembled video is transcoded based on an audio codec and a video codec. The audio codec and the video codec may be any generic or proprietary codec. Example of the video codecs include but may not be limited to H.265/MPEG-H HEVC codec, H.264/MPEG-4 AVC codec, H.263/MPEG-4 codec, H.263/MPEG- 4 Part 2 codec, H.262/MPEG-2 codec and ACT-L3 codec. The compression performed by the video codecs on the assembled video is a lossy compression. In addition, the interactive video system 112 utilizes a media streaming communication protocol to stream the real time and dynamically assembled video on each of the one or more communication devices 102. In an embodiment of the present disclosure, the media streaming communication protocol is a HTTP live streaming (hereinafter "HLS") protocol. In another embodiment of the present disclosure, the media streaming communication protocol is a MPEG based dynamic adaptive streaming over HTTP (hereinafter "MPEG-DASH") protocol.

[0048] The interactive video system 112 renders the assembled video for addition of one or more interactive elements and a bi-directional flow. The one or more interactive elements include forward playback, reverse playback, fast playback and slow playback. In addition, the one or more interactive elements include touch based navigation option, swipe based navigation option, click based navigation option, voice based navigation, motion based navigation and the like.

[0049] In an embodiment of the present disclosure, the interactive video system 112 includes a question management system 112a (as shown in FIG. IF). The interactive video system 112 is associated with a question database 120. The question database 120 includes a large repository of structured, indexed and classified questions. The large repository of questions includes contest based questions, product specific questions, sponsor questions, publisher based question, quality of service questions, service feedback questions, marketing questions, ad targeting questions and the like. In addition, each question may be associated with graphical media. The graphical media may include flash files, images, animations, GIFs, clips and the like. In addition, the question management system 112a manages and serves a pre-defined set of interactive questions to the user 102a. The question management system 112a sorts and rates the difficulty and rank of each interactive question. In an example, a contest of 10 questions may be ranked based on the difficulty level of each question. Each question may be assigned a score of difficulty in order to assess the feedbacks for each interactive question in the real time. In an embodiment of the present disclosure, the difficulty level of each question may be set manually by the one or more administrators. In another embodiment of the present disclosure, the question management system 112a may determine the difficulty level based on supervised machine learning algorithms. In yet another embodiment of the present disclosure, the question management system may determine the difficulty level based on delay and errors in responses for each question shared among multiple users. In addition, the question database 112a is updated with metadata of each question and the set of user feedbacks on each question. The question management system 112a records the id of each posted question in the question database 120 in order to avoid repetitive posting. The graphical media for each question may be presented for call to action objective. For example, the user may be presented with an advertisement of a specific car and an interactive questionnaire overlaid on top of the advertisement video. The user 102a answers the questionnaire and is targeted with an option for test drive. The details of the user 102a may be transferred to an associated car dealer for arranging test drive for the user. In addition, the interactive questionnaire may be supplemented or complemented by images or text on videos hotspots.

[0050] Further, the interactive video system 112 fetches the pre-defined set of interactive questions from the large repository of interactive questions present in the question database 120. The pre-defined set of interactive questions is based on content of each assembled video, a quality and satisfaction associated with the web based platform, user experience, quality of service, requirement of recommendations and the user profile. In addition, the pre-defined set of interactive questions may be any type of interactive questions. In an embodiment of the present disclosure, each interactive question is text based multiple choice questions. In another embodiment of the present disclosure, each interactive question is touch enabled question. In yet another embodiment of the present disclosure, each interactive question is an image based interactive question. Each interactive question may provide user 102a with an option to interact and provide feedback. The pre-defined set of interactive questions is fetched to invoke the set of user feedbacks for a pre-defined set of feedback subjects. The predefined set of feedback subjects includes but may not be limited to quality of service, user experiences, user satisfaction and suggestions for any change in the user interface.

[0051] The interactive video system 112 posts the pre-defined set of interactive questions during the streaming of the assembled video in the communication device 102 of the user 102a. The pre-defined set of interactive questions is posted based on the semantic context information mined from the set of preference data of the user 102a and the one or more tagged videos. In addition, the pre-defined set of interactive questions may be posted based on a real time viewing and selection behavior of the user 102a, a request of the user 102a for a series of questions and dynamic requests of posting questions from one or more publishers 106. Each interactive question of the pre-defined set of interactive questions is posted in at least one of one or more pre-defined regions of the user interface. The one or more regions includes but may not be limited to one or more video interactivity windows, recommendation sidebars and video player. In an embodiment of the present disclosure, the first pre-defined order of preference is set by one or more administrators associated with management of the platform hosted by the main server 110. In another embodiment of the present disclosure, the first pre-defined order of preference is set based on previous feedbacks from the user 102a. In yet another embodiment of the present disclosure, the first pre-defined order of preference is based on increasing order of difficulty. In yet another embodiment of the present disclosure, the first pre-defined order of preference is based on decreasing order of difficulty.

[0052] Each interactive question of the pre-defined set of interactive questions may be posted based on any suitable condition. In an embodiment of the present disclosure, each interactive question is posted based on a time based trigger. In another embodiment of the present disclosure, each interactive question is posted based on a scene based trigger. In yet another embodiment of the present disclosure, each interactive question is posted based on manual trigger set by the one or more administrators. In yet another embodiment of the present disclosure, each interactive question may be posted based on context mining of previous set of user feedbacks. In an embodiment of the present disclosure, the interactive video system 112 dynamically fetches the set of interactive questions from the repository of questions. The set of interactive question are fetched to obtain the set of user feedbacks. In another embodiment of the present disclosure, the interactive video system 112 dynamically enables the publishers, advertisers or any third party websites to directly push interactive questions to the user 102a during the streaming of each assembled logical cluster in the assembled video. The set of user feedbacks are collected dynamically in the real time corresponding to each posted interactive question. The user 102a may be given an option to either reply to the interactive question or skip the question. The interactivity may be further enhanced by introducing interactive elements like shake to reply, voice based feedback, tap to reply, haptic feedbacks, gesture based feedbacks and the like. In another embodiment of the present disclosure, the interactive video system 112 enables pushing of the pre-defined set of interactive questions from at least one of one or more publishers, one or more system administrators and the like. The pre-defined set of interactive questions is pushed in the real time. The interactive video system 112 collects the set of user feedbacks corresponding to a posted pre-defined set of interactive questions from the user 102a. The collection of the set of user feedbacks is done in the real time. Further, the interactive video system 112 updates the user profile corresponding to the user 102a based on variation in the set of preference data in the first database 114. In addition, the interactive video system 112 updates the assembled video and the set of user feedbacks in the real time.

[0053] It may be noted that in FIG. 1A, FIG. IB, FIG. 1C and FIG. ID, the communication device 102 corresponding to the user 102a is connected to the main server 110; however, those skilled in the art would appreciate that more number of communication devices associated with more number of users is connected to the main server 110. It may be noted that in FIG. 1A, FIG. IB, FIG. 1C and FIG. ID, the main server 110 is the provider of video assembling and posting of pre-defined set of interactive questions; however, those skilled in the art would appreciate that more number of main servers synchronously provide video assembling and interactive question posting services. It may be noted that in FIG. 1A, FIG. IB, FIG. 1C and FIG. ID, the communication device 102 associated with the user 102a is connected to the main server 110 through the communication network 104; however, those skilled in the art would appreciate that more number of communication devices associated with more number of users is connected to more number of main servers through more number of communication networks.

[0054] FIG. 2A illustrates an example of the web platform for posting the interactive questions during streaming of a real time, dynamic, adaptive and non-sequentially assembled video. The web platform includes an assembled video player (PI), a recommendation sidebar (RSI) and a video interactivity window (BS1). The web platform is accessed from the communication device 102 of the user 102a. The user 102a selects an assembled video (AVI) from the recommendation sidebar (RSI). The assembled video player (PI) loads the logical clusters of mapped fragments of the assembled video (AVI) virtually in the memory. The question management system 112a fetches a first interactive question (Ql), a second interactive question (Q2), a third interactive question (Q3), a fourth interactive question (Q4) from the question database 120. In addition, the question management system fetches a fifth interactive question (Q5) and a sixth interactive question (Q6) from the question database 120. Each question may be associated with a context of specific segment of the assembled videos or may be associated with a specific contest. In an example, the question management system 112 may post the first question (Ql), the second question (Q2), the third question (Q3), the fourth question (Q4), the fifth question (Q5), the sixth question (Q6) on specific regions of the platform. Each question may be supplemented by video, text or image hotspots. The interactive questions may be posted on any specific region of the platform. The specific regions may include video interactive windows, video player, header, footer and the like. In another example, the one or more publishers 106 pushes the first question (Ql), the second question (Q2), the third question (Q3), the fourth question (Q4), the fifth question (Q5) and the sixth question (Q6). The user 102a may interact with any of the first question (Ql), the second question (Q2), the third question (Q3) and the fourth question (Q4) by providing feedbacks. In addition, the user 102a may interact with any one of the fifth question (Q5) and the sixth question (Q6). The question management system updates and responds to the feedbacks associated with the first question (Ql), the second question (Q2), the third question (Q3), the fourth question (Q4) the fifth question (Q5) and the sixth question (Q6).

[0055] FIG. 2B illustrates an example of the real time, dynamic, adaptive and non-sequential assembling of the one or more mapped fragments of the one or more tagged videos. In the example, the one or more tagged videos include a first video (VI), a second video (V2) and a third video (V3). The interactive video system 112 receives the request of service from the user 102a through the communication network 104. The user 102a provides the set of preference data and the set of authentication data to the interactive video system 112. The interactive video system 112 fetches the first video (VI), the second video (V2) and the third video (V3) from the second database 116. In addition, the one or more pre-defined sections of the first video (VI), the second video (V2) and the third video (V3) are tagged with the set of tags. The interactive video system 112 fragments and logically clusters the first video (VI) into a first logical cluster (V1C1), a second logical cluster (V1C2), a third logical cluster (V1C3), a fourth logical cluster (V1C4), a fifth logical cluster (V1C5) and a sixth logical cluster (V1C6). In addition, the interactive video system fragments and logically clusters a seventh logical cluster (V1C7) and an eighth logical cluster (V1C8). Accordingly, the interactive video system 112 fragments and logically clusters the second video (V2) into a first logical cluster (V2C1), a second logical cluster (V2C2) and a third logical cluster (V2C3). The interactive video system 112 clusters a fourth logical cluster (V2C4), a fifth logical cluster (V2C5) and a sixth logical cluster (V2C6). In addition, the interactive video system 112 clusters a seventh logical cluster (V2C7), an eighth logical cluster (V2C8) and a ninth logical cluster (V2C9). The interactive video system performs similar operations on the third video (V3). The fragmentation of the first video (VI), the second video (V2) and third video (V3) is done for a pre-determined interval of time. The first set of logical clusters (V1C1-V1C8), the second set of logical clusters (V2C1-V2C9) and the third set of logical clusters (V3C1-V3C6) includes 8, 9 and 6 logical clusters of fragments respectively.

[0056] The interactive video system 112 non-linearly and non-sequentially assembles the fourth logical cluster (V2C4), the second logical cluster (V1C2), the fourth logical cluster (V3C4) and the third logical cluster (V1C3). In addition, the interactive video system assembles the fifth logical cluster (V2C5), the first logical cluster (V3C1) and the sixth logical cluster (V1C6) to obtain the assembled video. The pre-defined order of preference of corresponding clips is derived from the set of the preference data of the user 102a and the user profile corresponding to the user 102a. The assembled video is transcoded into the pre-defined format by the interactive video system 112. The assembled video in transcoded format is streamed to the user 102a in the real time.

[0057] FIG. 3 illustrates a block diagram 300 of the interactive video system 112, in accordance with various embodiments of the present disclosure. The block diagram 300 illustrates one or more components of the interactive video system 112. The one or more components collectively enable display of the interactive questions during streaming of the real time, dynamic, adaptive and non-sequentially assembled video. It may be noted that to explain the system elements of the FIG. 3, references will be made to the system elements of the FIG. 1A, FIG. IB, FIG. 1C and FIG. ID. The block diagram 300 includes a reception module 302, a creation module 304, an extraction module 306, a fragmentation module 308, a segregation module 310 and a mining module 312. In addition, the block diagram 300 includes a clustering module 314, assembling module 316, transcoding module 318, a rendering module 320, a serving module 322, a fetching module 324, a posting module 326, a collection module 328 and an update module 330.

[0058] The reception module 302 receives the set of preference data associated with the user 102a. The reception module 302 receives the set of preference data from the pre-defined selection criteria. The pre-defined selection criteria are related to the digitally processed repository of videos (as discussed above in the detailed description of FIG. 1A). Further, the creation module 304 creates the user profile based on the received set of user authentication data and the set of preference data (as described above in the detailed description of FIG. 1A).

[0059] The extraction module 306 extracts the one or more tagged videos based on the set of preference data of the user 102a from the digitally processed repository of videos. The extraction module 306 extracts the one or more videos based on the correlation of the set of tags with the set of preference data associated with the user 102a (as discussed above in the detailed description of FIG. 1A). Further, the fragmentation module 308 fragments each tagged video of the one or more tagged videos into the one or more tagged fragments. Each tagged video is fragmented into the one or more tagged fragments and each tagged fragment is characterized by a pre-determined interval of time. Each tagged video is fragmented based on segmentation of the tagged video for each pre-determined interval of time (as discussed above in the detailed description of FIG. 1A).

[0060] Going further, the segregation module 310 segregates one or more mapped fragments of the one or more tagged fragments into one or more logical sets of mapped fragments. The segregation module 310 the one or more mapped fragments based on a positive mapping of keywords from the set of preference data with the set of tags associated with each tagged fragment of the one or more tagged fragments (as described above in the detailed description of FIG. 1A). The mining module 312 mines semantic context information from each mapped fragment of the one or more mapped fragments and each logical set of mapped fragments of the one or more logical sets of mapped fragments. The semantic context information includes an object specific context information and scene specific context information of each mapped fragment and each logical set of mapped fragments (as discussed in detailed description of FIG. 1A). Further, the clustering module 314 clusters the one or more logical sets of mapped fragments into the one or more logical clusters of mapped fragments (as discussed above in the detailed description of FIG. 1A).

[0061] Further, the assembling module 316 assembles at least one of the one or more logical clusters of mapped fragments in the pre-defined order of preference to obtain the assembled video. Each logical cluster of mapped fragments is assembled based on the analysis of the set of preference data of the user and the semantic context information (as discussed in detailed description of FIG. 1A). The transcoding module 318 transcodes assembled video into a pre-defined video format. The transcoding module 318 utilizes a codec. The codec may be any standard codec or proprietary codec. The transcoding module 318 transcodes assembled video to enable adaptive bitrate streaming on each of the one or more communication devices 102. The adaptive bitrate streaming is based on one or more device parameters and one or more network parameters (as discussed above in the detailed description of FIG. 1A). The rendering module 320 renders the assembled video for addition of one or more interactive elements and a bi-directional flow (as discussed above in the detailed description of FIG. 1A). The serving module 322 serves the assembled video based on the received set of preference data. The assembled video is served to the user 102a in the real time (as stated above in the detailed description of the FIG. 1A). The fetching module 324 fetches the pre-defined set of interactive questions. The pre-defined set of interactive question is fetched to obtain the set of user feedbacks (as mentioned above in the detailed description of the FIG. 1A).

[0062] The posting module 326 posts the pre-defined set of interactive questions during the serving of the assembled video. The pre-defined set of interactive questions is posted in the first pre-defined order of preference. In addition, each interactive question of the pre-defined set of interactive question is posted in the one of one or more pre-defined regions of the user interface (as provided above in the detailed description of the FIG. 1A). The collection module 328 collects the set of user feedbacks associated to the posted pre-defined set of interactive questions from the user 102a (as mentioned above in the detailed description of the FIG. 1A). The update module 330 updates the set of preference data associated with the user 102a. In addition, the update module 330 updates the user profile and the set of user feedbacks. The user profile is updated based on variations in the set of preference data Moreover, the update module 330 updates the assembled video in the digitally processed repository of videos in the real time (as described above in the detailed description of the FIG. 1A).

[0063] FIG. 4 illustrates a flow chart 400 for enabling navigation to the one or more discrete segments of the real time dynamic and adaptive non-linear, non-sequentially assembled video, in accordance with various embodiments of the present disclosure. It may be noted that to explain the process steps of flowchart 400, references will be made to the system elements of the FIG. 1A, FIG. IB, FIG. 1C, FIG. ID, FIG. IE and the FIG.3. It may also be noted that the flowchart 400 may have lesser or more number of steps.

[0064] The flowchart 400 initiates at step 402. Following step 402, at step 404, the reception module 302 receives the set of preference data associated with the user 102a from the pre-defined selection criteria. The predefined selection criteria correspond to digitally processed repository of videos. In addition, the reception module 302 receives the set of user authentication data. At step 406, the serving module 318 serves the assembled video to the user in the real time. The assembled video is served based on the received set of preference data. At step 408, the fetching module 320 fetches the pre-defined set of interactive questions. At step 410, the posting module 322 posts the pre-defined set of interactive questions during the serving of the assembled video. At step 412, the collection module 324 collects the set of user feedbacks from the user 102a. The set of user feedbacks corresponds to a posted pre-defined set of interactive questions. The flowchart 400 terminates at step 416.

[0065] It may be noted that the flowchart 400 is explained to have above stated process steps; however, those skilled in the art would appreciate that the flowchart 400 may have more/less number of process steps which may enable all the above stated embodiments of the present disclosure.

[0066] FIG. 5 illustrates a block diagram of a communication device 500, in accordance with various embodiments of the present disclosure. The communication device 500 enables host process of the interactive video system 112. The communication device 500 includes the interactive video system 112, a control circuitry module 502, a storage module 504, an input/output circuitry module 506, and a communication circuitry module 508. The communication device 500 includes any suitable type of portable electronic device. The communication device 500 includes but may not be limited to a personal e-mail device and a cellular telephone. In addition, the communication device 500 includes a smartphone, the laptop, computer and the tablet. In another embodiment of the present disclosure, the communication device 500 can be a desktop computer.