Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
CONTROLLING DELIVERY OF CAPTURED STREAMS
Document Type and Number:
WIPO Patent Application WO/2016/202884
Kind Code:
A1
Abstract:
There is provided a technique for providing streaming services, comprising: a plurality of capture devices, each for generating a captured stream of content; a server, for receiving the plurality of captured streams, and for outputting at least one output stream; and an editing device for outputting a control signal to the server, wherein the server processes captured streams to provide one or more modified output stream in dependence on the control signal.

Inventors:
SHAW PHILIP (GB)
TILMANN RALF (DE)
MAAS HANS- JURGEN (DE)
HEILAND PETER (US)
HAMAIDE FABRICE (FR)
BULLET KRISTAN (GB)
WEAVER MILES (GB)
EVERETT SEAN (US)
CHRISTIE MARK (GB)
Application Number:
PCT/EP2016/063797
Publication Date:
December 22, 2016
Filing Date:
June 15, 2016
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
PIKSEL INC (US)
SHAW PHILIP (GB)
International Classes:
H04N21/218; H04N21/232; H04N21/242; H04N21/2665; H04N21/414; H04N21/4223; H04N21/81; H04N21/8547
Domestic Patent References:
WO2007117613A22007-10-18
Foreign References:
US20090148124A12009-06-11
US20150043892A12015-02-12
US20140281011A12014-09-18
US20140169766A12014-06-19
EP2403236A12012-01-04
US20130339539A12013-12-19
US20120265621A12012-10-18
Other References:
CRISTIAN HESSELMAN ET AL: "Sharing enriched multimedia experiences across heterogeneous network infrastructures", IEEE COMMUNICATIONS MAGAZINE, IEEE SERVICE CENTER, PISCATAWAY, US, vol. 48, no. 6, 1 June 2010 (2010-06-01), pages 54 - 65, XP011469851, ISSN: 0163-6804, DOI: 10.1109/MCOM.2010.5473865
WEN GAO ET AL: "Vlogging", ACM COMPUTING SURVEYS, ACM, NEW YORK, NY, US, US, vol. 42, no. 4, 23 June 2010 (2010-06-23), pages 1 - 57, XP058090589, ISSN: 0360-0300, DOI: 10.1145/1749603.1749606
PEREIRA F ET AL: "Multimedia Retrieval and Delivery: Essential Metadata Challenges and Standards", PROCEEDINGS OF THE IEEE, IEEE. NEW YORK, US, vol. 96, no. 4, 1 April 2008 (2008-04-01), pages 721 - 744, XP011206030, ISSN: 0018-9219
BENJAMIN TAG ET AL: "Collaborative storyboarding through democratization of content production", ADVANCES IN COMPUTER ENTERTAINMENT TECHNOLOGY, ACM, 2 PENN PLAZA, SUITE 701 NEW YORK NY 10121-0701 USA, 11 November 2014 (2014-11-11), pages 1 - 4, XP058064259, ISBN: 978-1-4503-2945-3, DOI: 10.1145/2663806.2663875
Attorney, Agent or Firm:
WILLIAMS, David John et al. (Bedford HouseJohn Street, London WC1N 2BF, GB)
Download PDF:
Claims:
Claims

1. A system for providing streaming services, comprising: a plurality of capture devices, each for generating a captured stream of content;

a server, for receiving the plurality of captured streams, and for delivering one or more output streams; and an editing device for outputting a control signal to the server,

wherein the server delivers the one or more outputs streams as a modified version of one or more of the input streams in dependence on the control signal.

2. The system of claim 1 wherein the server aligns the timing of the control signal with the timing of the captured streams .

3. The system of claim 2 wherein the control signal comprises one or more timing marks, and the server is configured to align said one or more timing marks with a time line of received capture data streams.

4. The system of any preceding claim wherein the server is provided with metadata associated with or embedded in each captured data stream, the server being configured to align the timing of the control signal additionally with the metadata of the captured data stream.

5. The system of any preceding claim wherein the server automatically generates the output stream in dependence on the control signal.

6. The system of any preceding claim wherein the control signal identifies which of the plurality of captured streams is to be output from the server.

7. The system of any preceding claim 6 wherein one or more output streams is a modified version of one or more input streams .

8. The system of claim 7 wherein one or more output streams comprise two or more input streams as a composite or videos, effects, transitions or combination thereof, such as in a spit screen arrangement or two or more input streams with a crossfade.

9. The system of any preceding claim wherein the control signal is generated by the editing device under the control of a user of the editing device, said user observing events being captured by the plurality of capture devices, wherein the user adjusts the control signal in accordance with a determination as to which capture device should be chosen at any instant in time, the determination identifying which of the plurality of capture streams should be selected at any instant in time in dependence on the users determination of the capture device or devices to select at that instant.

10. The system of any preceding claim wherein the server is configured to apply an overlay to the output stream in dependence on the control signal.

11. The system of claim 10 wherein one or more output streams comprise two or more input streams to which an overlay or video effect is applied.

12. The system of any preceding claim wherein the plurality of captured streams capture live events, and the output stream is a live event stream.

13. The system of any preceding claim wherein the editing device receives a representation of each or a subset of the captured data streams.

14. The system of claim 13 wherein the editing device receives a lower bandwidth version of the one or more captured data streams .

15. The system of any preceding claim wherein an editing application is provided on the editing device.

16. The system of any preceding claim wherein the editing device is a mobile device.

17. The system of any preceding claim wherein the editing device is a device of a rights holder, wherein the output stream is associated with a rights holder event.

18. The system of claim 17 wherein the control data is an instruction for the server to apply or alter rights control information applied to the output stream.

19. The system of claim 18 wherein the rights control information is digital rights management information.

20. A method for providing streaming services, comprising: generating, at a plurality of capture devices, captured streams of content;

receiving, at a server, the plurality of captured streams, and outputting one or more output streams; and

outputting a control signal to the server from an editing device ,

wherein the server delivers the one or more outputs streams as a modified version of one or more of the input streams in dependence on the control signal.

21. The method of claim 20 further comprising aligning the timing of the control signal with the timing of the captured streams .

22. The method of claim 21 wherein the control signal comprises one or more timing marks, and the server is configured to align said one- or more timing marks with a time line of received capture data streams. 23. The method of any one; of claims 20 to 22 wherein the server is provided with metadata associated with or embedded in each captured data stream, the server being' configured to align the timing of the control signal additionally with the metadata of the captured data stream. 24. The method of any one of claims 20 to 23 wherein the server automatically generates the output stream in dependence on the control signal.

25 . The method of any one of claims 20 to 24 wherein the; control signal identifies which of the plurality of. captured streams is to be output from the server.

26. The method of claim 25 further comprising mapping one or more captured streams to an output . 27 . The method of claim 26 wherein one or more output streams comprise two or more input streams as a composite or videos, effects, transitions or combination thereof, such as in a spit screen arrangement or two or more input streams with a crossfade. 28. The method of any one of claims 20 to 27 wherein the control signal is generated by the editing device under the control of a user of the editing device, said user observing the events being captured by the plurality of capture devices, wherein the user adjusts the control signal in accordance with a determination as to which capture device should be chosen at any instant in time, the determination identifying which of the plurality of capture streams should be selected at any instant in dependence on the users determination of the capture device or devices to select at that instant ,

29. The method of any one of claims 20 to 28 wherein the server is configured to apply an overlay to the output stream in dependence on the control signal.

30. The method of claim 29 wherein the server applies an overlay to a captured data stream, which captured data stream is routed to the output stream.

31. The method of claim 30 wherein the server applies the overlay to at least two captured, data streams which as a composite comprise an output stream.

32. The method of any one of claims 20 to 31 wherein the plurality of captured streams capture live events, and the viewing stream is a live event stream.

33. The method of any one of claims 20 to 32 wherein Che editing device receives a representation of each or a subset of the captured data, streams ,

34. The method of claim 33 wherein the editing' device receive© a lower bandwidth version of the one or more captured data streams .

35. The method of any one of claims 20 to 34 wherein an editing application is pi~ovided on the editing device.

36. The method of any one of claims 20 to 35 wherein the editing device is a mobile device,

37. The method of any one of claims 20 to 3S wherein the editing device is a device of a rights holder, wherein the output stream is associated with a rights holder event.

38. The method of claim 37 wherein the control data is an instruction for the server to apply or alter rights control information applied to the output stream.

39. The method of claim 38 wherein the rights control information is digital rights management information.

Description:
CONTROLLING DELIVERY OF CAPTURED STREAMS

BACKGROUND TO THE INVENTION:

Field of the Invention:

The invention is concerned with a technique for receiving streams from a plurality of capture devices at a central server, and processing the captured streams such that a modified output stream is generated therefrom in dependence on a control signal.

Description of the Related Art:

It is known to provide data streams from a plurality of different capture devices. The data streams may be provided for viewing by viewing devices. The capture and viewing devices may be, for example, mobile phones.

Locating captured streams to view can be difficult, and in particular selecting streams to view from a plurality of existing streams can be difficult.

It is an aim to provide improvements.

SUMMARY OF THE INVENTION:

There is provided a system for providing streaming services, comprising: a plurality of capture devices, each for generating a captured stream of content; a server, for receiving the plurality of captured streams, and for outputting one or more output streams; and an editing device for outputting a control signal to the server, wherein the server delivers the one or more outputs streams as a modified version of one or more of the input streams in dependence on the control signal.

The server may align the timing of the control signal with the timing of the captured streams. The control signal may comprise one or more timing marks, and the server is configured to align said one or more timing marks with a time line of received capture data streams.

The server may be provided with metadata associated with or embedded in each captured data stream, the server being configured to align the timing of the control signal additionally with the metadata of the captured data stream.

The server may automatically generate the output stream in dependence on the control signal.

The control signal may identify which of the plurality of captured streams is to be output from the server. Each captured stream may be mapped to an output. One or more captured streams may be mapped to an output. One or more captured streams may be mapped to one or more outputs.

One or more output streams may comprise two or more input streams as a composite of videos, effects, transitions or a combination thereof. Examples area split screen arrangement or two or more input streams with a crossfade.

In general, the control signal is used to determine which of a plurality of input signals are mapped to one or more output signals, and in addition to apply any modification to those input signals as mapped to an output signal. These modifications may include applying video effects, applying transitions, applying an overlay etc.

The control signal may be generated by the editing device under the control of a user of the editing device, said user observing the events being captured by the plurality of capture devices, wherein the user adjusts the control signal in accordance with a determination as to which capture device should be chosen at any instant in time, the determination identifying which of the plurality of capture streams should be selected at any instant in dependence on the users determination of the capture device or devices to select at that instant. The server may be configured to apply an overlay to the output stream in dependence on the control signal. The server may apply an overlay to a captured data stream, which captured data stream is routed to the output stream. The server may apply the overlay to at least two captured data streams which as a composite comprise an output stream.

The plurality of captured streams may capture live events, and the viewing stream is a live event stream.

The editing device may receive a representation of each or a subset of the captured data streams. The editing device may receive a lower bandwidth version of the one or more captured data streams.

An editing application may be provided on the editing device .

The editing device may be a mobile device.

The editing device may be a device of a rights holder, wherein the output stream is associated with a rights holder event. The control data may be an instruction for the server to apply or alter rights control information applied to the output stream. The rights control information may be digital rights management information.

There is provided a method for providing streaming services, comprising: generating, at a plurality of capture devices, a captured stream of content; receiving, at a server, the plurality of captured streams, and outputting one or more output streams; and outputting a control signal to the server from an editing device, wherein the server delivers the one or more outputs streams as a modified version of one or more of the input streams in dependence on the control signal.

The method may further comprise aligning the timing of the control signal with the timing of the captured streams. The control signal may comprise one or more timing marks, and the server is configured to align said one or more timing marks with a time line of received capture data streams.

The method may be provided with metadata associated with or embedded in each captured data stream, the server being configured to align the timing of the control signal additionally with the metadata of the captured data stream.

The method may automatically generate the output stream in dependence on the control signal.

The control signal may identify which of the plurality of captured streams is to be output from the server. Each captured stream may be mapped to an output.

The control signal may be generated by the editing device under the control of a user of the editing device, said user observing the events being captured by the plurality of capture devices, wherein the user adjusts the control signal in accordance with a determination as to which capture device should be chosen at any instant in time, the determination identifying which of the plurality of capture streams should be selected at any instant in dependence on the users determination of the capture device or devices to select at that instant.

The server may be configured to apply an overlay to the output stream in dependence on the control signal. The server may apply an overlay to a captured data stream, which captured data stream is routed to the output stream. The server may apply the overlay to at least two captured data streams which as a composite comprise an output stream.

The plurality of captured streams may capture live events, and the viewing stream may be a live event stream. The editing device may receive a representation of each or a subset of the captured data streams. The editing device may receive a lower bandwidth version of the one or more captured data streams. The method may provide an editing application on the editing device.

In an example there is provided a method of: receiving a plurality of data streams from a plurality of capture devices at a streaming server; receiving stream edit control signals for an edit device; processing the plurality of data streams under the control of the edit device; and in dependence on the processing, generating one or more viewing streams. This aspect may be referred to as director app.

The plurality of capture devices and the edit device are connected to the streaming server by a network. The network may be a public network, e.g. the Internet. The connection of the plurality of capture devices and the edit device to the streaming server by a network is independent connection.

The step of processing the plurality of data streams comprises receiving the plurality of data streams, and editing at least one of the plurality of data streams to provide one or more output streams.

The step of processing may comprise transmitting a control signal to a capture device. The control signal may be to manipulate the capture device. The control signal may control a parameter of the capture device.

Since the edit device is connected via a network such as the open Internet, the edit control signals may be synchronised to be applied to the video at the correct times.

Whilst synchronisation need to be maintained between streams, the edit device also needs to have its signals synchronised to those streams and/or from one set of low quality streams used for editing to the high quality streams used for viewing .

The edit device may display previews of the video content in one or more video streams received from the plurality of capture devices. Where the edit device displays previews of a plurality of streams, the edit device selects one or more streams to be viewed.

The edit device may display stored video content. The previously stored video content may be live-streamed. The edit device may edit the content for retransmission.

BRIEF DESCRIPTION OF THE FIGURES:

The invention is now described by way of reference to the following figures, in which:

Figure 1 illustrates an example of a system architecture in which described examples and embodiments as described may be implemented;

Figure 2 illustrates an example scenario for controlling captured streams; and

Figure 3 illustrates a further example scenario for controlling captured streams.

DESCRIPTION OF PREFERRED EMBODIMENTS:

With reference to Figure 1 there is illustrated a system architecture within which embodiments may be implemented.

With reference to Figure 1 there is illustrated: a plurality of devices, labelled capture devices, denoted by reference numerals 12a, 12b, 12c; a plurality of devices, labelled viewing devices, denoted by reference numerals 16a, 16b; a device, labelled editing device, denoted by reference numeral 20a; a network denoted by reference numeral 4; and a server denoted by reference numeral 2.

Each of the devices 12a, 12b, 12c is referred to as a capture device as in the described embodiments of the invention the devices capture content. However the devices are not limited to capturing content, and may have other functionality and purposes. In examples each capture device 12a, 12b 12c may be a mobile device such as a mobile phone.

Each of the capture devices 12a, 12b, 12c may capture an image utilising a preferably integrated image capture device (such as a video camera), and may thus generate a video stream on a respective communication line 14a, 14b, 14c. The respective communication lines 14a, 14b, 14c provide inputs to the network 4, which is preferably a public network such as the Internet. The communication lines 14a, 14b, 14c are illustrated as bi ¬ directional, to show that the capture devices 12a, 12b, 12c may receive signals as well as generate signals.

The server 2 is configured to receive inputs from the capture devices 12a, 12b, 12c as denoted by the bi-directional communication lines 6, connected between the server 2 and the network 4. In embodiments, the server 2 receives a plurality of video streams from the capture devices, as the signals on lines 14a, 14b, 14c are video streams.

The server 2 may process the video streams received from the capture devices as will be discussed further hereinbelow.

The server 2 may generate further video streams on bidirectional communication line 6 to the network 4, to the bi ¬ directional communication lines 18a, 18b, associated with the devices 16a, 16b respectively.

Each of the devices 16a, 16b is referred to as a viewing device as in the described embodiments of the invention the devices allow content to be viewed. However the devices are not limited to providing viewing of content, and may have other functionality and purposes. In examples each viewing device 16a, 16b may be a mobile device such as a mobile phone.

The viewing devices 16a and 16b may be associated with a display (preferably an integrated display) for viewing the video streams provided on the respective communication lines 18a, 18b.

A single device may be both a capture device and a viewing device. Thus, for example, a mobile phone device may be enabled in order to operate as both a capture device and a viewing device .

A device operating as a capture device may generate multiple video streams, such that a capture device such as capture device 12a may be connected to the network 4 via multiple video streams, with multiple video streams being provided on communication line 14a.

A viewing device may be arranged in order to receive multiple video streams. Thus a viewing device such as viewing device 16a may be arranged to receive multiple video streams on communication line 18a.

A single device may be a capture device providing multiple video streams and may be a viewing device receiving multiple video streams.

Each capture device and viewing device is connected to the network 4 with a bi-directional communication link, and thus one or all of the viewing devices 16A, 16B may provide a signal to the network 6 in order to provide a feedback or control signal to the server 2. The server 2 may provide control signals to the network 4 in order to provide control signals to one or more of the capture devices 12a, 12b, 12c.

The capture devices 12a, 12b, 12c are preferably independent of each other, and are independent of the server 2.

Similarly the viewing devices 16a, 16b are preferably independent of each other, and are independent of the server 2.

The capture devices 12a, 12b, 12b are shown in Figure 1 as communicating with the server 2 via a single network 4. In practice the capture devices 12a, 12b, 12c may be connected to the server 2 via multiple networks, and there may not be a common network path for the multiple capture devices to the server 2. Similarly the viewing devices 16a, 16b may be connected to the server 2 via multiple networks, and there may not be a single common network path from the server 2 to the viewing devices 16a, 16b.

The system architecture of Figure 1 may be used to stream live content from capture devices to the server, and then for viewing devices to access the live content in streams from the server .

Figure 2 illustrates two capture devices 500 and 502 which may correspond to the capture devices 12a, 12b, 12c of Figure 1 ; a so-called director capture device which correspond to one of the capture devices of Figure 1, or the editing device 20 of Figure 1, and a network 506 which may correspond to the network 4 of Figure 1.

In Figure 2, as shown the device 504 may receive video streams from capture devices 502, and provide a consolidated video stream to a network 506.

It will be understood that the arrangement illustrated in Figure 2 is exemplary. In embodiments the director device 504 may receive the captured streams directly. In other examples the captured streams may be provided directly to the server as described above, and the director device may provide control inputs to the server 2.

Figure 3 illustrates streaming functional modules of the server 4 of Figure 1, including an event recognition module 514 and a director module 516. Reference numeral 508 denotes the streaming functional module of the server 2.

As shown in Figure 3, the streaming server 508 may be associated with the module 514 which determines a recognition of an event, and provide a video stream as either being associated with an event or not associated with an event. In general, there is provided an application that offers features for controlling, editing, processing or otherwise managing one or more live streams. This application may be a secondary application, provided in addition to a main application associated with a service for delivering and accessing live streams.

For a given event, a publisher may use this application to take one or more input streams from capture devices and brand them as a single, coherent collection of synchronised streams. These streams may originate from devices the publisher operates, or may be crowd-sourced from consumer devices present at an event, a concert, a collection of outside broadcast locations etc .

Much of the content captured at a particular device may be configured, marshalled and controlled by a managing director application .

For a user of the system such as a content provider, a value of this application is that in creating content for viewers, there is provided a consistent set of reliable professional tools with which to shape their output. The convenience of the application being on a mobile device, such as a tablet, means that there is no specialist or expensive hardware required. Consumer-class devices can be used to provide professional grade material. This allows any event, no matter how casual, short or low-budget to benefit from the same production tools quality.

For an end user - i.e. an editor - content originating from any contributing source can be managed and controlled in a consistent way, meaning that it is easy to perform common tasks with content no matter what its origin. This frees the editor from having to be concerned with compatible frame rates, stream qualities, synchronisation issues, behaviour of overlays across transitions from one camera o another. A single contributor may be able to create an event with a number of different devices contributing live content and from which a director edits together, in real time, at least one published output stream. While there may be a single main published stream, the director may also choose one or more raw camera feeds for viewers to switch to.

A use case may be a church which wishes to broadcast an event. Two cameras are set up on the speaker, another two on the audience, and final one on a speaker at an off-site location. During the event the director publishes a stream that uses one or other shot of the speaker, interspersed with reaction shots from the congregation. For one segment of the event the director may switch to the off-site speaker. Throughout, the director allows viewers to watch either the main stream of any one of the other streams except the off-site speaker.

Another use case is where a breaking news story has brought a number of reporters to various locations around an event. A director has feeds from either of these, plus from a studio. As events unfold the director wants to bring content from the filed into the news programme by switching between the studio and one of the reporters.

For a customer, there is less reliance on a single point of view being sufficient to capture the mood or full range of content from an event. Greater flexibility is provided when comprising a published stream. If one cameras video becomes uninteresting, a switch can be made to another camera feed and all the viewers can be brought along.

For a contributor end user there is an ability to set up a range of shots rather than having to be a moving cameraman. Each camera can be configured for optimal video capture in its given location meaning it is possible to concentrate on content capture, not on by momentary setting adjustments.

For a viewer content is more interesting. There are fewer moments where action slackens off. The overall feel provided is more like television, but with the added advantage of being able to see real-time behind the scenes and/or alternative angle content .

Another example is the use of metrics that can be offered to a single contributor through the camera application or a user of a director application using the professional publishing application. This can allow maintenance of recording consistency across a number of remotely controlled devices for example.

Example metrics are ambient noise level indicator, white balance, colour saturation, light meter etc. A content capture monitor can be provided that can be used to allow a better setup and configuration or on-the-fly adjustments of capture properties .

For the customer, this brings studio quality controls to ad-hoc video capture, meaning that the same production values, planning and directorial control can be deployed allowing for smooth re-use of talent and/or recording resources.

For the contributor, a previously unavailable level of shot configuration can be provided for a professional recording session .

All the examples and embodiments described herein may be implemented as processed in software. When implemented as processes in software, the processes (or methods) may be provided as executable code which, when run on a device having computer capability, implements a process or method as described. The execute code may be stored on a computer device, or may be stored on a memory and may be connected to or downloaded to a computer device.

Examples and embodiments are described herein, and any part of any example or embodiment may be combined with any part of any other example or embodiment. Parts of example are embodiments are not limited to being implemented in combination with a part of any other example or embodiment described. Features described are not limited to being only in the combination as presented.