Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
APPARATUS AND METHOD FOR GENERATING AN EXTRAPOLATED IMAGE USING A RECURSIVE HIERARCHICAL PROCESS
Document Type and Number:
WIPO Patent Application WO/2016/102355
Kind Code:
A1
Abstract:
Apparatus and Method for Generating an Extrapolated Image Using a Recursive Hierarchical Process A method and apparatus for generating an extrapolated image from an existing film or video content, which can be displayed beyond the borders of the existing film or video content to increase viewer immersiveness, are provided. The present principles provide for hierarchical processing in which higher resolution images are generated at each higher level, and wherein the higher level image is generated based on prediction and weighting derived from a current level image, and the current level is refined for the prediction based on overlapping data.

Inventors:
URBAN FABRICE (FR)
GUILLOTEL PHILIPPE (FR)
TURBAN LAURA (FR)
Application Number:
PCT/EP2015/080462
Publication Date:
June 30, 2016
Filing Date:
December 18, 2015
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
THOMSON LICENSING (FR)
International Classes:
G11B27/034; G06T3/40; G11B27/34
Domestic Patent References:
WO2004006570A12004-01-15
WO2006003600A12006-01-12
Other References:
TAMAR AVRAHAM ET AL: "Ultrawide Foveated Video Extrapolation", IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, IEEE, US, vol. 5, no. 2, 1 April 2011 (2011-04-01), pages 321 - 334, XP011350695, ISSN: 1932-4553, DOI: 10.1109/JSTSP.2010.2065213
GARY J SULLIVAN ET AL: "Overview of the High Efficiency Video Coding (HEVC) Standard", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 22, no. 12, 1 December 2012 (2012-12-01), pages 1649 - 1668, XP011486324, ISSN: 1051-8215, DOI: 10.1109/TCSVT.2012.2221191
LAURA TURBAN ET AL: "Extrafoveal Video Extension for an Immersive Viewing Experience", IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS., 31 December 2015 (2015-12-31), US, pages 1 - 1, XP055259817, ISSN: 1077-2626, DOI: 10.1109/TVCG.2016.2527649
PETER MILLS, ALIA SHEIKH, GRAHAM THOMAS, PAUL DEBENHAM, BBC BRITISH BROADCASTING CORPORATION: "Research & Development, White Paper Surround Video, WHP 208", 1 December 2011 (2011-12-01), XP055259627, Retrieved from the Internet [retrieved on 20160318]
DANIEL E NOVY ET AL: "COMPUTATIONAL IMMERSIVE DISPLAYS AP" Wgg MASSACHOSETTSIN7iiTUYE OF TECHNOLOGY LIBRARIES Certified By aniel E. Novy Program in Media Arts and Sciences", 30 June 2013 (2013-06-30), MIT, XP055259641, Retrieved from the Internet [retrieved on 20160318]
BRETT R JONES ET AL: "IllumiRoom", HUMAN FACTORS IN COMPUTING SYSTEMS, ACM, 2 PENN PLAZA, SUITE 701 NEW YORK NY 10121-0701 USA, 27 April 2013 (2013-04-27), pages 869 - 878, XP058043019, ISBN: 978-1-4503-1899-0, DOI: 10.1145/2470654.2466112
AMIT AIDES ET AL: "Multiscale ultrawide foveated video extrapolation", COMPUTATIONAL PHOTOGRAPHY (ICCP), 2011 IEEE INTERNATIONAL CONFERENCE ON, IEEE, 8 April 2011 (2011-04-08), pages 1 - 8, XP031943271, ISBN: 978-1-61284-707-8, DOI: 10.1109/ICCPHOT.2011.5753126
JANI LAINEMA ET AL: "Intra Coding of the HEVC Standard", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 22, no. 12, 1 December 2012 (2012-12-01), pages 1792 - 1801, XP011487148, ISSN: 1051-8215, DOI: 10.1109/TCSVT.2012.2221525
ROWEIS S T ET AL: "Nonlinear dimensionality reduction by locally linear embedding", SCIENCE, AMERICAN ASSOCIATION FOR THE ADVANCEMENT OF SCIENCE, US, vol. 290, no. 5500, 22 December 2000 (2000-12-22), pages 2323 - 2326, XP002406560, ISSN: 0036-8075, DOI: 10.1126/SCIENCE.290.5500.2323
Attorney, Agent or Firm:
HUCHET, Anne et al. (Issy-Les-Moulineaux, FR)
Download PDF:
Claims:
CLAIMS

1 . A method of generating an extrapolated image (720) that extends beyond a border of an image (710), comprising:

accessing (1 120) a portion of the image (814);

generating (1030) a first extrapolated image (820) based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle; generating a second extrapolated image (840) based on prediction and weighting determined from blocks (822) of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

2. The method according to claim 1 , wherein generating a first extrapolated image (820) comprises resizing the portion of the image (810) to fit a size of an extension portion to be displayed adjacent the border of the image; rotating the resized portion about a vertical axis; and blurring the resized portion.

3. The method according to claim 1 , wherein the generating a second extrapolated image (840) is performed based on prediction and weighting of overlapping blocks to be included in the second extrapolated image, including using a block matching algorithm to determine at least one matching block (822) in the image.

4. The method according to claim 3, wherein the Local Linear Embedding is used to determine the weights of the blocks to be included in the second extrapolated image (840).

5. The method according to claim 3, wherein the generating a second extrapolated image (840) includes blending the overlapping blocks included in the second extrapolated image using previously determined blocks of the second extrapolated image.

6. The method according to one of claims 1 to 5, further comprising generating a third extrapolated image based on prediction determined from the second extrapolated image (840), wherein the third extrapolated image is higher resolution than the second extrapolated image, and wherein the steps of generating of a higher resolution extrapolation image than a previous extrapolation image is repeated N number of times.

7. The method according to one of claims 1 to 6, further comprising generating extrapolated images for the top (140), left (120), and right edges (130), respectively, of the image.

8. An apparatus (1 1 1 0) for generating an extrapolated image that extends beyond a border of an image, comprising: an input (1 1 10) configured to access an image (820); and at least one processor (1 140) configured to access a portion (814) of the image; generate a first extrapolated image (820) based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle; and generate a second extrapolated image (840) based on prediction and weighting determined from blocks of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

9. The apparatus according to claim 8, wherein the processor (1 140) is configured to generate a first extrapolated image (820) by: resizing the portion of the image to fit a size of an extension portion to be displayed adjacent the border of the image; rotating the resized portion about a vertical axis; and blurring the resized portion.

10. The apparatus according to claim 8, wherein the processor (1 140) is configured to generate the extrapolated image (840) based on prediction and weighting of overlapping blocks to be included in the second extrapolated image, including using a block matching algorithm to determine at least one matching block in the image.

1 1 . The apparatus according to claim 10, wherein the processor (1 140) is configured to determine the weights of the blocks to be included in the second extrapolated image using Local Linear Embedding.

12. The apparatus according to claim 10, wherein the processor (1 140) is configured to generate a second extrapolated image (840) by blending the overlapping blocks included in the second extrapolated image using previously determined blocks of the second extrapolated image.

13. The apparatus according to one of claims 8 to 12, wherein the processor (1 140) is further configured to generate a third extrapolated image based on prediction determined from the second extrapolated image (840), wherein the third extrapolated image is higher resolution than the second extrapolated image, and repeating the process of generating a higher resolution extrapolation image than a previous extrapolation image for N number of times.

14. The apparatus according to one of claims 8 to 13, wherein the processor (1 140) is configured to generate extrapolated images for the top, left, and right edges, respectively, of the image.

15. A computer program product stored in a non-transitory computer-readable storage media (1 160), comprising computer-executable instructions for:

accessing (1020) a portion of the image (812);

generating (1030) a first extrapolated image (820) based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle; and

generating (1040-1070) a second extrapolated image (840) based on prediction and weighting determined from blocks of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

Description:
Apparatus and Method for Generating an Extrapolated Image

Using a Recursive Hierarchical Process

Technical field This invention relates to an apparatus and a method for processing images, and in particular, an apparatus and method for extrapolating an existing film or video to generate extrapolated images that extend beyond the borders of the existing film or video.

Background

This section is intended to introduce the reader to various aspects of art, which may be related to various aspects of the present invention that are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present invention. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art. Referenced Documents

[1 ] METHOD OF AND SYSTEM FOR CONTROLLING AN AMBIENT LIGHT AND

LIGHTING UNIT, WO2004006570 (A1 )— 2004-01 -15

[2] DOMINANT COLOR EXTRACTION USING PERCEPTUAL RULES TO PRODUCE AMBIENT LIGHT DERIVED FROM VIDEO CONTENT, WO2006003600 (A1 )— 2006- 01 -12

[3] P. Mills, A. Sheikh, G. Thomas, and P. Debenham. BBC research & development, white paper WHP 208 - surround video, page 34. 201 1 .

[4] D. E. Novy. Computational immersive display. 2013.

[5] Jones, Brett R; Benko, Hrvoje; Ofek, Eyal; Wilson, Andrew D (2013). "NlumiRoom:

Peripheral Projected Illusions for Interactive Experiences". 2013.

[6] A. Aides, T. Avraham, and Y. Schechner. Multiscale ultrawide foveated video extrapolation. In 201 1 IEEE International Conference on Computational Photography (ICCP), pages 1 -8. 201 1 .

[7] J. Lainema, F. Bossen, W. Han, J. min, K. Ugur. Intra Coding of the HEVC Standard.

In IEEE Transactions on Circuits and Systems for Video Technology, vol.22, no.12, pages 1792-1801 . December 2012.

[8] Sam Roweis & Lawrence Saul. Nonlinear dimensionality reduction. Image Extrapolation techniques may be used to generate extrapolated images for display beyond the borders of an existing film or video image to improve viewer immersiveness while viewing a film or video content. The image extrapolation may be performed based on content in the images to be extended. Such extension seeks to increase the field of vision by projecting content around the screen to surround the spectator with the extended images along with the main image of the film or video.

An example arrangement for displaying such extrapolated images is shown in Figure 1 . The main image from the film or video content is displayed on main screen 1 10, while the extrapolated images are displayed beyond the borders of the main screen 1 10 on display areas 120, 130 and/or 140 to provide the feeling of immersiveness.

The images may be projected onto the various display areas using different arrangements including those shown in Figures 2 and 3. In Figure 2, a single projection 250 is displayed on main screen 210, and portions of projection 250 is displayed on screens 220, 230 and/or 240. In Figure 3, arrangement 310 includes a center projection that projects the film or video content onto the center display screen, while 320 includes a left projection that projects the extrapolated image onto the left display screen and 330 includes a right projection that projects the extrapolated image onto the right display screen. Although not shown, a projection may also be provided for an above or below display screen. Various systems for projecting specific displays or images around the borders or a film or video have been proposed.

In an Ambilight System developed by Philips Corporation^ ], the distribution of colors in the image is studied [2] and a selected main color is projected around the screen using RGB LEDs mounted on the edges of the TV. Smart light bulbs can be added to the setup that can change colors according to the image displayed on the screen. The light projection is done around the TV in a limited area, but provides no detail of background nor shape nor motion impression.

A system proposed by Mills et al. [3] comprises a projection system for content acquired with 2 cameras. However, this system doesn't focus on content creation and does not apply image extrapolation to existing content. The Infinity-by-nine, developed by the MIT Media Lab [4], aims to create an immersive display for TV. The system uses three projectors and 3 screens (on the ceiling and on both sides of the room) to enhance the program displayed on a TV screen. The content on the additional screens is computed in real time from the TV content. The content is created at scene changes and updated from estimated camera motion. Because the technique relies on the motion estimation, it may not work so well for fast changing scenes, scenes with flashes, or scenes with large motion.

IHumiRoom, proposed by Microsoft [5], uses a Kinect sensor and a projector to enhance the TV content by projecting images around a television display. It was mostly developed for immersive gaming to give the illusion that what happens in the game also happens in the player's room. The system projects images directly onto the furniture and walls surrounding the television display to convey events happening beyond the main image on the screen. However, this system does not address the content creation by image extrapolation, that is extrapolating based on content of the image. Content is mostly computer generated by the game renderer.

The Ultrawide Foveated Video Extrapolation [6] extrapolates video content around the center image. The key idea of the algorithm is to use a Patch-Match algorithm in space and time dimensions (using video cubes, instead of image blocks) to extrapolate the video outside of its borders. When choosing the right patch, they also penalize patches that are likely to cause distraction (i.e. patches that add many edges outside of the video). The patch used is also selected at random among a few patches with similar cost, to avoid an unnatural repetition of the same patch. To limit computation, the search for a fitting patch is limited to a certain area in the original input video, close to the pixels that are being filled, and the search area is extended if necessary. This algorithm takes several minutes to extrapolate one frame, and as such, may be difficult to implement in real time. The algorithm uses a multi-scale approach to begin the extrapolation at a coarse resolution and refining the resolution in further steps. This technique ensures smooth extrapolation but is very computationally intensive. In addition, the penalization of edge areas does not take into account the human visual system to determine whether the edge is distracting or not. The extrapolated video can miss some relevant structures.

Summary

The present principles are directed to improvement in an apparatus and a method for extrapolating images from original film and video content.

In one respect, the present principles provide a method of generating an extrapolated image that extends beyond a border of an image, comprising: accessing a portion of the image; generating a first extrapolated image based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle; and generating a second extrapolated image based on prediction and weighting determined from blocks of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

In an embodiment according to the present principles, the steps of generating a higher resolution level based on prediction and weighting from a current level may be repeated N number of times based on the desired resolution and the computing resources available.

In another respect, the present principles provides an apparatus for generating an extrapolated image that extends beyond a border of an image, comprising: an input configured to access an image; and a processor configure to access a portion of the image, generate a first extrapolated image based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle, and generate a second extrapolated image based on prediction and weighting determined from blocks of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

In an embodiment according to the present principles, the processor is configured to repeat the process of generating a higher resolution level based on prediction and weighting from a current level N number of times based on the desired resolution and the computing resources available.

In another respect, the present principles provide a computer program product stored in a non- transitory computer-readable storage media, comprising computer-executable instructions for: accessing a portion of the image; generating a first extrapolated image based on the portion of the image, including using Intra prediction to determine an estimated angle of prediction and propagating a predicted block in the first extrapolated image at the estimated angle; and generating a second extrapolated image based on prediction and weighting determined from blocks of the first extrapolated image, wherein the second extrapolated image is higher resolution than the first extrapolated image.

In an embodiment according to the present principles, the computer-executable instructions repeat the process of generating a higher resolution level based on prediction and weighting from a current level N number of times based on the desired resolution and the computing resources available. Brief Description of the Drawings

The above-mentioned and other features and advantages of this invention, and the manner of attaining them, will become more apparent and the invention will be better understood by reference to the following description of embodiments of the invention taken in conjunction with the accompanying drawings, wherein:

Figure 1 shows an exemplary arrangement for displaying extrapolated images beyond the borders of a film and/or video display;

Figure 2 shows an exemplary arrangement for projecting a image onto various display screens;

Figure 3 shows an exemplary arrangement for projection images onto various display screens; Figure 4 illustrates an exemplary image including an original image and extrapolated images generated and displayed beyond the left and right borders of the original image;

Figure 5 illustrates an aspect of generating an extrapolated image in accordance with the present principles;

Figure 6 illustrates an aspect of generating an extrapolated image in accordance with the present principles;

Figure 7 illustrates an aspect of generating an extrapolated image in accordance with the present principles;

Figure 8 illustrates an aspect of generating an extrapolated image in accordance with the present principles; Figure 9 illustrates an aspect of block matching in an extrapolated image in accordance with the present principles;

Figure 10 illustrates an exemplary process for generating an extrapolated image in accordance with the present principles; and

Figure 1 1 illustrates an exemplary apparatus for generating and/or processing an extrapolated image in accordance with the present principles.

The description set out herein illustrates exemplary embodiments for implementing various aspects of the present principles. Such examples are not to be construed as limiting the scope of the claims in any manner. Detailed Description

The present principles relate to an improved apparatus and method for generating extrapolated images that extend beyond the borders, or edges, of an existing film or video content. The present principles advantageously provide the features of: - Prediction (or prefilling) of the extrapolation image with blurring and HEVC angular prediction used for block propagation. To estimate the angular prediction, intra mode prediction from HEVC is used. In this case, the prediction is used to propagate structures rather than its conventional use in terms of compression. HEVC Intra prediction provides an angle for predicting content based on its surrounding, and an estimate of the efficiency of the estimation (through coding cost). Another possible technique is using different tools and integrating them together (detecting the orientation of gradients, estimating the main gradient, and finding a prediction).

Hierarchical computing of the extrapolated image (computation is made at a lower resolution and adapted to the higher resolution to save computational time). Usually, computations done at lower resolution are used as prediction and computations are made once again at the current level, typically resulting in a reduced local search. The multi-resolution allows to narrow the search range around a prediction. In our case, the search is done at lower resolution, and the result is directly applied at higher resolution. Of course, this is in the case of 2 resolution levels. For more resolution levels, the result at level N is used as a prediction to narrow the search at this level and apply the result at level N-1 , recursively until level 0

- "Overlapping patches" based recursive completion. The overlapping area is used as already known data for block matching (or patch-match). So the available data to find a matching block for the current block is both the coarse estimation and the overlapping area. If no overlapping is used, every block is computed independently. There will be no consistency between neighboring block, and block artifacts will be visible. Moreover, the block artifacts are even more reduced using blending between overlapping areas of the block.

Local linear embedding on N layers (LLE) allows using multiple references to compute a block prediction. The combining weights estimated on already computed area allow a closer match for the current block. This also avoids creating mosaic artifacts because combining several predictions allows creating new content.

Figure 4 illustrates an example image display having extrapolated images 420 and 430 generated based on an original image 410. It can be seen that the respective extrapolated images are different and generated in a manner to extend the original image beyond both the left and right borders of the original image 410.

The present principles rely on hierarchical processing to generate the extrapolated image, and therefore a Gaussian multi-resolution pyramid of N levels of the input image is computed (lowest resolution at level N-1 and highest resolution at level 0). The extension of level i is done using computations made at lower resolution level i+1 .

In addition, texture propagation is done from block to block spatially, leading to some recursive texture propagation.

The pyramidal decomposition as well as the filtering process takes advantage of the human visual properties, and especially the visual properties at the periphery to provide a better, more immersive experience. In particular the extrapolated images take advantage of low resolution acuity and high sensibility to moving objects and structures.

In our explanation we will consider generating an extrapolated image to extend the image on the right side. It is to be understood that similar processing, with appropriate change in orientation, can be used to generate an extrapolated image for extending the image beyond the left edge of the image. Also, the same procedure can also be used to extend the image above and below the screen. Small adjustments in the scanning order may be made to extend the image all around the central screen.

According to the present principles, hierarchical processing is performed in which higher resolution images are generated at each higher level, and wherein the higher level image is generated based on prediction and weighting derived from a current level image, and the current level is refined for the prediction based on overlapping data.

Lowest level (level i = N-1 )

A coarse prediction of the extension is made at the lowest level of the pyramid. This process is shown in Figure 5.

To do this first prediction, an image portion 530 of the input image 510 on the right border is extracted and resized to fit extension area using bilinear interpolation. The extracted, resized image is then flipped horizontally (in order not to lose edge continuity on the border) and blurred, for example, using the normalized box filter of size 25 to produce an initial extrapolated image 520. The size of the extracted image portion 530 may be varied as desired. The portion that is extracted must be large enough to contain data (at least a few pixels), and small enough not to recognize the image. 1 /10 of the image width was found to be a good width based on the tradeoffs.

As an option, the rightmost pixels on the right edge of input image 510 can be discarded by shifting this selected area to avoid extending a black border or unwanted pixels that can appear in the border of the image.

On top of this blurry prediction, an estimate of the direction of each block on the border is computed, for example, using the HEVC angular prediction [7], where the 34 HEVC directions are used. After evaluating the angle of the block, if the predicted block is similar enough to the original block, it (the predicted block) is propagated on the extension with the estimated angle starting from the position of the original block as shown in Figure 6. Here, block 630 is propagated in direction 650, and block 640 is propagated in direction 660. This prediction aims at propagating regular structures. This process propagates structures with the correct direction which may not be the case of the mirrored coarse estimate above.

Essentially, the angular prediction is performed for each block at the right most end of the image, wherein the estimation and prediction is done on the original image and then propagated on the extrapolated image. Some rightmost pixels can be discarded to avoid wrong prediction when a black border is present on the edge of the image. These blocks can be overlapping. First the best HEVC intra prediction is estimated. The chosen mode defines the angle of the structure in the block. Then if the prediction is similar enough to the block, the block is propagated on the extension following the estimated angle. 2 methods are possible.

1 . The block is copied and pasted, following a line starting at the block and having the direction of the predicted angle. This solution can propagate details of the block, existing texture.

2. Computing the HEVC intra prediction with chosen mode (angle) for the block.

Pasting this prediction following the estimated structure direction angle. This solution propagates only main structure in the block.

Level i, i<N-1

The prediction at a level i<N-1 is the extension computed during the refining step at level i+1 . Refininq the extension for a level i, ≠0

When computing at a level i, the algorithm actually computes an extension for both levels i and i-1 while doing all the computing at level i, inspired by a similar technique as the one used for super resolution. If i>1 , the extension of level i-1 computed at level i will serve as a prediction, otherwise if i=1 the extension of level i- 1 =0 computed at level i=1 will be the final extension.

Recursive overlapping block

The predicted extension is refined by selecting blocks that will be patched, starting from a block at the top left corner of the extension half overlapping the input image to a block at the lower right corner of the extension (kind of zig-zag scanning order), with a step of x pixels along the horizontal axis and y pixels along the vertical axis (it means that extrapolated blocks overlap partly each other). Final fusion between overlapping pixels will be done by blending the pixel values to generate a blended block. Blocks are selected from left to right and top to bottom. This aspect of the present principles is illustrated in Figure 9, wherein overlapping exemplary blocks 920, 930 and 940 are processed. The processing sequentially continues for the blocks disposed along the direction of arrows 950 for extension portion 910.

It is to be understood that the values of x and y may be adjusted as desired. Lower values of x and y will result in less block effect on the final extension, while higher values of x and y, will result in faster processing.

Block Matching

The selected block at level i will be called B l and is associated with the corresponding block β ί_1 at level i-1 (co-located block). The goal is to reconstruct the block B' of the extrapolated image. In this step we find good candidates for this unknown block. We are reconstructing overlapping blocks, which means part of the block is already reconstructed (left and top part). The known part is used as a template to find nearest blocks in the reference search area. This is also known as "template matching" technique. A block matching algorithm finds the K nearest neighbors (K-NN) of the block B'. In our case, the block matching relies on the sum of the squared difference. Those K-NN will be referenced from now on as the patches Pf with 1 <j<K. Where SSD is the sum of squared difference and Ω the search window in the input. Local Linear embedding (LLE) on template at level i

A technique called LLE is used to compute the patch that refines the extension. The LLE guarantees stability in the extension because it combines various patches which prevents adding distractions or details in the extension. Details are not necessary since human peripheral vision can't detect them and it could distract the eye (i.e. force the user to turn his/her head to see the details with foveal vision) instead of improving immersion.

The weights wj minimizing the following equation are computed through LLE optimization:

Where T is a template created to take into account the pixels that were refined previously and we ignore the pixels that were only predicted but not yet refined (typically an inverse-L shape).

Alternatively, the template is not used and the whole current block is considered.

A final patch P l is obtained:

As an improvement, in order to ensure that only coherent blocks, in the sense that the blocks are good matching blocks in terms of the block matching criterion, are used, blocks Pj with an SSD value too high (SSD (B 1 , Pj)≥ q) can be rejected: From the patches Pj we deduce corresponding patches Pj '1 at the corresponding positions (co-located) at level i-1 , with the weights computed at level i, and a final patch P i_1 :

or (with rejection): Note that the weights are only computed at level i.

Alternatively the PatchMatch algorithm could be used instead of K-NN search and LLE. PatchMatch is a fast algorithm for computing dense approximate nearest neighbor correspondences between patches of two image regions. It is well-known from the computer vision community.

Figure 7 shows the original image and the coarse resolution image that was generated as described above. The process of determining the blocks of the extrapolated image in hierarchical manner is shown in Figure 8. Images 810 and 820 are associated with level i, while images 830 and 840 are associated with level i-1 . In image 820, block 822 is associated with one or more blocks 814 within search area 812. Weights are estimated to determine block 822. The position of block(s) 814 is used to determine the block(s) 832 in level i-1 . The weights determined for block 822 are then associated with block(s) 832 and used, for example, by the blending technique described below, to generate block 844. The remaining blocks are sequentially processed in a similar manner to generate the blocks to be filled in for portion 842 of the extrapolated image 840.

Blending

The final patches P l and P i_1 are blended into the extensions of levels i and i-1 respectively, to prevent visible seams or block effect, in a way similar to alpha blending. We consider blending-masks Μ[, Μ , { _1 , ½ _1 where:

Vx, y, 0 < {(x, y)≤ 1

Vx, y, 0 < M 2 l (x, y)≤ 1

Vx, y, 0 < M^ix. y)≤ 1

Vx, y, 0 < ½ _1 ( y)≤ 1

Vx, y, M[(x, y) + M 2 l (x, y) = 1 and

Vx, y, { -1 (x, y) + Mi,- 1 (x, y) = 1.

The blending equation becomes:

F l = M[ P l + M 2 l B l i_1 = { -1 ··· P i_1 + M^ 1 ··· β ί_1

Where ··· is an element-wise multiplication and F l and F l 1 the final refined blocks in the extension of levels i and i-1 respectively.

This blending is done so that pixels p xy that were not refined are taken from the new patch ( {(x, y) = 1 and M 2 l (x, y) = 0). Pixels that were already refined in B l are weighted with the pixels in P'in a way that the farther pixels are from the not-refined-area of the patch (i.e. the more pixels were already refined in a previous step), the higher the weight of the original pixel in B l is. The same goes for level i-1 .

Hierarchical computing

As soon as a level i is done, the algorithm moves on to the level i-1 until it reaches the level 0. The prediction of level i-1 is based on the extrapolated image that was just computed during the refining step at level i. The final extrapolated image is the one computed for level 0 at level 1 . It is to be understood that the number of levels processed to generate the final extrapolated image may be varied as desired. Higher number of levels will require increased amount of processing, but will generally provide a higher resoluton final image, while lower number of levels will require less amount of processing, but will generally provide a lower resolution.

The resolution of the final image also depends on the resolution of the lowest resolution level and the size ratio between 2 levels. The ratio between two levels can be any value. We can work with 2 levels and have very low resolution and HD. In sum, the point of the hierarchical process is to start from easy processing (low detail resolution) and progressively add resolution using a reliable prediction. With an increased number of levels, the robustness will be improved. The start and end resolutions and the number of levels can vary to have a trade- off between computation cost and robustness.

An exemplary process for generating an extrapolated image according to the present principles is illustrated in Figure 10. Process 1000 begins at step 1010, and first accesses a portion of an image in 1020. A coarse prediction is generated in step 1030, followed by the hierachical processing of the levels. The processing includes selecting the block to be patched in 1040, determining a block in the image based on the prediction and weighting from level i in 1050, performing a blending of the pixel values to generate the new block in 1060, and determining whether all of the blocks of the extrapolated image has been processed in 1070. If not, the process repeats steps 1040-1060 until all of the blocks have been processed. If so, the process checks whether the final level has been processed. If not, the process repeats steps 1040-1070 until the final level has been processed. If so, the process stops and the final extrapolated image has been generated.

FIG. 1 1 illustrates a block diagram of an exemplary system in which various aspects of the exemplary embodiments of the present principles may be implemented. System 1 100 may be embodied as a device including the various components described below and is configured to perform the processes described above. Examples of such devices, include, but are not limited to, personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances, and servers. System 1 100 may be communicatively coupled to other components, for example, a signal receiving apparatus, a projection system, a display system, or the like, to both receive film and video data, process the film and video data as described above, and project or display the extrapolated images with the received film and video data to provide the immersive viewer experience. The system 1 100 may include at least one processor(s) 1 140 configured to execute instructions loaded therein for implementing the various processes as discussed above. Processor(s) 1 140 may include embedded memory, input output interface and various other circuitry as known in the art. The system 1 100 may also include at least one memory 1 150 (e.g., a volatile memory device, a non-volatile memory device). System 1 100 may additionally include a storage device 1 160, which may include non-volatile memory, including, but not limited to, EEPROM, ROM, PROM, RAM, DRAM, SRAM, flash, magnetic disk drive, and/or optical disk drive. The storage device 1 160 may comprise an internal storage device, an attached storage device and/or a network accessible storage device, as non-limiting examples. Program code to be loaded onto processor(s) 1 140 to perform the various processes described hereinabove may be stored in storage device 1 160 and subsequently loaded onto memory 1 150 for execution by processor(s) 1 140. System 1 100 may also include a decryption module 1 120 configured to receive and decrypt encrypted film and/or video data. The image data generated by processor(s) 1 140 may be output via an output port to display device(s), or alternatively system 1 100 may be integrated with a display, wherein the output signal is directly applied to display the images on a display of system 1 100.

System 1 100 may include a decoder 1 130 that decodes the received input signals as necessary, which may be in compressed form in accordance with any of the known compression formats, for example, MPEG2 or H.264. Decoder 1 130 may be implemented as a separate component within system 1 100 or may be incorporated as a part of processor(s) 1 140 as known in the art. Decryption module 1 120 represents the module(s) that may be included in a device to perform the decryption functions. As is known, a device may include one or both of the encryption and decryption modules, for example, encryption may be done on a regular PC since encryption does not involve secret key so that the PC need not include secure memory for storing the input parameters (i.e., the public system parameters and the user's identity). Decryption however, requires secret keys (i.e., the decryption key) and is done in a secure device, for example a smart card. Additionally, decryption module 1 120 may be implemented as a separate element of system 1 100 or may be incorporated within processor(s) 1 140 as a combination of hardware and software as known to those skilled in the art. The system 1 100 may also include communications interface 1 1 10 that enables communication with other devices via a communications channel, either wired or wireless. The communications interface 1 1 10 may include, but is not limited to a transceiver configured to transmit and receive data from the communications channel. The communications interface may include, but is not limited to, a modem or network card and the communications channel may be implemented within a wired and/or wireless medium. The various components of system 1 100 may be connected or communicatively coupled together using various suitable connections, including, but not limited to internal buses, wires, and printed circuit boards.

The exemplary embodiments of this invention may be carried out by computer software implemented by the processor(s) 1 140 or by hardware, or by a combination of hardware and software. As a non-limiting example, the exemplary embodiments of this invention may be implemented by one or more integrated circuits. The memory 1 150 may be of any type appropriate to the technical environment and may be implemented using any appropriate data storage technology, such as optical memory devices, magnetic memory devices, semiconductor-based memory devices, fixed memory and removable memory, as non-limiting examples. The processor(s) 1 140 may be of any type appropriate to the technical environment, and may encompass one or more of microprocessors, general purpose computers, special purpose computers and processors based on a multi-core architecture, as non-limiting examples.

The foregoing has provided by way of exemplary embodiments and non-limiting examples a description of the method and systems contemplated by the inventor. It is clear that various modifications and adaptations may become apparent to those skilled in the art in view of the description. However, such various modifications and adaptations fall within the scope of the teachings of the various embodiments described above. The embodiments described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed above may also be implemented in other forms (for example, an apparatus or program). An apparatus may be implemented in, for example, appropriate hardware, software, and firmware. The methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs"), and other devices that facilitate communication of information between end-users.

Reference to "one embodiment" or "an embodiment" or "one implementation" or "an implementation" of the present principles, as well as other variations thereof, mean that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles. Thus, the appearances of the phrase "in one embodiment" or "in an embodiment" or "in one implementation" or "in an implementation", as well any other variations, appearing in various places throughout the specification are not necessarily all referring to the same embodiment.

Additionally, this application or its claims may refer to "determining" various pieces of information. Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting the information, or retrieving the information from memory.

Further, this application or its claims may refer to "accessing" various pieces of information. Accessing the information may include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.

Additionally, this application or its claims may refer to "receiving" various pieces of information. Receiving is, as with "accessing", intended to be a broad term. Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory). Further, "receiving" is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.

As will be evident to one of skill in the art, implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted. The information may include, for example, instructions for performing a method, or data produced by one of the described embodiments. For example, a signal may be formatted to carry the bitstream of a described embodiment. Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal. The formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream. The information that the signal carries may be, for example, analog or digital information. The signal may be transmitted over a variety of different wired and/or wireless links, as is known. The signal may be stored on a processor-readable medium.