Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
WIDE VIEW WHITE BALANCE
Document Type and Number:
WIPO Patent Application WO/2021/104644
Kind Code:
A1
Abstract:
An apparatus is configured to detect an activation of an image sensor, automatically capture at least one intermediate image of a scene with the image sensor, detect a capture of a main image, detect at least one illuminant from visual information of the scene associated with the captured at least one intermediate image, determine white balance coefficients associated with the detected at least one illuminant and apply the determined white balance coefficients to the captured main image. The intermediate images of the scene create a field of view of the scene that is wider than a field of view of the captured main image. The intermediate images can be used to build a three-dimensional model of the scene. This provides more information to calculate the illuminant.

Inventors:
MUUKKI MIKKO (SE)
BILCU RADU (SE)
Application Number:
PCT/EP2019/083104
Publication Date:
June 03, 2021
Filing Date:
November 29, 2019
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
HUAWEI TECH CO LTD (CN)
MUUKKI MIKKO (FI)
International Classes:
G06T5/50; H04N9/73
Foreign References:
EP3490252A12019-05-29
Attorney, Agent or Firm:
KREUZ, Georg (DE)
Download PDF:
Claims:
CLAIMS

1. An apparatus (100) comprising a processor (102), the processor (102) configured to: detect an activation of an image sensor (104); automatically capture at least one intermediate image of a scene (110) with the image sensor (104); detect a capture of a main image (112) of at least a portion of the scene (110) with the image sensor (104); detect at least one illuminant from visual information of the scene (110) associated with the captured at least one intermediate image; determine white balance coefficients associated with the detected at least one illuminant; and apply the determined white balance coefficients to the captured main image.

2. The apparatus (100) according to claim 1 wherein the processor (102) is configured to build the visual information of the scene (110) by stitching together the main image and at least one captured intermediate image to create a stitched image with a field of view that is larger than a field of view of the captured main image.

3. The apparatus (100) according to any one of claims 1 or 2 wherein the processor (102) is configured to detect the activation of the image sensor (104) from one or more of an initiation of a camera application (106) of the apparatus (100), a movement of the apparatus (100), a change of a frame view (108) of the image sensor (104), a framing of the main image, or an inability to detect an illuminant associated with the capture of the main image.

4. The apparatus (100) according to any one o f the preceding claims wherein the processor (102) is configured to create a three-dimensional model of the scene (110) from the captured at least one intermediate image; calculate at least one illuminant from the three-dimensional model of the scene (110); and provide the calculated illuminant as the detected illuminant.

5. The apparatus (100) according to claim 4, wherein the processor (102) is further configured to delete the three-dimensional model of the scene (110) and initiate a capture of new intermediate images of the scene (110) after the illuminant is detected.

6. The apparatus (100) according to any one of claims 4 or 5, wherein the processor (102) is further configured to detect a change in an orientation of the image sensor (104), delete the captured at least one intermediate image, capture at least one new intermediate image and create the three-dimensional model of the scene (110) from the captured at least one new intermediate image.

7. The apparatus (100) according to any one of the preceding claims, wherein automatically capturing the at least one intermediate image of the scene (110) with the image sensor (104) comprises creating a field of view of the scene that is wider than a field of view of the image sensor (104).

8. A method (200) comprising: detecting (202) an activation of an image sensor of an apparatus; automatically capturing (204) at least one intermediate image of a scene with the image sensor; detecting (206) a capture of a main image of at least a portion of the scene with the image sensor; detecting (208) at least one illuminant from visual information of the scene associated with the captured at least one intermediate image; determining (210) white balance coefficients associated with the detected at least one illuminant; and applying (212) the determined white balance coefficients to the captured main image.

9. The method (200) according to claim 8 further comprising creating (220) a three dimensional model or map of the scene. 10. The method (200) according to claim 9 wherein creating (220) the three dimensional model or map further comprises building the visual information of the scene by stitching (224) together the main image and at least one captured intermediate image to create (226) a stitched image with a field of view that is larger than a field of view of the captured main image. 11. The method (200) according to any one of claims 8 to 10 further comprising detecting

(202) the activation of the image sensor from one or more of an initialization (231) of a camera application of the apparatus, a manual activation (232) of a camera, a movement (233) of the apparatus, a change (234) of a frame view of the image sensor, a framing (235) of the main image, or an inability (236) to detect an illuminant associated with the capture of the main image.

12. The method (200) according to any one of claims 8 to 11 wherein creating (214) the three dimensional model or map further comprises calculating (222) at least one illuminant from the three-dimensional model or map of the scene; and using (208) the calculated illuminant as the detected illuminant.

13. The method (200) according to claim 12, further comprising deleting (242) the three- dimensional model or map of the scene and initiating (244) a capture of new intermediate images of the scene after the illuminant is detected.

14. The method (200) according to any one of claims 8 to 13 further comprising detecting (252) a change in an orientation of the image sensor, deleting (254) the captured at least one intermediate image, capturing (256) at least one new intermediate image and creating (258) the three-dimensional model or map of the scene from the captured at least one new intermediate image.

15. The method (200) according to any one of claims 8-14, wherein automatically capturing the at least one intermediate image of a scene with the image sensor further comprises creating a field of view of the scene that is wider than a field of view of the image sensor.

Description:
WIDE VIEW WHITE BALANCE

TECHNICAL FIELD

[0001] The aspects of the present disclosure relate generally to imaging devices and more particularly to detecting an illuminant in a scene. BACKGROUND

[0002] Automatic white balance (AWB) algorithms are used in imaging devices such as cameras to modify the pixel values in order to obtain images with correct colors. More precisely, these algorithms are used to correctly set the white point of the image. In white balancing, the first step is to identify the illuminant(s) and then, based on the identified illuminant, calculate and apply proper correction. Typically, white balance correction is done by global algorithms, where one offset and one gain are calculated for each color component. These offsets and gains are then used to map the pixel values to the correct ones, which ideally will match the real scene color. While this color correction is a trivial task, robust identification of the correct illuminant, even in most challenging conditions, still remains a difficult task. Failing to correctly identify the illuminant will result in erroneous estimation of the color correction parameters, which will produce images with incorrect colors.

[0003] A single color object in a scene is traditionally difficult for automatic white balance algorithms. In a scene containing only an object with a single color, such as a uniformly colored wall or grass foliage, there is no means to identify the correct color. The many combinations of illuminant spectrum and object color can produce the same pixel values into the camera sensor. For example, if a flat, single colored object has a reddish cast, it is not possible to know if this is a reddish object under illuminant 1 or a grey object under illuminant 2. In order to perform proper color processing, it needs to be understood if the illuminant is illuminant 1 or illuminant 2. Scenes having single color object are more common if the camera field of view (FOV) is narrow.

[0004] In prior art solutions, the automatic white balance algorithm may use historical information. For instance, in a case where the illuminant is not correctly identified, the last known reliable result is used. However this is just a guess since there is no guarantee that the last reliable illuminant was estimated from the current scene. For example, changing the perspective from which the scene is imaged might change the illuminant.

[0005] Accordingly, it would be desirable to be able to provide a more robust manner of illuminant identification that addresses at least some of the problems identified above. SUMMARY

[0006] The aspects of the disclosed embodiments are directed to detecting an illuminant. This object is solved by the subject matter of the independent claims. Further advantageous modifications can be found in the dependent claims.

[0007] According to a first aspect the above and further objects and advantages are obtained by an apparatus. In one embodiment, the apparatus includes a processor that is configured to detect an activation of an image sensor, automatically capture at least one intermediate image of a scene with the image sensor, detect a capture of a main image of at least a portion of the scene with the image sensor, detect at least one illuminant from visual information of the scene associated with the captured at least one intermediate image, determine white balance coefficients associated with the detected at least one illuminant and apply the determined white balance coefficients to the captured main image. The aspects of the disclosed embodiments improve automatic white balance performance by using a wider field of view for illuminant detection. The increased information obtained from the intermediate images is used by the automatic white balance algorithm to increase the robustness of illuminant detection.

[0008] In a possible implementation form of the apparatus the processor is configured to build the visual information of the scene by stitching together the main image and at least one captured intermediate image to create a stitched image with a field of view that is larger than a field of view of the captured main image. The aspects of the disclosed embodiments use neighboring scenes to create a three-dimensional model and detect the temperature of an illuminant of a captured image from the three-dimensional model.

[0009] In a possible implementation form of the apparatus the processor is configured to detect the activation of the image sensor from one or more of an initiation of a camera or camera application of the apparatus, a movement of the apparatus, a change of a frame view of the image sensor, a framing of the main image, or an inability to detect an illuminant associated with the capture of the main image. The aspects of the disclosed embodiments are directed to automatically building a three-dimensional view of a scene, and then using the three- dimensional view to detect or determine the illuminant for an image taken from the scene.

[0010] In a possible implementation form of the apparatus the processor is configured to create a three-dimensional model of the scene from the captured at least one intermediate image, calculate at least one illuminant from the three-dimensional model of the scene, and provide the calculated illuminant as the detected illuminant. The aspects of the disclosed embodiments use neighboring scenes to detect the temperature of the illuminant of a captured image.

[0011] In a possible implementation form of the apparatus the processor is further configured to delete the three-dimensional model of the scene and initiate a capture of new intermediate images of the scene after the illuminant is detected. In the system of the disclosed embodiments, the wider view is based on current information, and is not random historical information.

[0012] In a possible implementation form of the apparatus the processor is further configured to detect a change in an orientation of the image sensor, delete the captured at least one intermediate image, capture at least one new intermediate image and create the three- dimensional model of the scene from the captured at least one new intermediate image. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information. A change in orientation can indicate a new scene.

[0013] In a possible implementation form of the apparatus, automatically capturing the at least one intermediate image of a scene with the image sensor comprises creating a field of view of the scene that is wider than a field of view of the image sensor. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to detect and calculate the illuminant.

[0014] In a possible implementation form of the apparatus the automatic capture of initial images starts independently of the user’s starting of the camera application. The aspects of the disclosed embodiments are configured to gather information of the scene before the main image is captured.

[0015] In a possible implementation form of the apparatus the automatic capture of initial images is triggered by an unsuccessful illuminant detection. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information.

[0016] In a possible implementation form of the apparatus any of a number of known methods are used to build the three-dimensional model of the scene. The aspects of the disclosed embodiments use neighboring scenes to create a three-dimensional model and detect the temperature of an illuminant of a captured image from the three-dimensional model.

[0017] In a possible implementation form of the apparatus the three-dimensional model consists only of information related to the camera pose relative to the scene. The aspects of the disclosed embodiments use neighboring scenes to create a three-dimensional model and detect the temperature of an illuminant of a captured image from the three-dimensional model.

[0018] In a possible implementation form of the apparatus the captured intermediate images are used to calculate one illuminant. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information. [0019] In a possible implementation form of the apparatus the three-dimensional scene data is used to select which intermediate images are used to select the illuminant. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information.

[0020] In a possible implementation form of the apparatus, the processor is configured to detect a mixed illuminant. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information.

[0021] In a possible implementation form of the apparatus the processor is configured to build an image with a larger field of view by stitching the intermediate images. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant. The aspects of the disclosed embodiments improve automatic white balance performance by using a wider field of view for illuminant detection. The increased information from the intermediate images is used by the automatic white balance algorithm to increase the robustness of illuminant detection.

[0022] In a possible implementation form of the apparatus the stitched image is used to calculate one illuminant. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0023] In a possible implementation form of the apparatus a stitched image is used to compute a mixed illuminant. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0024] In a possible implementation form of the apparatus the already computed three- dimensional scene model is deleted and the capture of new intermediate images is started immediately after the main image is processed. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information. [0025] In a possible implementation form of the apparatus the created image with a larger field of view is deleted after the main image is processed. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information.

[0026] In a possible implementation form of the apparatus the captured intermediate images are deleted when the main image is processed. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information. [0027] In a possible implementation form of the apparatus a three-dimensional model of the scene is built from newly captured intermediate images. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information. [0028] In a possible implementation form of the apparatus a stitched image, with a larger field of view, is built from the newly captured set of intermediate images. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information.

[0029] In a possible implementation form of the apparatus the three-dimensional model of the scene and the intermediate images are used to calculate one illuminant. The intermediate images create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0030] In a possible implementation form of the apparatus the three-dimensional model of the scene and the intermediate images are used to calculate a mixed illuminant. The intermediate images create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0031] In a possible implementation form of the apparatus the obtained image with the larger field of view is used to calculate one illuminant. The intermediate images create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0032] In a possible implementation form of the apparatus the obtained image with the larger field of view is used to calculate a mixed illuminant. The intermediate images create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0033] In a possible implementation form of the apparatus the detected illuminant is used to calculate the white balance coefficients for next captured main image. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information.

[0034] In a possible implementation form of the apparatus the user has the option to stop the background capturing of the intermediate images. The aspects of the disclosed embodiments can control the capture of intermediate images used to create the three- dimensional model of the scene.

[0035] In a possible implementation form of the apparatus the user has the option to select the scene for which the intermediate images are captured. The aspects of the disclosed embodiments can control the capture of intermediate images used to create the three- dimensional model of the scene. [0036] In a possible implementation form of the apparatus the illuminant is calculated only from the main image. The aspects of the disclosed embodiments can control the capture of intermediate images used to create the three-dimensional model of the scene.

[0037] In a possible implementation form of the apparatus the failure of the illuminant estimation from main image is detected. The aspects of the disclosed embodiments can control when intermediate images are captured to create the three-dimensional model of the scene.

[0038] In a possible implementation form of the apparatus the failure of illuminant detection is communicated to the user with the option to start collecting intermediate images. The aspects of the disclosed embodiments can control when intermediate images are captured to create the three-dimensional model of the scene.

[0039] According to a second aspect the above and further objects and advantages are obtained by a method. In one embodiment, the method includes detecting an activation of an image sensor of an apparatus, automatically capturing at least one intermediate image of a scene with the image sensor, detecting a capture of a main image of at least a portion of the scene with the image sensor, detecting at least one illuminant from visual information of the scene associated with the captured at least one intermediate image, determining white balance coefficients associated with the detected at least one illuminant and applying the determined white balance coefficients to the captured main image.

[0040] In a possible implementation form of the method, the method further includes building the visual information of the scene by stitching together the main image and at least one captured intermediate image to create a stitched image with a field of view that is larger than a field of view of the captured main image. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to detect or calculate the illuminant.

[0041] In a possible implementation form of the method the method further includes detecting the activation of the image sensor from one or more of an initiation of a camera or camera application of the apparatus, a movement of the apparatus, a change of a frame view of the image sensor, a framing of the main image, or an inability to detect an illuminant associated with the capture of the main image. The aspects of the disclosed embodiments are directed to automatically building a three-dimensional view of a scene, and then using the three- dimensional view to detect or determine the illuminant for an image taken from the scene. [0042] In a possible implementation form of the method, the method further includes creating a three-dimensional model of the scene from the captured at least one intermediate image, calculating at least one illuminant from the three-dimensional model of the scene and providing the calculated illuminant as the detected illuminant. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant.

[0043] In a possible implementation form of the method, the method further includes deleting the three-dimensional model of the scene and initiating a capture of new intermediate images of the scene after the illuminant is detected. The aspects of the disclosed embodiments rely upon current information to calculate the illuminant rather than historical information.

[0044] In a possible implementation form of the method, the method further includes detecting a change in an orientation of the image sensor, deleting the captured at least one intermediate image, capturing at least one new intermediate image and creating the three- dimensional model of the scene from the captured at least one new intermediate image. The aspects of the disclosed embodiments use current information of the scene to calculate the illuminant and not historical information. A change in orientation can indicate a new scene.

[0045] In a possible implementation form of the method, automatically capturing the at least one intermediate image of a scene with the image sensor further comprises creating a field of view of the scene that is wider than a field of view of the image sensor. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant. The aspects of the disclosed embodiments improve automatic white balance performance by using a wider field of view for illuminant detection. The increased information from the intermediate images is used by the automatic white balance algorithm to increase the robustness of illuminant detection.

[0046] These and other aspects, implementation forms, and advantages of the exemplary embodiments will become apparent from the embodiments described herein considered in conjunction with the accompanying drawings. It is to be understood, however, that the description and drawings are designed solely for purposes of illustration and not as a definition of the limits of the disclosed invention, for which reference should be made to the appended claims. Additional aspects and advantages of the invention will be set forth in the description that follows, and in part will be obvious from the description, or may be learned by practice of the invention. Moreover, the aspects and advantages of the invention may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the appended claims.

BRIEF DESCRIPTION OF THE DRAWINGS [0047] In the following detailed portion of the present disclosure, the invention will be explained in more detail with reference to the example embodiments shown in the drawings, in which:

[0048] Figures 1 illustrates a schematic block diagram of an exemplary apparatus incorporating aspects of the disclosed embodiments. [0049] Figure 2 is a flow chart illustrating aspects of an exemplary method incorporating aspects of the disclosed embodiments.

[0050] Figure 3 is a flow chart illustrating aspects of an exemplary method incorporating aspects of the disclosed embodiments. [0051] Figure 4 is a flow chart illustrating aspects of an exemplary method incorporating aspects of the disclosed embodiments.

[0052] Figure 5 is a flow chart illustrating aspects of an exemplary method incorporating aspects of the disclosed embodiments. [0053] Figure 6 is a flow chart illustrating aspects of an exemplary method incorporating aspects of the disclosed embodiments.

[0054] Figure 7 illustrates a schematic block diagram of an exemplary apparatus that can be used to practice aspects of the disclosed embodiments.

DETAILED DESCRIPTION OF THE DISCLOSED EMBODIMENTS [0055] Referring to Figure 1, a schematic block diagram of an exemplary apparatus

100 incorporating aspects of the disclosed embodiments is illustrated. The apparatus 100 is configured to capture at least one intermediate image of a scene 110 and determine the white balance coefficients to apply to a captured main image based on an illuminant associated with the at least one intermediate image of the scene 110. The aspects of the disclosed embodiments improve Automatic White Balance (AWB) performance by using a wider Field of View (FOV) for illuminant detection. The intermediate images of a scene create a field of view of the scene that is wider than a field of view of the image sensor, which provides more information to calculate the illuminant. The increased information from the intermediate images is used by the automatic white balance algorithm to increase the robustness of illuminant detection. In the system of the disclosed embodiments, the wider view is based on current information, and not random historical information.

[0056] As is illustrated in Figure 1, an apparatus 100 comprises a processor 102. In one embodiment, the processor 102 is configured to detect an activation of an image sensor 104 and automatically capture at least one intermediate image of a scene with the image sensor 104. The activation of the image sensor 104 can occur automatically or manually.

[0057] For example, in one embodiment, the apparatus 100 is or includes a camera device or camera application 106. For the purposes of the description herein the camera device or camera application 106 will generally be referred to as “camera 106” herein. The term “camera device” or “camera application” is generally used herein to refer to any apparatus or device that includes a camera or imaging device, such as a mobile communication device, smartphone or other computing device. The user starts, or otherwise initializes or activates, the camera 106 of the apparatus 100. In one embodiment, the initialization of the camera 106 can cause the image sensor 104 to automatically capture images, referred to herein as “intermediate images.” The intermediate images can be images of a scene that appear within the field of view of the view finder of the camera 106.

[0058] As an example, in one embodiment, the user initializes and aims the camera

106. During the aiming of the camera 106, the image sensor 104 is configured to capture intermediate images.

[0059] As another example, during the aiming of the camera 106, the framing can change or the user may zoom to an object 112 in the scene 110. When an action such as a change in framing or change in zoom is detected by the processor 102, the image sensor 104 is configured to automatically capture the intermediate images. As will be described further herein, the intermediate images are used to form or create a three-dimensional model or image of the scene 110. The three-dimensional model can then be used to detect an illuminant for a captured main image of a portion 112 of the scene 110.

[0060] The processor 102 is configured to detect a capture of a main image of at least a portion of the scene with the image sensor 104 and detect at least one illuminant from visual information of the scene 110 associated with the captured intermediate images. In one embodiment, when processing the captured main image, the processor 102 is configured to run an automatic white balance algorithm. The automatic white balance algorithm can be run for example, with panorama statistics from the created three-dimensional image of the scene 110. As the panorama statistics have more content than the camera field of view when image 112 was captured, the results for determining the illuminant should be more reliable.

[0061] As another example, in one embodiment, when processing the captured main image, the processor 102 can be configured to run the algorithm with single camera field of view statistics. However, the results may not be reliable. Using the three-dimensional model of the scene created from the intermediate images as disclosed herein, a reliable automatic white balance result can be obtained even if the intermediate images are fully or partially outside of the field of view of the captured main image.

[0062] In one embodiment, the processor 102 is configured to specify which areas of the three dimensional model outside of the current camera field of view are relevant with respect to the current camera field of view. The processor 102 may also be configured to use additional information from the created three dimensional model to understand which light sources provide light, directly or via reflective surfaces, to the camera field of view.

[0063] The processor 102 is configured to determine the white balance coefficients associated with the detected at least one illuminant and apply the determined white balance coefficients to the captured main image. Being able to understand the scene outside of the camera field of view improves automatic white balance performance, making it more robust and with better colors.

[0064] The aspects of the disclosed embodiments are configured to use neighboring scenes to create the three-dimensional map or model of a scene and detect the temperature of an illuminant of a captured image from the three-dimensional model. One system for building the three-dimensional model is referred to as Simultaneous Localization and Mapping (SLAM). SLAM is the generic term for an exemplary system configured to build three dimensional models of an environment and, at the same time, localize the device, running the SLAM algorithms, into the mapped scene. Although SLAM is referenced herein, the aspects of the disclosed embodiments can utilize any suitable three-dimensional mapping algorithm. SLAM can be utilized in many applications from three-dimensional modeling to Augmented Reality (AR) and Virtual Reality (VR). In one embodiment, the SLAM elements can be used with the apparatus 100 to increase the accuracy of the illuminant detection. [0065] The aspects of the disclosed embodiments are configured to run a simplified three-dimensional mapping process whenever the camera application 106 of the apparatus 100 is initialized. While the user is framing the picture to be taken, for example, a mapping process such as SLAM will calculate and update a simplified version of the scene, by using viewfinder images. The picture to be taken, referred to herein as the captured main image, is an aspect of the scene.

[0066] In one embodiment, each viewfinder image can also be checked to determine if the user has changed the orientation of the camera 106 or apparatus 100 so that a new scene, previously unseen is now imaged. This can be referred to as the reset point. The processor 102 is configured to build a simplified three-dimensional model of the scene 110 for each position and/or orientation of the camera 106, or the camera view finder. In this manner, there is more information that can be utilized in the automatic white balance algorithm, other than just the current frame.

[0067] In one embodiment, from the previous frames used in building the three dimensional model of the current scene 110, statistics and/or illuminant information can be extracted and stored. Thus, whenever a frame is detected for which the illuminant cannot be identified, statistics and/or illuminant information detected previously from the same scene are used. In this manner, it is ensured that the illuminant is always detected from the same or current scene and not from random historical data. [0068] Moreover when a change in scene is detected, historical data is not used since the user might have changed the position and orientation of the camera 106 or apparatus 100 towards a completely different scene. In addition, for each intermediate frame used to build the three-dimensional model of the scene 110, the approximate spatial positions of the pixels and their detected illuminant is known. In this manner, it is possible to build a more precise illuminant model, particularly in a mixed illuminant scene.

[0069] Referring to Figure 2, one example of a method 200 incorporating aspects of the disclosed embodiments is illustrated. In this example, the initialization of an image sensor of an apparatus is detected 202. The apparatus could include or comprise a camera device. In one embodiment, as illustrated in Figure 3, the initialization 202 of the image sensor could comprise one or more of an initialization 231 or opening of a camera application, manual activation 232 of the camera or camera application, a movement 233 of the camera, a change 234 in framing of the camera, a change 235 in focus or zoom, or an inability 236 to detect an illuminant associated with the capture of the main image. In alternate embodiments, any suitable mechanism can be used to initiate the capture of the intermediate images as described herein.

[0070] For example, the aspects of the disclosed embodiments are configured to enable the capture of intermediate images from the time the camera 106 is moved. Alternatively, the capture of intermediate images does not occur until the camera 106 is turned on, or a camera application of the apparatus 100 is activated. [0071] In one embodiment, the apparatus 100 is a smart phone with a camera application 106 and image sensor 104. It is detected that the user picks up or otherwise moves the smart phone. Intermediate image capture can start when the smart phone is moved, or when the camera application 106 of the smart phone is initialized or opened. [0072] At least one intermediate image of the scene is captured 204. An intermediate image is an image of a scene that is within the field of view of the viewfinder of the camera 106 at the time the image is captured. The aspects of the disclosed embodiments enable the capture of one or more intermediate images of the scene. For example, in one embodiment, the main image is stitched with the first intermediate image and the illuminant is detected. If the illuminant is not detected, another intermediate image is captured and stitched with the combined image. The combined image is the result of the previous stitching.

[0073] In one embodiment, it is detected 206 that a main image of the scene is captured.

The main image of the scene is typically the image desired to be captured by the user.

[0074] At least one illuminant from the visual information of the scene that is associated with or determined from the intermediate image is detected 208. As will be further described below, this can also include one or more illuminants determined from a three- dimensional map or model of the scene created by the captured intermediate images. The white balance coefficients associated with the detected at least one illuminant are determined 210. The determined white balance coefficients are then applied 212 to the captured main image. [0075] In one embodiment, when the intermediate images are captured, a three- dimensional model or map of the scene is created 220. Referring to Figure 4, in one embodiment, the creation 220 of the three-dimensional model or map of the scene can include for example, stitching 224 together at least two of the captured intermediate images to create 226 a stitched image. The stitched image will have a field of view that is larger than a field of view of the captured main image. In this example, the calculated 222 illuminant(s) can be determined from the stitched image.

[0076] As an example, in one embodiment, a SLAM three-dimensional mapping process is used to build the three-dimensional model. This allows the camera to see a wider field of view than it would otherwise. This three-dimensional model enables providing reliable current and historical information for automatic white balance system needs.

[0077] The aspects of the disclosed embodiments can be used improve reliability in detecting an illuminant for a scene. For example, if the illuminant detection method of the disclosed embodiments is always used, reliability of illuminant detection will improve due to the increased amount of information provided by the three-dimensional model. Alternatively, the aspects of the disclosed embodiments might only be utilized when a difficult scene is detected. A difficult scene can be one in which the illuminant cannot be detected from the captured main image. An example of this is the capture of a picture of a unique object with constant color. In this case the illuminant detection from the main image will fail. By capturing intermediate images and creating a three-dimensional model, as described herein, the field of view increases and illuminant detection is more reliable.

[0078] The aspects of the disclosed embodiments rely upon current information related to the scene rather than historical information. In one embodiment, referring to Figure 5, the method includes deleting 242 the three-dimensional model of the scene and initiating 244 a capture of new intermediate image(s) of the scene after the illuminant is detected. In this manner, the scene information is kept up to date.

[0079] In one embodiment, referring to Figure 6, after a three-dimensional model is created and the illuminant detected, a change in an orientation of the image sensor is detected. Any captured intermediate images are deleted 254. At least one new intermediate image is captured 256 and a three-dimensional model of the scene is created from the captured at least one new intermediate image. In this manner, the information of the scene is kept up to date and old information discarded.

[0080] Figure 7 illustrates a block diagram of an exemplary apparatus 1000 appropriate for implementing aspects of the disclosed embodiments. The apparatus 100 referred to in Figure 1 can be a part of or comprise the apparatus 1000 of Figure 7. In the example of Figure 7, the apparatus 1000 includes or is coupled to a processor or computing hardware 102, a memory 1004, a radio frequency (RF) unit 1006 and a user interface (UI) 1008. An image sensor device 104 is also connected to the processor 102. Although not shown in Figure 7, the image sensor device 104 could include, be comprised of, or otherwise be connected to, the camera 106.

[0081] The processor 1002 may be a single processing device or may comprise a plurality of processing devices including special purpose devices, such as for example, digital signal processing (DSP) devices, microprocessors, graphics processing units (GPU), specialized processing devices, or general purpose computer processing unit (CPU). The processor 102 often includes a CPU working in tandem with a DSP to handle signal processing tasks. The processor 102 is configured to implement any one or more of the methods described herein.

[0082] In the example of Figure 7, the processor 102 is configured to be coupled to a memory 1004 which may be a combination of various types of volatile and non-volatile computer memory such as for example read only memory (ROM), random access memory (RAM), magnetic or optical disk, or other types of computer memory. The memory 1004 is configured to store computer program instructions that may be accessed and executed by the processor 102 to cause the processor 102 to perform a variety of desirable computer implemented processes or methods such as the methods as described herein.

[0083] The program instructions stored in memory 1004 are organized as sets or groups of program instructions referred to in the industry with various terms such as programs, software components, software modules, units, etc. Each module may include a set of functionality designed to support a certain purpose. Also included in the memory 1004 are program data and data files which may be stored and processed by the processor 102 while executing a set of computer program instructions.

[0084] The apparatus 1000 can also include or be coupled to an RF Unit 1006 such as a transceiver, coupled to the processor 1002 that is configured to transmit and receive RF signals based on digital data 1012 exchanged with the processor 102 and may be configured to transmit and receive radio signals with other nodes in a wireless network. In certain embodiments, the RF Unit 1006 includes receivers capable of receiving and interpreting messages sent from satellites in the global positioning system (GPS) and work together with information received from other transmitters to obtain positioning information pertaining to the location of the computing device 1000. To facilitate transmitting and receiving RF signals the RF unit 1006 includes an antenna unit 1010 which in certain embodiments may include a plurality of antenna elements.

[0085] The UI 1008 may include one or more user interface elements such as a touch screen, keypad, buttons, voice command processor, as well as other elements adapted for exchanging information with a user. The UI 1008 may also include a display unit configured to display a variety of information appropriate for a computing device or mobile user equipment and may be implemented using any appropriate display type such as for example organic light emitting diodes (OFED), liquid crystal display (FCD), as well as less complex elements such as LEDs or indicator lamps. For example, in one embodiment, an icon for the camera application 106 can be presented on the UI 1008.

[0086] Thus, while there have been shown, described and pointed out, fundamental novel features of the invention as applied to the exemplary embodiments thereof, it will be understood that various omissions, substitutions and changes in the form and details of devices and methods illustrated, and in their operation, may be made by those skilled in the art without departing from the spirit and scope of the presently disclosed invention. Further, it is expressly intended that all combinations of those elements, which perform substantially the same function in substantially the same way to achieve the same results, are within the scope of the invention. Moreover, it should be recognized that structures and/or elements shown and/or described in connection with any disclosed form or embodiment of the invention may be incorporated in any other disclosed or described or suggested form or embodiment as a general matter of design choice. It is the intention, therefore, to be limited only as indicated by the scope of the claims appended hereto.