Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
CAMERA IMAGING SYSTEMS AND METHODS
Document Type and Number:
WIPO Patent Application WO/2012/051720
Kind Code:
A2
Abstract:
A video imaging system comprising: a low resolution colour digital video camera and a high resolution monochromatic digital video camera operably connected to a digital processing system.

Inventors:
ZHANG YUN (CA)
Application Number:
PCT/CA2011/050666
Publication Date:
April 26, 2012
Filing Date:
October 24, 2011
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
UNIV NEW BRUNSWICK (CA)
ZHANG YUN (CA)
Foreign References:
US5852502A1998-12-22
US20060125936A12006-06-15
US20020122113A12002-09-05
US4101916A1978-07-18
US7340099B22008-03-04
Other References:
See references of EP 2630802A4
Attorney, Agent or Firm:
DERENYI, Eugene (46 Elgin StreetOttawa, Ontario K1P 5K6, CA)
Download PDF:
Claims:
I claim:

\. A video imaging system comprising: a low resolution colour digital video camera and a high resolution monochromatic digital video camera operably connected to a digital processing system.

2. The system of claim 1 further comprising: an object motion module for detecting the movement of objects within the fields of view of the cameras, and an object position module for determining the position of an object in the overlapping field of view of the cameras.

3. A method comprising: providing an image frame from a low resolution colour digital video camera; providing a corresponding image frame from a high resolution monochromatic digital video camera; and fusing the two image frames to obtain a colour image having a higher resolution than the image frame from the low resolution colour digital video camera.

4. The method according to claim 3 further comprising: providing a three dimensional coordinate system for determining the position of an object in an overlap area of the fields of view of the cameras, by determining the two dimensional coordinates of the object from the position of the object on at least one of the two images, and, determining a distance from the cameras to the object in an axis perpendicular to the plane of the image frames from the parallax between the two image frames comprising the fused image.

5. The method of claim 3 wherein the image frame from the colour camera comprises a stream of video image frames and, the image frame from the monochromatic camera comprises a stream of video image frames.

6. The method of claim 5 wherein the image frames are fused using a method selected from the group consisting of fusion methods disclosed in U.S. Patent No. 7,340,099, arithmetic based, statistics based, ratio based and wavelet based fusion.

7. The method of claim 5 further comprising: isolating two image frames from at least one of the cameras; comparing the two image frames and identifying at least one difference in the two image frames.

8. The method of claim 7 further comprising: relating the at least one difference to motion of an object in the overlap area.

9. An imaging system comprising: a low resolution colour digital sensor, a high resolution monochromatic digital sensor operably connected to a digital processing system, a lens and a beam splitter for directing incident light passing through the lens to the colour and monochromatic sensors.

10. An imaging system comprising: a low resolution colour digital sensor, a high resolution monochromatic digital sensor operably connected to a digital processing system, a lens and a mirror movable between a first position for directing incident light passing through the lens to the colour sensor and a second position for directing the incident light to the monochromatic sensor.

11. A method comprising: providing an image from a low resolution colour digital sensor; providing a corresponding image from a high resolution monochromatic digital sensor; and fusing the two images to obtain a colour image having a higher resolution than the image from the low resolution colour sensor.

12. The method of claim 11 wherein the image from the colour sensor comprises a stream of video image frames and, the image from the monochromatic sensor comprises a stream of video image frames.

13. The method of claim 12 wherein the image frames are fused using a method selected from the group consisting of fusion methods disclosed in U.S. Patent No. 7,340,099, arithmetic based, statistics based, ratio based and wavelet based fusion.

14. The method of claim 12 further comprising: isolating two image frames from at least one of the cameras; comparing the two image frames and identifying at least one difference in the two image frames.

Description:
CAMERA IMAGING SYSTEMS AND METHODS

FIELD

This invention relates to camera imaging.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is a schematic diagram of a dual video image system according to one or more embodiments of the present invention;

FIG. 2 is a schematic diagram of consecutive image frames according to one or more embodiments of the present invention;

FIG. 3 is a schematic diagram of a two camera system according to one or more embodiments of the present invention;

FIG. 4 is a schematic diagram of a two sensor single lens system according to one or more embodiments of the present invention; and

Fig. 5 is a schematic diagram of a rotating half mirror system according to one or more embodiments of the present invention.

SUMMARY

According to one or more embodiments of the present invention, a video imaging system comprising a low resolution colour digital video camera and a high resolution monochromatic digital video camera operably connected to a digital processing system. The system can further comprise an object motion module for detecting objects moving within the fields of view of the cameras, and an object position module for determining the position of an object in the overlapping field of view of the cameras.

According to one or more embodiments of the present invention, a method comprising providing an image frame from a low resolution colour digital video camera and a corresponding image frame from high resolution monochromatic digital video camera and fusing the two image frames to obtain a colour image having higher resolution than the image frame from the low resolution colour digital video camera. The method can further comprise providing a three dimensional coordinate system for determining the position of a moving object in the overlapping fields of view of the cameras whereby the two dimensional position of the moving object is determined according its position in the images, whereas the distance from the cameras to the object in the axis perpendicular to the plane of the images is derived from the parallax error between the two image frames to be fused.

According to one or more embodiments of the present invention, a camera imaging system comprising a low resolution colour digital sensor chip, a high resolution monochromatic digital sensor chip, a beam splitter, and a lens, wherein the lens gathers incident light towards the beam splitter, and the beam splitter splits the light towards the two sensor chips. The system further comprises a digital processing system which fuses a low resolution colour image from the colour sensor and a high resolution monochromatic image from monochromatic sensor to produce a high resolution colour image.

DETAILED DESCRIPTION

Referring to FIG. 1, an overall system configuration for a dual video imaging system according to an embodiment of the present invention comprises a colour digital video camera 2 having lens 20 and a monochromatic digital video camera 4 having lens 22. The cameras 2 and 4 each generate a digital signal of scene 6, which is then transmitted to digital processing system ("DPS") 12.

The cameras 2 and 4 employ charge-coupled device ("CCD") sensors or complementary metal-oxide-semiconductor ("CMOS") sensors. Camera 2 is a low resolution colour ("LC") video camera while camera 4 is a high resolution monochrome ("HM") video camera. Cameras 2 and 4 are capable of providing streaming video signals as part of a security, surveillance or monitoring system. It will be understood, however, that the applications for the cameras 2 and 4 are not limited to such systems.

Camera 2 has a field of view defined by light rays 8 while camera 4 has a field of view defined by light rays 10. Colour camera 2 and monochrome camera 4 produce separate streaming video signals which are then supplied to the DPS 12. The cameras 2 and 4 are adjacent and can be housed together in a single camera housing (not shown).

The low resolution colour streaming video signals from camera 2 are fused by image fusing module ("FM") 26 in processor 12 with corresponding high resolution monochrome streaming video signals from camera 4 to produce a fused high resolution colour streaming video signal ("HC") 28. Fusing the colour and monochrome video signals provides the dual camera system with improved sensitivity capable of acquiring high resolution colour video signals under poor lighting conditions due to the inclusion of the high resolution signal from the monochrome camera and the colour signal from the colour camera.

The colour and monochrome video signals are comprised of individual image frames. Corresponding pairs of video image frames from cameras 2 and 4 are isolated and then fused. Various methods for fusing the frame pairs can be used. For example, image fusion methods for fusing a low resolution multispectral satellite images with high resolution panchromatic satellite images are known in the field of remote sensing and can be adapted to fuse video image frames from cameras 2 and 4. One such fusion method is disclosed in U.S. Patent No. 7,340,099 (Zhang) which is incorporated herein by reference in its entirety. Other image fusion methods used for satellite imagery include arithmetic based, statistics based, ratio based and wavelet based methods. By substituting colour and monochrome video image frame pairs according to the present invention for multispectral and panchromatic images respectively, prior art image fusing methods can be adapted to fuse video image frames acquired by camera 2 with video image frames acquired by camera 4.

In a further aspect, referring to FIG. 2, a moving object 30 in the scene 6 can be detected by both video cameras 2 and 4 based on finding changes in the consecutive image frames 32 of each video camera. If there is no moving object in the scene 6, the images in the consecutive frames will be the same. If there is a moving object 30 in scene 6, the images in the consecutive frames will be not the same. The changed area between two adjacent frames is the location of the moving object on the image. The changed areas can be found by comparing images in adjacent frames. Suitable conventional methods, techniques and algorithms for comparing consecutive image frames and finding changes in such image frames can be adopted for this system to find moving objects in consecutive images frames from each of the two cameras 2 and 4.

In a still further aspect, referring to FIG. 3, the position of objects Ol and 02 in the scene 6 is provided in a three dimensional coordinate system. Since cameras 2 and 4 are adjacent, not overlaid with each other, the light rays from lens 20 and lens 22 to any object in the scene 6 are not parallel. The closer the object to the two lenses, the larger the angle between the two light rays from the two lenses to the object. For example, the object 02 is closer to the two lenses 20 and 22 than is object Ol. The angle A2 is therefore larger than angle Al. The distance from lenses 20 and 22 to any object in the scene 6 can be calculated according to the base line distance between lenses 20 and 22 and the viewing angle between the two light rays. This distance gives the coordinate along the Z-axis of the three- dimensional coordinate system. Using the base line distance between the two lenses and the viewing angle between the two light rays to determine the distance from an airborne sensor to a ground object is well known in the fields of photogrammetry and computer vision. Such methods can be adapted to determine the distance to objects in the scene 6 because cameras 2 and 4 view essentially the same scene, but there is a parallax between the field of view 8 of camera 2 and the field of view 10 of camera 4.

The plane of the image frames 32 in FIG. 2 is defined by an X-Y coordinate system which is used to position objects (such as object 30) in two dimensions on the image frames 32. The X-Y axis position of an object plus its Z-axis position provides the object's position in three dimensions relative to cameras 2 and 4. The X-Y-Z positions are provided to a position calculating module in the DPS 12 which calculates the position of objects in the scene 6. The position calculating module is programmed with suitable computer algorithms based on prior art photogrammetric or computer vision methods as described above.

In one or more embodiments, the dual camera system according to the present invention provides colour video with improved sensitivity compared with a conventional video camera, the detection of moving objects, and the three dimensional position of the objects in the common field of view of the cameras 2 and 4. According to one or more embodiments of the present invention, methods of the present invention can be applied to image frames from two corresponding still cameras.

In a still further aspect, referring to FIG.4, the dual imaging system described above with initial reference to FIG. 1 is modified by replacing the dual cameras and dual lenses with a dual sensor single lens camera. Lens 22 is omitted and a light splitter 40 is added. The light beam splitter 40 splits the incoming light into two directions. Camera 2 is reconfigured with its low resolution digital colour sensor 42 towards one of the split light beams and camera 4 is re-configured with its high resolution monochrome digital sensor 44 towards the other split beam.

Cameras 2 and 4 are positioned such that when the light splitter 40 splits the incoming light into two directions, about half of the incident light is directed towards the colour digital sensor 42 and about the other half of the incident light is directed towards the monochromatic digital sensor 44. In this embodiment, the capacity of detecting distance from the camera to a moving object is reduced.

Separate streaming video signals from sensors 42 and 44 are then supplied to the DPS 12 in a similar manner to the signals from cameras 2 and 4 in the system described with initial reference to FIG. 1.

Low resolution colour streaming video signals from sensor 42 are fused by the FM 26 in processor 12 with corresponding high resolution monochrome streaming video signals from sensor 44 to produce a fused high resolution colour streaming video signal ("HC") 28 using the methods described herein.

In a still further embodiment, referring to FIG. 5, a rotating half mirror 50 or vibrating mirror (not shown) can be used in place of the splitter 40 of FIG. 4. The half mirror 50 rotates between a first position (PI) where incident light from the lens 20 passes through the empty half of the mirror directly reaching the sensor 44 and a second position (P2) where incident light passing through the lens 20 is directed by the half mirror to the sensor 42. The mirror 50 rotates between the first and second positions sufficiently quickly to obtain a suitable image pair of a scene. It is understood that other devices can be used in place of splitter 40 or a mirror 50, as long as they can direct incident light from lens 20 towards both sensor 42 and sensor 44 simultaneously.