Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
AUTOFOCUS IMAGING CIRCUITRY, AUTOFOCUS IMAGING APPARATUS, AND AUTOFOCUS IMAGING METHOD FOR AN EVENT CAMERA
Document Type and Number:
WIPO Patent Application WO/2021/228463
Kind Code:
A1
Abstract:
The present disclosure generally pertains to autofocus imaging circuitry, configured to: obtain, for a first adjusted focus, a first plurality of brightness change events from a brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

Inventors:
SCHEPER KIRK (DE)
DÜRR PETER (DE)
ERDLER OLIVER (DE)
Application Number:
PCT/EP2021/057767
Publication Date:
November 18, 2021
Filing Date:
March 25, 2021
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
SONY GROUP CORP (JP)
SONY EUROPE BV (GB)
International Classes:
G02B7/38; H04N5/232
Foreign References:
US20170337469A12017-11-23
CN109151315A2019-01-04
EP3809692A22021-04-21
Other References:
GALLEGO GUILLERMO ET AL: "Focus Is All You Need: Loss Functions for Event-Based Vision", 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), IEEE, 15 June 2019 (2019-06-15), pages 12272 - 12281, XP033687566, DOI: 10.1109/CVPR.2019.01256
BASIL HUBER: "Department of Informatics High-Speed Pose Estimation using a Dynamic Vision Sensor", MASTER THESIS, 1 March 2014 (2014-03-01), pages 1 - 60, XP055162714, Retrieved from the Internet [retrieved on 20150116]
DANIEL GEHRIG ET AL: "Asynchronous, Photometric Feature Tracking using Events and Frames", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 25 July 2018 (2018-07-25), XP081118977, DOI: 10.1007/978-3-030-01258-8_46
Attorney, Agent or Firm:
MFG PATENTANWÄLTE (DE)
Download PDF:
Claims:
CLAIMS

1. Autofocus imaging circuitry, configured to: obtain, for a first adjusted focus, a first plurality of brightness change events from a bright ness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

2. The autofocus imaging circuitry of claim 1, wherein the first and the second plurality of brightness change events is indicative of an event density.

3. The autofocus imaging circuitry of claim 2, further configured to: determine the event density.

4. The autofocus imaging circuitry of claim 3, further configured to: determine a contrast based on the determined event density.

5. The autofocus imaging circuitry of claim 1, further configured to: track an object based on the determined focus.

6. The autofocus imaging circuitry of claim 1, wherein the determination of the focus is further based on a neural network processing the first and the second plurality of brightness change events.

7. The autofocus imaging circuitry of claim 6, wherein the neural network includes a spiking neural network.

8. The autofocus imaging circuitry of claim 1, further configured to: obtain at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern.

9. The autofocus imaging circuitry of claim 1, wherein the determination of the focus is based on a feedback loop based on the first and second plurality of brightness change events.

10. An autofocus imaging apparatus, comprising: an imaging device including a plurality of brightness change event detection elements and a plurality of imaging elements; and autofocus imaging circuitry, configured to, for each brightness change event detection ele ment: obtain, for a first adjusted focus, a first plurality of brightness change events from the brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

11. The autofocus imaging apparatus of claim 10, the autofocus imaging circuitry being further configured to, for each brightness change event detection element: determine a brightness change event density for determining a spatial brightness change event density for the plurality of brightness change event elements.

12. The autofocus imaging apparatus of claim 10, wherein the imaging device has a hybrid sen sor with a predetermined pattern of the brightness change event detection elements and the plurality of imaging elements.

13. An autofocus imaging method, comprising: obtaining, for a first adjusted focus, a first plurality of brightness change events from a brightness change event detection element; obtaining, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determining a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

14. The autofocus imaging method of claim 13, further comprising: determining an event density based on the first and the second plurality of brightness change events.

15. The autofocus imaging method of claim 14, further comprising: determining a contrast based on the determined event density.

16. The autofocus imaging method of claim 13, further comprising: tracking an object based on the determined focus.

17. The autofocus imaging method of claim 13, wherein the determination of the focus is fur ther based on a neural network processing the first and the second plurality of brightness change events.

18. The autofocus imaging method of claim 17, wherein the neural network includes a spiking neural network.

19. The autofocus imaging method of claim 13, further comprising: obtaining at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern.

20. The autofocus imaging method of claim 13, wherein the determination of the focus is based on a feedback loop based on the first and second plurality of brightness change events.

Description:
AUTOFOCUS IMAGING CIRCUITRY, AUTOFOCUS IMAGING APPARATUS, AND AUTOFOCUS IMAGING METHOD FOR AN EVENT CAMERA

TECHNICAL FIELD

The present disclosure generally pertains to autofocus imaging circuitry, an autofocus imaging appa- ratus, and an autofocus imaging method.

TECHNICAL BACKGROUND

Generally, it is known to automatically adapt a focus of a lens (stack) of an imaging system (e.g. a camera), also known as autofocusing. For example, known mirrorless cameras typically have passive autofocus by using phase detection or contrast detection.

In phase detection, dedicated focus pixels are utilized to determine the focus at a frequency, which is decoupled from the main imaging frequency.

In contrast detection, it is swept over a range of focal lengths to determine an optimum focus point.

Although there exist techniques for autofocusing, it is generally desirable to provide autofocus imag- ing circuitry, an autofocus imaging apparatus, and an autofocus imaging method.

SUMMARY

According to a first aspect, the disclosure provides autofocus imaging circuitry, configured to: ob- tain, for a first adjusted focus, a first plurality of brightness change events from a brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

According to a second aspect, the disclosure provides an autofocus imaging apparatus, comprising: an imaging device including a plurality of brightness change event detection elements and a plurality of imaging elements; and autofocus imaging circuitry, configured to, for each brightness change event detection element: obtain, for a first adjusted focus, a first plurality of brightness change events from the brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for au- tofocusing a scene.

According to a third aspect, the disclosure provides an autofocus imaging method, comprising: ob- taining, for a first adjusted focus, a first plurality of brightness change events from a brightness change event detection element; obtaining, for a second adjusted focus, a second plurality of bright- ness change events from the brightness change event detection element; and determining a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

Further aspects are set forth in the dependent claims, the following description and the drawings.

BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments are explained by way of example with respect to the accompanying drawings, in which:

Fig. 1 depicts a block diagram of an autofocus imaging apparatus according to the present disclo- sure;

Fig. 2 depicts a block diagram of a further embodiment of an autofocus imaging apparatus accord- ing to the present disclosure;

Fig. 3 depicts an embodiment of a probabilistic filter for determining a focus based on an event den- sity;

Fig. 4 depicts a schematic view of a hybrid sensor according to the present disclosure;

Fig. 5 depicts an autofocus imaging method according to the present disclosure;

Fig. 6 depicts a further embodiment of an autofocus imaging method according to the present dis- closure;

Fig. 7 depicts a further embodiment of an autofocus imaging method according to the present dis- closure, and

Fig. 8 depicts a further embodiment of an autofocus imaging method according to the present dis- closure.

DETAILED DESCRIPTION OF EMBODIMENTS

Before a detailed description of the embodiments under reference of Fig. 1 is given, general explana- tions are made.

As mentioned in the outset, known imaging systems or cameras may utilize one of phase detection or contrast detection in order to autofocus.

However, phase detection requires lenses with a comparably high numerical aperture. Therefore, phase detection is rather used in large imaging systems (e.g. microscopes, DSLR (digital single-lens reflex) camera) and not suitable for small devices, such as mobile phones. On the other hand, contrast detection can generally be used regardless of the numerical aperture of the lens. However, a full range of focal lengths of a used imaging system is typically swept in order to autofocus. Therefore, contrast detection is limited by the measurement rate of the used image sensor and the velocity and accuracy of the used motor, which adjusts the focal length by changing the position of the focus lens(es). Hence, contrast detection may be considered as time inefficient.

Moreover, phase detection as well as contrast detection require image processing in order to focus on a moving object (commonly known as focus tracking or AI (artificial intelligence) servoing). However, the image processing in known systems is therefore limited by the hardware which is used for image processing. For example, in small devices, (e.g. mobile phones), in which a battery power is limited and is therefore desired to be saved, e.g. by a user of the device, the processing power typi- cally needs to be kept small, or conversely need a large battery, if the processing is performed on the device, which is typically undesirable.

For these reasons, especially in a scene with a (fast) moving object, known systems may not be able to focus correctly. The same may be true for difficult lighting conditions (e.g. low ambient light, night, dark room, etc.) or for a scene with a low contrast (e.g. similar colors of different objects).

Therefore, some embodiments pertain to autofocus imaging circuitry, configured to: obtain, for a first adjusted focus, a first plurality of brightness change events from a brightness change event de- tection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

Autofocus imaging circuitry according to the present disclosure may include any known circuitry, e.g. one or multiple processors (e.g. CPU (Central Processing Unit), GPU (Graphic Processing Unit), FPGA(s) (Field Programmable Gate Array(s)), a neuromorphic computing platform, and the like. Moreover, autofocus imaging circuitry may include a system of computers, servers, and the like, as it is generally known.

Such autofocus imaging circuitry may be configured to obtain, for a first (and/ or second) adjusted focus, a first (and/ or second) plurality of brightness change events from a brightness change event detection element.

For example, autofocus imaging circuitry may be adopted in a camera system including a lens system for focusing incoming light, as it is generally known. The lens system may be adjusted to focus on the first (or second) focus (or focal plane, focal point, and the like), such that light being reflected at or deriving from the adjusted focal plane may be bundled and light from other focal planes may not be bundled, as it is generally known. If a change of the amount of light in the focal plane occurs, this change may be detected by the brightness change event detection element, if the change is above a predetermined threshold of the amount of light.

For example, in the first adjusted focus, a certain amount of light may be incoming on a brightness change event detection element, thereby generating the first plurality of brightness change events, whereas in the second adjusted focus, which is adapted based on a feedback-loop based on the first plurality of brightness change events, a(nother) certain amount of light may be incoming on the brightness change event detection element, thereby generating the second plurality of brightness change events. The focus may be determined (and autofocus may be achieved) based on the second plurality of brightness change events, which are in turn based on the second adjusted focus, which is based on the first plurality of brightness change events, such that the focus in total is determined based on the first and the second plurality of brightness change events. However, in some embodiments, the sec- ond plurality of brightness change events is fed to the autofocus imaging circuitry in ways of a feed- back-loop, such that an iterative method is established to further refine the determination of the focus.

The amount of light detected by an(y) imaging element may be represented as brightness. Hence, if a change in brightness above a predetermined threshold in the adjusted focus is detected, the bright- ness change event detection element may generate an event indicating whether a positive change of brightness (e.g. brighter) or a negative change of brightness (e.g. darker) is detected, without limiting the present disclosure in that regard, since only absolute values of brightness changes may be de- tected, in some embodiments.

Such a brightness change event detection element may be configured of (a pixel) of an event camera, a neuromorphic camera, a silicon retina, a dynamic vision sensor (DVS), and the like. For example, a brightness change event detection element may be a part of an optical sensor (e.g. a DVS), i.e. a pixel or a group of pixels thereof.

Each (group of) pixel(s) of the DVS may independendy, asynchronously produce a sparse, time-re- solved measurement of the change of perceived brightness.

When a pixel measures a change of brightness, which is larger than a predetermined threshold, it generates an event including the (absolute or relative) time, at which the change was detected (e.g. with microsecond resolution), as well as the polarity of the change (positive (i.e. brighter) or negative (i.e. darker)). However, in some embodiments, it is not necessary to determine the polarity of the event. In some embodiments, the brightness change event detection element may be further config- ured to determine an absolute intensity of the light.

Such a DVS may have a dynamic range above hundred and twenty decibels, for example, a sensitiv- ity of less than fifteen percent with respect to a minimum contrast ratio, and a latency below hun- dred microseconds, without limiting the present disclosure to any of these properties.

However, such a configuration may maximize a speed while reducing redundant data at the same time for a predetermined lighting scenario.

Since redundant information is minimized, the processing power can be minimized as well. Thereby, focus tracking, for example, may be implemented in a time-efficient manner making it suitable for devices with a limited computational capability (e.g. a mobile phone)

If a lighting condition (e.g. light intensity) is roughly constant, according to the present disclosure, a (temporal) change in intra-scene contrast may be identified. Based on this, a motion estimation, ob- ject tracking, video data compression, gesture recognition, and the like may be performed.

As already discussed, a first and a second focus may be adjusted, and therefore, a first plurality of brightness change events for the first adjusted focus (however, it may generally be possible to detect no brightness change events at all or only one) and a second plurality of brightness change events for the second adjusted focus (or none or only one) may be obtained by the autofocus imaging cir- cuitry according to the present disclosure by evaluating data provided (directly or indirectly) from the brightness change event detection element.

Based on the first and second plurality of brightness change events, the autofocus imaging circuitry may be configured to determine a focus.

For example, if a scene (e.g. an object) is to be imaged, the focal plane in which the scene is localized may be found by comparing the number of brightness change events of the first and the second plu- rality of brightness change events.

It may be envisaged that a lower number of brightness change events indicates a better focal plane (without limiting the present disclosure in that regard), such that a minimum of events may indicate the focus corresponding to the first or the second adjusted focus.

Moreover, it may be envisaged to perform a focus sweep, i.e. adjusting a plurality of focuses and ob- taining for each adjusted focus a plurality of brightness change events in order to generate a bright- ness change curve, from which the ideal focus can be determined by determining a minimum of the curve, and the like. Hence, autofocus imaging circuitry according to the present disclosure may be able to determine an autofocus based on an event sensing in different adjusted focuses, such that it may be autofocused on a scene (e.g. an object (of interest)) based on the determined focus.

In some embodiments, the first and the second plurality of brightness change events is indicative of an event density.

That is, for each adjusted focus, a corresponding brightness change event may be determined, such that the brightness change events may be expressed as a function of the adjusted focus, thereby re- sulting in an event density, which may, in some embodiments, be determined by the autofocus imag- ing circuitry in order to determine the focus, e.g. based on a minimum of the event density, without limiting the present disclosure in that regard.

However, in some embodiments, the focus may be known as a function of time and the event den- sity may also be known as a function of time, such that the two graphs may be mapped in order to determine the focus.

For example, if a moving object (e.g. a bird, soccer ball, etc.) is to be tracked, the motion of the ob- ject may be estimated with a probabilistic filter (e.g. an extended Kalman filter (EKF)) representing the event density. In such a case, a focus sweep, for example, may be limited to a range smaller than the whole focus range and be centered on the estimated trajectory of the tracked object, such that it is possible to increase the speed of determining the focus.

In some embodiments, the autofocus imaging circuitry is further configured to: determine a contrast based on the determined event density.

Since the contrast may be intertwined with the brightness, as it is generally known, according to the present disclosure, it may be sufficient to determine a contrast based on the event density without the need to do a brightness imaging, but a computationally less complex event sensing instead.

In some, embodiments, a focus sweep is performed for determining a contrast for each focus, and the contrast of an image projected on an image sensor is detected by measuring the rate of change of the event rate or by the rate of change of spatial density of the events. Since the events are produced asynchronously, detecting the contrast according to the present disclosure may be performed faster (e.g. in orders of magnitude) than in known systems since the events are produced asynchronously, as discussed above.

Moreover, in some embodiments, an autofocus imaging apparatus including a high-bandwidth lens (e.g. a liquid lens), is provided, such that a low-latency measurement can be performed, which may enable a closed-loop autofocus. However, the present disclosure is not limited to the case of a high- bandwidth lens, as any lens with an arbitrary bandwidth may be used. Generally, if a video of a (fast) moving object is captured, there is a need for an autofocus, which adapts its speed to the speed of the object propagating through different focus planes.

Therefore, in some embodiments, the autofocus imaging circuitry is further configured to: track an object based on the determined focus.

The object may be any object, who's relative position to the camera may change due to a movement of the object and/ or the camera.

The tracking may be performed with known tracking techniques after the focus is determined or by a re-determining of the focus according to the present disclosure (in order to find the object).

Moreover, if DVS pixels are used, such a re-determining of the focus (i.e. re- focusing) may be car- ried out in between frames of (color) imaging pixels, such that the re-focusing may not (or only rarely) be visible in a taken image (or in consecutive images).

Generally, the disclosure is not limited to finding (determining) the focus based on the above-men- tioned methods (e.g. based on event density), since, in some embodiments, the determination of the focus is further based on a neural network processing the first and the second plurality of brightness change events.

Hence, in some embodiments, the determination of the focus and/ or the tracking may be per- formed with a neural network or any other artificial intelligence, which may utilize known algo- rithms.

In some embodiments, at least one of the determination of the focus and the tracking is based on training of an artificial intelligence.

The artificial intelligence (AI) may use machine learning based methods or explicit feature based methods, such as shape matching, for example by edge detection, histogram based methods, tem- plate match based methods, color match based methods, or the like. In some embodiments, a ma- chine learning algorithm may be used for performing object recognition, e.g. for comparing a detected predefined object with a recognized object to increase a correctness of a detection, which may be based on at least one of: Scale Invariant Feature Transfer (SIFT), Gray Level Co-occurrence Matrix (GLCM), Gabor Features, Tubeness, or the like. Moreover, the machine learning algorithm may be based on a classifier technique, wherein such a machine learning algorithm may be based on least one of: Random Forest; Support Vector Machine; Neural Net, Bayes Net, or the like. Further- more, the machine learning algorithm may apply deep-learning techniques, wherein such deep-learn- ing techniques may be based on at least one of: Autoencoders, Generative Adversarial Network, weakly supervised learning, boot-strapping, or the like. Supervised learning may further be based on a regression algorithm, a perception algorithm, Bayes- classification, Naiver Bayer classification, next-neighbor classification, artificial neural network, and the like.

The artificial intelligence may, in such embodiments, be fed with ground truth data, which may cor- respond to or be based on the predefined object, such that the artificial intelligence may learn to de- termine the focus.

However, the neural network may employ semi-supervised or unsupervised learning, as well.

For example, for semi-supervised or unsupervised learning, an existing autofocus method may be used, which may be a high-performance method, but may determine the autofocus rather slow. However, with such a method, (raw) data may be captured, which may allow to transfer the proper- ties of the (slow) autofocus method to the (faster) method according to the present disclosure. Such a method may be regarded as a type of semi-supervised learning, in a case of a trained autofocus method according to an existing method, or a unsupervised learning, in a case of a manually imple- mented autofocus.

In some embodiments, the neural network is based on or includes a spiking neural network.

A spiking neural network may employ an analog, a digital or a hybrid neuron model or any related or unrelated asynchronous computational hardware or software.

The brightness change event detection element may produce sparse, time-resolved data asynchro- nously, in some embodiments, which may be processed by an asynchronous and/ or sparse compu- tation platform, such as a spiking neural network, without limiting the present disclosure in that regard.

The spiking neural network may be implemented in (neuromorphic hardware) which may be imple- mented within an imaging system, or the like. However, the spiking neural network may derive an algorithm for determining the focus and/ or for tracking an object externally, and the found algo- rithm(s) may be implemented in the autofocus imaging circuitry according to the present disclosure.

Such a configuration may lead to a low power consumption and a low latency control of the autofo- cus imaging circuitry, such that a fast and efficient autofocus determination may be performed.

In some embodiments, the autofocus imaging circuitry is further configured to: obtain at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern. For example, a light source (e.g. of a camera system), which may include a plurality of lasers, laser diodes, VCSELs (vertical cavity surface emitting lasers), and the like, may emit structured light, e.g. a plurality of points, a grid, and the like.

Based on the reflection of the emission pattern (and based on the first and/ or second adjusted fo- cus), the autofocus imaging circuitry may determine the focus, e.g. by determining an event density based on the structured emission pattern. For example, the focus may already be determined after a first (and only) adjusted focus since a deterioration or a smearing out of the structure may be indica- tive of the focal event density.

However, generally, a first and a second focus may be adjusted in order to determine the focus based on the structured light, as discussed herein.

Moreover, the emission pattern may include one or different wavelengths, wherein the present dis- closure is not limited to an emission of visible light, as any suitable electromagnetic frequency may be utilized, such as infrared light.

For example, a pixel of a dynamic vision sensor may be responsive in an infrared light range, as well as in a visible light range, without limiting the present disclosure in that regard.

Thereby, the focus may be determined (quickly) in a dark area (i.e. an area with ambient light below a predetermined threshold), e.g. a (dark) room, a laboratory, and the like.

In some embodiments, the determination of the focus is based on a feedback loop based on the first and second plurality of brightness change events.

For example, it may be desired to find a focus based on the plurality of events which has the least absolute number of events. In such an example, the number of events in the second adjusted focus may be compared with the first adjusted focus. If the number of events of the second adjusted focus is higher (or smaller) than the number of events of the first adjusted focus, a further focus may be adjusted corresponding to the direction of focal length in which still a lower number of events is ex- pected.

Feedback loops are generally known, such that an unnecessary explanation thereof is omitted here. According to the present disclosure, a feedback loop may envisaged since a brightness change event detection element (e.g. a DVS pixel) does not perform imaging, such that it is possible to have a fast feedback loop compared to known autofocus methods, such as contrast sweep, phase detection, and the like.

Some embodiments pertain to an autofocus imaging apparatus, including: an imaging device includ- ing a plurality of brightness change event detection elements and a plurality of imaging elements; and autofocus imaging circuitry, configured to, for each brightness change event detection element: obtain, for a first adjusted focus, a first plurality of brightness change events from the brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene, as discussed herein.

The autofocus imaging apparatus may be implemented in or correspond to any device with a varia- ble focus lens, such as mobile phone, a digital (still or video) camera, a laptop, a notebook, etc., a smart television, and the like.

The autofocus imaging apparatus may include a hybrid imaging device, such as a hybrid DVS (dy- namic vision sensor) camera, which may include an image sensor having conventional imaging ele- ments as well as brightness change event detection elements.

Such an image sensor may be based on known image sensor technologies, such as CMOS (comple- mentary metal oxide semiconductor), CCD (charge coupled device), and the like. The imaging ele- ments of the sensor may be based on known technologies, as well, such as CAPD (current assisted photodiode), SPAD (single photon avalanche diode), or any other diode technology, as it is generally known.

The number of the plurality of imaging elements may be higher than the number of brightness change event detection elements (e.g. a ratio of three to one), since it may not be needed to have a large number of brightness change event detection elements, if it is desired to have a highly resolved image. On the other hand, the ratio may be changed in favor of the brightness change event detec- tion elements, if it is desired to have a more exact focus determination.

Moreover, the arrangement of the imaging elements and the brightness change event detection ele- ments may be any arrangement envisaged by the skilled person and may depend on the application. In some embodiments, the arrangement may be based on a predetermined pattern of the brightness change event detection elements and the imaging elements, e.g. a checkerboard pattern, a reduced checkerboard pattern (i.e. regular, but the respective elements are different in number), an irregular pattern, and the like. Moreover, the predetermined pattern may be based on an even distribution of the brightness change event detection elements in order to find a sufficient average focus of the whole scene, but with “blind” pixels between the imaging elements. In some embodiments, on the other hand, the brightness change event detection elements may be arranged to be in one (or multi- ple) line(s) (or column(s)) of the sensor, if it is not important to have a good average focus of the scene, but if it is important to have a maximum of connected image points (i.e. pixels). The present disclosure is not limited to these two embodiments of arrangement of the image sensor as any (also arbitrary or random) arrangement may be envisaged.

In some embodiments, however, the imaging device may be based on a dual sensor, i.e. an image sensor and a brightness change event detection sensor having a plurality of brightness change event detection elements. In such embodiments, a beam splitter may be adopted to direct incoming light onto each of the sensors.

In some embodiments, the autofocus imaging circuitry is further configured to, for each brightness change event detection element: determine a brightness change event density for determining a spa- tial brightness change event density for the plurality of brightness change event elements.

In such embodiments, for each brightness change event detection element, an event density may be determined. This leads to a plurality of event densities distributed on a two-dimensional plane of the image sensor, i.e. a spatial brightness change event density may be concluded therefrom in order to determine a focus which suits the two-dimensional nature of the image sensor in sufficient manner.

In some embodiments, the imaging device has a hybrid sensor with a predetermined pattern of the brightness change event detection elements and the plurality of imaging elements, as discussed herein.

Some embodiments pertain to an autofocus imaging method, including: obtaining, for a first ad- justed focus, a first plurality of brightness change events from a brightness change event detection element; obtaining, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determining a focus based on the first and the second plurality of brightness change events for autofocusing a scene, as discussed herein.

An autofocus imaging method according to the present disclosure may be carried out with an auto- focus imaging circuitry, an autofocus imaging apparatus, and the like, as discussed herein.

In some embodiments, the autofocus imaging methods further includes: determining an event den- sity based on the first and the second plurality of brightness change events, as discussed herein. In some embodiments, the autofocus imaging methods further includes: determining a contrast based on the determined event density, as discussed herein. In some embodiments, the autofocus imaging methods further includes: tracking an object based on the determined focus. In some embodiments, the determination of the focus is further based on a neural network processing the first and the sec ond plurality of brightness change events, as discussed herein. In some embodiments, the neural network includes a spiking neural network, as discussed herein. In some embodiments, the autofo cus imaging methods further includes: obtaining at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern, as discussed herein. In some embodiments, the determination of the focus is based on a feedback loop based on the first and second plurality of brightness change events, as discussed herein.

The methods as described herein are also implemented in some embodiments as a computer pro gram causing a computer and/ or a processor to perform the method, when being carried out on the computer and/ or processor. In some embodiments, also a non-transitory computer-readable record ing medium is provided that stores therein a computer program product, which, when executed by a processor, such as the processor described above, causes the methods described herein to be per formed.

Returning to Fig. 1, there is depicted an embodiment of an auto focus imaging apparatus 1 according to the present disclosure in a block diagram. Such an imaging apparatus may be implemented in a mirrorless camera.

The autofocus imaging apparatus 1 has a lens stack 2, which is configured to focus incoming light onto a hybrid sensor 3 according to the present disclosure, which includes imaging elements and brightness change event detection elements. The hybrid sensor 3 is, in some embodiments, also im plemented as a hybrid DVS (dynamic vision sensor) camera.

The lens stack 2 has a motor for adjusting distances between lenses of the lens stack 2, such that a focus can be adjusted.

The hybrid sensor 3 is configured to generate a plurality brightness change event in response to a change of the incoming light intensity.

Brightness change event data is transmitted to autofocus imaging circuitry 4 according to the present disclosure, which is implemented as a neuromorphic manycore processor, which is a dedicated pro cessor for event data.

The autofocus imaging circuitry 4 induces the lens stack to adjust a second focus after a first plural ity of events in a first adjusted focus is obtained, such that a second plurality of brightness change events is generated and obtained by the autofocus imaging circuitry 4. The autofocus imaging cir cuitry 4 is configured to determine a focus based on the first and second plurality of brightness change events, as discussed herein. This focus is then transmitted to and adjusted in the lens stack 2.

In such a configuration an autofocus feedback loop is constituted, such that each consecutive ad justed focus is adjusted based on the number of events of the current iteration. That means that the present disclosure is not limited to the case of finding exacdy a two pluralities of brightness change events and the feedback loop may be iterated a (predetermined) number of times until an optimum focus is determined. Fig. 2 depicts a further embodiment of an autofocus imaging apparatus 1 ’ according to the present disclosure. Such an imaging apparatus may be implemented in a mirrored camera.

In contrast to the autofocus imaging apparatus 1 of Fig. 1 , instead of the hybrid sensor 3, an ar rangement of a beam splitter 5, an image sensor 6 having “conventional” (color) imaging pixels, and a brightness change event detection sensor 7.

In this embodiment, the lens stack 2 focuses the incoming light onto the beam splitter 5, which dis tributes the light onto the image sensor 6 and the brightness change event detection sensor 7.

From the brightness change event detector 7, the autofocus imaging circuitry 4 obtains a plurality of brightness change events, as discussed above.

It will be appreciated by a person skilled in the art that the present disclosure is not limited to the optical paths as depicted here.

For example, the beam splitter 5 may be replaced by an adjustable mirror configuration, such that the incoming light may be first transmitted to the brightness change event detection sensor 7 in or der to determine the focus and after the focus is found, the light may be directed to the image sen sor 6.

In other embodiments, one of the two optical paths going to either one of the image sensor 6 or the brightness change event detection sensor 7 may as well be blocked while determining the focus or while imaging the scene, respectively.

Fig. 3 depicts an embodiment of a probabilistic filter 20 for determining a focus based on an event density.

The lower graph depicts the event density (symbolic curve), which is generated based on a focus sweep (i.e. for a predetermined number of focuses, a corresponding number of pluralities of bright ness change events is obtained).

In this embodiment, it is searched for the minimum of the event density in order to find the opti mum focus, since, if the number of events is low, it can be concluded that the amount of scattered and diffracted light is low, as well, such that the focus in a plane of interest (e.g. an object plane) is concluded from the minimum of the event density.

As can be taken from Fig. 3, the abscissas of the two graphs are time, i.e. the two graphs are focal length versus time (top) and event density versus time (bottom). In this embodiment, it is assumed, that the adjusted focal length (of the focus sweep) has a direct impact on the event density, and, with corresponding delays taken into account, the event density can be mapped on a corresponding same time axis as the focal length is mapped onto. Therefore, a focus can be concluded from the corresponding event density minimum.

Fig. 4 depicts a schematic view of a hybrid sensor 30 according to the present disclosure.

The hybrid sensor 30 has brightness change event detection elements 31 and color imaging elements 32, as they are known.

The hybrid sensor 30 has a predetermined pattern of the brightness change event detection elements 31 and the color imaging elements 32, as discussed herein. Every second line of the hybrid sensor only has color imaging elements 32, wherein in the remaining lines, a line with one brightness change event detection element 31 in the middle of the line alternates with a line with two brightness change event detection elements, which are in the second column seen from the left and from the right, respectively.

In other words, the present hybrid sensor 30 has exemplary seven times seven pixels (of course, the present disclosure is not limited in that regard). In the first line, seen from the left, there are three color imaging elements 32, then a brightness change event detection element 31 , then again three color imaging elements 32. In the second line, there are seven color imaging elements. In the fourth line, again, there are seven color imaging elements 32. The fifth to seventh lines are a repetition of the first three lines.

As indicated, it should be recognized that the present disclosure is not limited to a hybrid sensor having seven times seven pixels, as any number of pixels may generally be provided (e.g. in a Mega pixel range), as it is generally known, and the embodiment of the hybrid sensor 30 may be construed as a simplification of such a case.

Fig. 5 depicts an autofocus imaging method 40 according to the present disclosure in a block dia gram.

The autofocus imaging method 40 can be carried out by the autofocus imaging circuitry 4 according to Figs. 1 and 2.

In 41, a first plurality of brightness change events is obtained from a brightness change event detec tion element, as discussed herein.

In 42, a second plurality of brightness change events is obtained from the brightness change event detection element, as discussed herein.

In 43, a focus is determined based on the first and second plurality of brightness change events, as discussed herein.

Fig. 6 depicts a further embodiment of an autofocus imaging method 50 according to the present disclosure. The autofocus imaging method 50 differs from the autofocus imaging method 40 in that an event density is determined on which basis a contrast is determined for determining the focus, as discussed herein.

The autofocus imaging method 50 can be carried out by the autofocus imaging circuitry 4 according to Figs. 1 and 2.

In 51, a first plurality of brightness change events is obtained from a brightness change event detec tion element, as discussed herein.

In 52, a second plurality of brightness change events is obtained from the brightness change event detection element, as discussed herein.

In 53, based on the first and second plurality of brightness change events, an event density is deter mined, as discussed herein.

Based on the event density, a contrast is determined as the derivative of the event density, in 54.

Based on the determined contrast, a focus is determined, in 55, by evaluating the zeros of the con trast.

Fig. 7 depicts a further embodiment of an autofocus imaging method 60 according to the present disclosure. In this embodiment, additionally to the autofocus imaging method 40 of Fig. 5, an object is recognized and tracked based on a spiking neural network, as discussed herein.

The autofocus imaging method 60 can be carried out by the autofocus imaging circuitry 4 according to Figs. 1 and 2.

In 61, a first plurality of brightness change events is obtained from a brightness change event detec tion element, as discussed herein.

In 62, a second plurality of brightness change events is obtained from the brightness change event detection element, as discussed herein.

In 63, a focus is determined based on the first and second plurality of brightness change events, as discussed herein.

In 64, an object is recognized by a spiking neural network, and the object is tracked. The focus is then readjusted based on the movement of the object and based on the algorithm developed by the spiking neural network.

Fig. 8 depicts a further embodiment of an autofocus imaging method 70 according to the present disclosure. In 71, brightness change events are obtained. The arrow 72 indicates that the brightness change events are obtained based on a feedback-loop, i.e. a first plurality of brightness change events is ob tained in a first adjusted focus, on basis of which a second focus is adjusted and a second plurality of brightness change events is obtained. The feedback-loop may have a predetermined number of itera tion (e.g. more than a first and a second adjusted focus) for determining the focus or may run as long as the focus can be determined with a certainty above a predetermined threshold.

In 73, the focus is determined based on the obtained brightness change events.

In other words, the focus is determined based on a closed loop to drive a lens of a camera system with an adjustable focus lens.

It should be recognized that the embodiments describe methods with an exemplary ordering of method steps. The specific ordering of method steps is however given for illustrative purposes only and should not be construed as binding. For example, the ordering of 41 and 42 in the embodiment of Fig. 5 may be exchanged. Also, the ordering of 51 and 52 in the embodiment of Fig. 6 may be ex changed. Further, also the ordering of 61 and 62 in the embodiment of Fig. 7 may be exchanged. Other changes of the ordering of method steps may be apparent to the skilled person.

All units and entities described in this specification and claimed in the appended claims can, if not stated otherwise, be implemented as integrated circuit logic, for example on a chip, and functionality provided by such units and entities can, if not stated otherwise, be implemented by software.

In so far as the embodiments of the disclosure described above are implemented, at least in part, us ing software-controlled data processing apparatus, it will be appreciated that a computer program providing such software control and a transmission, storage or other medium by which such a com puter program is provided are envisaged as aspects of the present disclosure.

Note that the present technology can also be configured as described below.

(1) Autofocus imaging circuitry, configured to: obtain, for a first adjusted focus, a first plurality of brightness change events from a bright ness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

(2) The autofocus imaging circuitry of (1), wherein the first and the second plurality of bright ness change events is indicative of an event density. (3) The autofocus imaging circuitry of (2), further configured to: determine the event density.

(4) The autofocus imaging circuitry of (3), further configured to: determine a contrast based on the determined event density.

(5) The autofocus imaging circuitry of anyone of (1) to (4), further configured to: track an object based on the determined focus.

(6) The autofocus imaging circuitry of anyone of (1) to (5), wherein the determination of the fo cus is further based on a neural network processing the first and the second plurality of brightness change events.

(7) The autofocus imaging circuitry of (6), wherein the neural network includes a spiking neural network.

(8). The autofocus imaging circuitry of anyone of (1) to (7), further configured to: obtain at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern.

(9) The autofocus imaging circuitry of anyone of (1) to (8), wherein the determination of the fo cus is based on a feedback loop based on the first and second plurality of brightness change events.

(10) An autofocus imaging apparatus, comprising: an imaging device including a plurality of brightness change event detection elements and a plurality of imaging elements; and autofocus imaging circuitry, configured to, for each brightness change event detection ele ment: obtain, for a first adjusted focus, a first plurality of brightness change events from the brightness change event detection element; obtain, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determine a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

(11) The autofocus imaging apparatus of (10), the autofocus imaging circuitry being further con figured to, for each brightness change event detection element: determine a brightness change event density for determining a spatial brightness change event density for the plurality of brightness change event elements. (12) The autofocus imaging apparatus of anyone of (10) and (11), wherein the imaging device has a hybrid sensor with a predetermined pattern of the brightness change event detection elements and the plurality of imaging elements.

(13) An autofocus imaging method, comprising: obtaining, for a first adjusted focus, a first plurality of brightness change events from a brightness change event detection element; obtaining, for a second adjusted focus, a second plurality of brightness change events from the brightness change event detection element; and determining a focus based on the first and the second plurality of brightness change events for autofocusing a scene.

(14) The auto focus imaging method of (13), further comprising: determining an event density based on the first and the second plurality of brightness change events.

(15) The autofocus imaging method of (14), further comprising: determining a contrast based on the determined event density.

(16) The autofocus imaging method of anyone of (13) to (15), further comprising: tracking an object based on the determined focus.

(17) The autofocus imaging method of anyone of (13) to (16), wherein the determination of the focus is further based on a neural network processing the first and the second plurality of brightness change events.

(18) The autofocus imaging method of (17), wherein the neural network includes a spiking neural network.

(19) The autofocus imaging method of anyone of (13) to (18), further comprising: obtaining at least one of the first and the second plurality of brightness change events based on a detection of a reflection of a structured emission pattern.

(20) The autofocus imaging method of anyone of (13) to (19), wherein the determination of the focus is based on a feedback loop based on the first and second plurality of brightness change events.

(21) A computer program comprising program code causing a computer to perform the method according to anyone of (11) to (20), when being carried out on a computer. (22) A non-transitory computer-readable recording medium that stores therein a computer program product, which, when executed by a processor, causes the method according to anyone of (11) to (20) to be performed.