Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
DEFECT DETECTION SYSTEM USING A CAMERA EQUIPPED UAV FOR BUILDING FACADES ON COMPLEX ASSET GEOMETRY WITH OPTIMAL AUTOMATIC OBSTACLE DECONFLICTED FLIGHTPATH
Document Type and Number:
WIPO Patent Application WO/2020/008344
Kind Code:
A1
Abstract:
In a described embodiment, a UAV and method for controlling a UAV are disclosed. The UAV is controlled along a predetermined flight path and the camera of UAV is controlled such that adjacent pictures of the captured pictures have an overlap in a predetermined range. In another embodiment, a device and method for detecting defects in image data of a UAV camera are disclosed. An image data management system for managing image data of a UAV and a client software application for performing surface scans with a UAV are also disclosed.

Inventors:
KOO TAO WEI (SG)
DONG WENJUAN (SG)
YONG SEE WEI (SG)
SOH CHENG LOCK (SG)
RUBIO JAIME (SG)
Application Number:
PCT/IB2019/055617
Publication Date:
January 09, 2020
Filing Date:
July 02, 2019
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
HUS UNMANNED SYSTEMS PTE LTD (SG)
JTC CORP (SG)
International Classes:
G01C11/02; B64C39/02; G05D1/10; G06K9/00; G06T7/00
Foreign References:
US20140168420A12014-06-19
US20180170540A12018-06-21
US20170206414A12017-07-20
CN102419598A2012-04-18
US20140168420A12014-06-19
AU2016201290A12016-09-15
Other References:
See also references of EP 3818337A4
Attorney, Agent or Firm:
SCHWEIGER, Martin et al. (SG)
Download PDF:
Claims:
CLAIMS

1. A UAV for detecting defects in a wall of a building, the UAV comprising

- a propulsion means for flying the UAV,

- a digital camera for capturing pictures of the wall,

- a flight control module, the flight control module be ing operative to control the propulsion means based on image data of the at least one camera, such that the UAV flies along the wall of the building along a pre

determined flight path,

- an image capturing module, wherein the flight control module is operative to control the propulsion means and wherein the image capturing module is operative to con trol the digital camera such that adjacent pictures cap tured by the digital camera have an overlap in a prede termined range.

2. The UAV of claim 1, wherein the image capturing module is operative to control the at least one camera to cap ture pictures of the wall from pre-determined capturing positions .

3. The UAV of claim 2, wherein the motor control module is operative to slow down the movement of the UAV at the pre-determined capturing positions.

4. The UAV of claim 2 or claim 3, wherein a distance be tween two adjacent capturing position is equal to an ex tension of a capturing area of a captured picture in a flight direction of the UAV minus a pre-determined over lap in the flight direction, wherein the predetermined overlap is within the pre-determined range.

5. The UAV of one of the claims 1 to 4, wherein the motor control module controls the propulsion means such that the pre-determined flight path results in a correspond ing scanning path, the scanning path comprising a mean dering path, the meandering path comprising capturing segments in a first direction and connecting segments in a second direction, the capturing segments being joined by the connecting segments.

6. The UAV according to one of the preceding claims, where in the digital camera comprises a visible light sensor.

7. The UAV according to one of the preceding claims, where in the digital camera comprises a radar sensor.

The UAV according to one of the preceding claims, where- in the digital camera comprises an infrared sensor.

9. The UAV according to one of the preceding claims, where in the digital camera comprises an acoustic transducer.

10. Method for controlling a UAV with at least one camera, the method comprising:

- control a propulsion means of the UAV to move the UAV along a pre-determined flight path

- controlling the at least one camera to capture pic tures of a wall of a building at pre-determined captur ing locations, such that adjacent pictures of the cap tured pictures have an overlap in a predetermined range.

11. The method according to claim 10, comprising - controlling the propulsion means to slow down the movement of the UAV at the pre-determined capturing po sitions .

12. The method according to claim 10, comprising

controlling the propulsion means such that a distance between two adjacent capturing positions is equal to an extension of a capturing area of a captured picture in a flight direction of the UAV minus a pre-determined over lap in the flight direction, wherein the predetermined overlap is within the pre-determined range.

13. The method according to one of the claims 10 to 12, com prising

controlling the propulsion means such that the pre determined flight path results in a corresponding scan ning path, the scanning path comprising a meandering path, the meandering path comprising capturing segments in a first direction and connecting segments in a second direction, the capturing segments being joined by the connecting segments.

14. A device for detecting defects in image data of a UAV camera comprising,

- a first data storage for storing the image data,

- a processing means for processing the image data,

- a second data storage for storing processed image da ta, wherein the processing means is operative to identi fy a first crack in a first picture and to identify a second crack in a second picture, wherein the first pic ture overlaps with the second picture, and wherein the processing means is furthermore operative to identify whether the first crack and the second crack form part of the same crack.

15. Method for detecting defects in image data of a UAV cam era, the image data comprising a plurality of pictures, the method comprising

- detecting a first crack in a first picture of the plu rality of pictures

- detecting a second crack in a second picture of the plurality of pictures, the second picture being adjacent to the first picture,

- identify whether the first crack and the second crack form part of the same crack.

16. The method of claim 15, wherein the detection of the first crack and of the second crack comprises presenting the respective picture to a neural network,

deriving from the output of the neural network whether a set of pixels of the respective picture corresponds to a crack .

17. The method according to one of the claims 15 or 16,

wherein the detection of the first crack and of the sec ond crack comprises

inputting the respective picture to a support vector ma chine,

deriving from the output of the support vector machine whether a set of pixels of the respective picture corre sponds to a crack.

18. The method according to one of the claims 15 to 17,

wherein the detection of the first crack and of the sec ond crack comprises inputting the respective picture to a nearest neighbour classifier,

deriving from the output of the nearest neighbour clas sifier whether a set of pixels of the respective picture corresponds to a crack.

19. Image data management system for managing image data of an unmanned aerial vehicle (UAV) , the management of the image data comprising acquisition, storage and distribu tion of the image data, the image data relating to pic tures of a digital camera of the UAV, the image data management system comprising

- a data acquisition module,

- a data processing module and a data compilation mod ule, wherein the data processing module and the data compilation module are provided on one or more server computers ,

- a client module, the client module being provided on a client device, wherein the client module is operative to receive client input data relating to an object to be scanned for defects, to a time interval between succes sive scans of the object, and to a type of output data, and to forward the client input data to the data compi lation module,

wherein the data compilation module is operative to store the client input data as configuration data, and wherein the data acquisition module is operative to trigger a UAV to perform scans of the object at prede termined times and to transmit image data of the object scans to the data acquisition module.

20. The Image data management system of claim 19, wherein the client module is furthermore operative to establish a data link with the server computer and to request out put data from the data compilation module.

21. A client software application for performing surface scans with an unmanned aerial vehicle (UAV) ,

the client software application being operative to

- submit a scan request, the scan request comprising da ta specifying a location to be scanned, a pre-determined scan interval and a type of data to be retrieved,

- retrieving image data of the surface scan.

22. A portable client device with a computer readable

memory, the computer readable memory comprising the cli ent software application of claim 21.

Description:
DEFECT DETECTION SYSTEM USING A CAMERA EQUIPPED UAV FOR

BUILDING FACADES ON COMPLEX ASSET GEOMETRY WITH OPTIMAL AUTO

MATIC OBSTACLE DECONFLICTED FLIGHTPATH

In a first aspect, the specification discloses an unmanned aerial vehicle (UAV) for detecting defects (a building flaw or design mistake that reduces the value of the building, and causes a dangerous condition; list of defects include Paint peeling, concrete cracks, wall tile cracks, water ponding, water stains, buckling, concrete spalling, loose sealant, tile delamination, glass crack, rust, wall opening,

block/brick wall crack, surface discoloration, mold on fagade rubber gasket, molding, sealant discoloration, etc. non- exhaustive ) in an outer wall of a building. Preferentially, the UAV is provided by an electrically driven rotary wing UAV, which comprises propellers. The UAV refers in particular to small size drones or drone boxes, such as small size mul ticopters, quadcopters.

The UAV comprises a propulsion means for flying the UAV, which can be provided by two or more propellers and electric motor which are connected to the propellers. The propulsion means is connected to an energy source of the UAV. In partic ular, the energy source can be an electric energy source such as an electric battery, fuel cell, hybrid battery/fuel cell, or any other power generator systems/methods that produce electrical energy needed to propel/move the UAV platform.

Furthermore, the UAV comprises a digital camera for capturing picture or video of the wall and a data storage for storing the pictures captured by the digital camera, which may be provided as part of the digital camera. The data storage can be provided by a computer memory, for example a magnetic or a semiconductor-based memory or any other type of computer memory. The digital camera can be provided with a visible light sensor, but also with a radar sensor, an infrared sen- sor, an acoustic transducer or any other type of image sensor which can provide an image or picture of a surface.

In particular, the digital camera can comprise an image sen sor and a transmitter or an emitter for irradiating the wall surface. This allows to capture images of the surface wall even in poor lighting conditions. The emitter can be an emit ter of acoustic, radar or infrared waves or it can also be provided by a light emitter.

Furthermore, the UAV can comprise further contactless sensors which provide complementary information to the image data.

For example, a radar sensor can provide depth information to improve the defect identification or to provide more detailed information about specific features of the defect.

A flight control module of the UAV is operative to autono mously control the propulsion means based on image or video data of the at least one camera. In particular, a module of the UAV can be operative by means of a computer program in a computer storage of the UAV and/or by means of electronic circuitry of the UAV.

Furthermore, the flight control module may receive other po sition data, for example from a satellite navigation system of the UAV or from a wireless network navigation system. The flight module controls the propulsion means such that the UAV flies in front of the wall of the building along a pre determined flight path. In particular, the propulsion means can be controlled such that the UAV flies in a predetermined distance and/or with a predetermined orientation to the wall. An image capturing module of the UAV is operative to send control signals to the digital camera which cause the digital camera to capture pictures or video of the wall.

The flight control module is operative to control the propul sion means and the image capturing module is operative to control the digital camera such that adjacent pictures cap tured by the digital camera have an overlap that is a prede termined range, for example an overlap may be in the range of 5% to 10% or 5% to 20%. In one embodiment, the adjacent pic tures are adjacent in a flight direction of the UAV.

In a simple implementation the flight control module controls the propulsion means to fly the UAV along the predetermined flight path and the image capturing module causes the digital camera to take pictures or video at suitable capturing posi tions .

A flightpath algorithm receives sensor information from the drone and calculates the optimal distance, image overlap, and obstacles of the complex asset geometry to provide safe flight as well as image taking accuracy and consistency. The algorithm automatically localises and crops the image for clarity .

The flightpath algorithm is also able to generate multiple optimised flightpaths of one to many complex asset geometries resulting in the production of swarming flight paths per com plex asset or multiple complex assets.

The flight control module and the image capturing module are connected to an electric battery or any other suitable power source located on the UAV platform. In one embodiment, the propulsion means comprises electric motors and the electric battery (or another suitable power source on the UAV) pro vides the energy source to which the propulsion means is con nected .

The control of the propulsion means can be based on a pre dicted/estimated flight path, on a real time analysis of im age data or video data or on a combination of both. In the context of the specification, a picture refers especially to a rectangular contiguous area of image pixels or data derived therefrom. The image pixels correspond to sensor signals of light intensity sensors.

In one embodiment, the image capturing module is operative to control the at least one camera to capture pictures of the wall from pre-determined capturing positions, especially to capture one picture per capturing position.

In a further embodiment, the motor control module is opera tive to slow down or even to halt the movement of the UAV at the pre-determined image capturing positions.

In a specific embodiment, a distance between two adjacent capturing position is equal to an extension of a capturing area of a captured picture in a flight direction of the UAV minus a pre-determined overlap in the flight direction, wherein the predetermined overlap is within the pre

determined range.

According to a further embodiment, the motor control module controls the propulsion means such that the pre-determined flight path results in a corresponding scanning path, which comprises a meandering path. The meandering path in turn com prises capturing segments in a first direction and connecting segments in a second direction, the capturing segments being joined by the connecting segments.

The scanning path is formed by the location on the wall that corresponds to a centre location of an imaging surface on which a camera lens of the at least one camera projects.

In particular, the flight path can be provided by a meander ing path that comprises capturing segments in a first direc tion and connecting segments in a second direction, the cap turing segments being joined by the connecting segments.

Furthermore, the specification discloses a computer imple mented method for controlling a UAV with at least one camera. According to this method a propulsion means of the UAV is controlled to move the UAV along a pre-determined flight path. Furthermore, the at least one camera is controlled to capture pictures of a wall of a building at or from pre determined capturing locations, such that adjacent pictures of the captured pictures have an overlap that is within a predetermined range relative to a picture extension.

According to a further embodiment, the method comprises con trolling the propulsion means to slow down or even halt the movement of the UAV at the pre-determined capturing posi tions .

According to a further embodiment, the method comprises con trolling the propulsion means such that a distance between two adjacent capturing positions is equal to an extension of a capturing area of a captured picture in a flight direction of the UAV minus a pre-determined overlap in the flight di- rection, wherein the predetermined overlap is within the pre determined range.

According to a yet a further embodiment, the method comprises controlling the propulsion means such that the pre-determined flight path results in a corresponding scanning path. The scanning path comprises a meandering path, which in turn com prises capturing segments in a first direction and connecting segments in a second direction. The capturing segments are joined by the connecting segments.

According to a further aspect, the current specification dis closes a device for detecting defects in image data of a UAV camera with a first data storage for storing the image data, a processing means for processing the image data, and a sec ond data storage for storing processed image data.

The processing means is operative to identify a first defect in a first picture and to identify a second defect in a sec ond picture, wherein the first picture overlaps with the sec ond picture, and wherein the processing means is furthermore operative to identify whether the first defect and the second defect form part of the same defect.

Furthermore, the present specification discloses a computer implemented method for detecting defects in image data of a UAV camera. The image data comprises a plurality of pictures captured by the UAV camera.

According to this method a first defect is detected in a first picture of the plurality of pictures, and a second de fect is detected in a second picture of the plurality of pic- tures, the second picture being adjacent to the first pic ture .

Furthermore, the method comprises identifying whether the first defect and the second defect form part of the same de fect .

According to an embodiment, the detection of the first defect and of the second defect comprises presenting the respective first or second picture to a neural network and deriving from the output of the neural network whether a set of pixels of the respective picture corresponds to a defect.

According to a further embodiment, the detection of the first defect and of the second defect comprises inputting the re spective first or second picture to a support vector machine and deriving from the output of the support vector machine whether a set of pixels of the respective picture corresponds to a defect.

According to a further embodiment, the detection of the first defect and of the second defect comprises inputting the re spective first or second picture to a nearest neighbour clas sifier and deriving from the output of the nearest neighbour classifier whether a set of pixels of the respective picture corresponds to a defect.

In a further aspect, the present specification discloses an image data management system for managing image data of an unmanned aerial vehicle. The system comprises three compo nents which are related to data acquisition, management and client interface, respectively. In particular, the management of the image data comprises the optimization, acquisition, storage and distribution of the image data, wherein the image data relates to pictures of a digital camera of the UAV, pictures of an image capturing sensor, or video file from an image recording device. More specifically, the image data management system comprises a data acquisition module, data optimization module, a data processing module, a data compilation module, and a client module .

The data processing module and the data compilation module are provided on one or more server computers, while the cli ent module is provided on a client device.

The client module is operative to receive client input data relating to an object to be scanned for defects, to a time interval between successive scans of the object, and to a type of output data, and to forward the client input data to the data compilation module.

The data compilation module is operative to store the client input data as configuration data. The data acquisition module is operative to trigger a UAV to perform scans of the object at predetermined times and to transmit image data of the ob ject scans to the data acquisition module.

According to a further embodiment, the client module is fur thermore operative to establish a data link with the server computer and to request output data from the data compilation module .

Moreover, the present specification discloses a client soft ware application for performing surface scans with an un- manned aerial vehicle (UAV) . The software application can be embodied by a set of computer readable instructions on a com puter readable storage medium. The client software applica tion is operative to submit a scan request over a communica tion network, which may wireless or wired or a combination thereof .

The scan request comprises data specifying parameters relat ing to the surface scan and the retrieved data, such as a lo cation to be scanned, a pre-determined scan interval and a type of data to be retrieved.

Furthermore, the client software application is operative to retrieve image data of the surface scan in accordance with a previous scan request over the same or over a different com munication network. The data retrieval can be a push retriev al that is initiated by a remote computer or a pull retriev al, which is initiated by the client software application.

The present application also discloses a portable client de vice with a computer readable memory, which comprises the client software application.

The subject matter of the present specification is now ex plained in further detail with reference to the below men tioned Figures, in which

Figure 1 shows a defect detecting UAV flying along a cali bration path during a calibration procedure,

Figure 2 shows the defect detecting UAV of Fig. 1 flying along a flight path during a scan procedure,

Figure 3 shows an enlarged section of the flight path of

Fig. 2 indicating captured pictures of the building fagade,

Figure 4 shows a schematic layout of the UAV of Fig. 1, Figure 5 shows a calibration procedure of the UAV of Fig. 1, Figure 6 shows a wall scanning procedure of the UAV of Fig.

1,

Figure 7 shows a system for the acquisition, storage and

distribution of defect data,

Figure 8 shows a vertical defect in a wall plaster,

Figure 9 shows a vertical and a diagonal defect in a con

crete wall,

Figure 10 shows a vertical and a diagonal defect in a masonry wall ,

Figure 11 shows a learning curve of a semi-supervised learn ing procedure on the basis of defect data,

Figure 12 shows a first flight path that is adapted to fea tures of a building, and

Figure 13 shows a second flight path that is adapted to the features of the building.

Fig. 1 shows a defect detecting unmanned aerial vehicle (UAV) 10 which flies along a calibration path 11 during a calibra tion procedure. The calibration path 11 of the UAV 10 is in dicated by dashed arrows. The dashed arrows 12 indicate a vertical projection of segments of the UAV's flight path 11 onto an outer wall 13 of a multi-level building 14.

During the calibration procedure, the UAV 10 moves in front of the outer wall 13. A camera of the UAV 10, which is not shown in Fig. 1, is directed towards the building wall 13 and captures a rectangular area 15 which is indicated in Fig. 1.

A dotted region 16 of the building wall 13 indicates a scan area of the building wall 13 which is to be scanned for de fects. A rectangular region 17 inside the dotted region indi cates a no-scan area which does not require scanning for de- fects. For example, the no-scan area may comprise a glass fa- gade which does not have defects to be investigated.

The scan area 16 and the no-scan area 17 of Fig. 1 are com puted based on image data taken during the calibration pro cess of Fig. 1. Preferentially, the computation of the scan areas and the no-scan areas is carried out by a processor of the UAV 10. Furthermore, the computation of the flight paths of Fig. 2 is preferentially carried out by the processor of the UAV 10. In this way, the UAV 10 can directly proceed to the scan phase of Fig. 2 without the need of a data transfer to a ground station.

In the example of Fig. 1, the UAV 10 is provided by a quad- copter as a preferred example, though the subject matter of the present specification is not limited to a specific type of UAV. The quadcopter is equipped with a digital camera, electronics for controlling the digital camera and for con trolling electric motors of the quadcopter and for evaluating image data taken by the digital camera.

The abovementioned components of the UAV 10 are illustrated in more detail in Fig. 4. The calibration process of Fig. 1 is illustrated in further detail by the flow diagram of Fig.

4.

Fig. 2 shows the defect detecting UAV 10 of Fig. 1 flying along a first flight path 20 during a scan procedure. The first flight path 20, which is indicated on the building wall 13, has the form of a meandering line and consists of hori zontal capturing segments 21 connected by vertical connection segments 22. A rectangular capturing area 29 of the UAV 10 is indicated on the building wall 13. The distance of the UAV 10 to the building wall 13 during the scan procedure, which his shown in Fig. 2, is smaller than the distance of the UAV 10 to the building wall 13 during the calibration procedure, which is shown in Fig. 1.

A second flight path 24, which has vertical capturing seg ments 25 connected by horizontal connection segments 26, is indicated on a second building wall 27 by a meandering line 24. Similar to Fig. 1, the lines indicate a vertical projec tion of the UAV's flight path onto the building wall 27.

A flight path with horizontal capturing segments is in many situations less energy consuming and preferable over a flight path with vertical capturing segments. However, there are al so situation in which a flight path with vertical capturing segments may be preferable, for example when the height of the building is much larger than the width. In general, a flight path with alternating horizontal and vertical seg ments, such as the flight paths shown in Fig. 2, can provide a convenient combination or stitching of adjacent pictures.

More in general, the paths 11, 20, and 24 of Fig. 1 and Fig.

2 can also refer to a scanning path instead of a flight path. In a case where the UAV 10 deviates from a pre-determined flight path, for example due to wind influence or due to avoidance of obstacles, the UAV 10 may compensate for such deviations by adjusting the alignment and the magnification of the camera such that a pre-determined scanning path is maintained within a given accuracy. Furthermore, a pre determined scanning path 20, 24 can be achieved through a combination of UAV 10 and camera movement. Fig. 3 shows a sectional enlargement of the first flight path 20, in which image capturing areas 29 and successive image capturing positions 30 of the UAV are indicated. Furthermore, Fig. 3 shows horizontal overlap regions 23 and vertical over lap regions 33 between the image capturing areas 29. For clarity reasons, only one of the capturing areas 29 and of the overlap regions 32, 33 is labelled.

A defect 35 of the building wall 13 extends over a first, a second and a third capturing area 29 and over a horizontal overlap region 32, a vertical overlap region 33 and a further horizontal overlap region 32 between adjacent capturing areas 29.

Experiments have shown that an overlap of between 5% - 10% of the respective extension of the capturing areas can help avoid that a defect appearing on more than one picture is identified as multiple defects as compared to a larger over lap of more than 10%.

Fig. 4 shows a schematic layout of the UAV 10 of Fig. 1. A UAV battery 40 is connected to a motor control module 41 and to an image capturing module 42. Furthermore, the battery 40 is connected to a charging connector 43 via a charging module 44. A control line 45 and a data line 46 of the image captur ing module 42 are connected to a camera 47 of the UAV 10 and a data line 48 of the image capturing module 42 is connected to an image data memory 49. By way of example, the image data memory 49 may be provided by semiconductor memory or a mag netic disk. The motor control module 41 is connected to four motors 50 of the UAV 10. More specifically, power electronics of the motor control module, which is not shown in Fig. 4, is connected to the motors 50. Furthermore, the motor control module 41 com prises processing means for computing motor control signals.

The image capturing module 42 is connected to a data transfer connector 51, for example a USB connector or any other type of connector that has data transfer functionality. The image capturing module 42 and the motor control module 41 are con nected by a further data line 53. The image capturing module 42 comprises processing means for deriving control signals to the camera and for evaluating and storing data of the camera 47.

A position sensing module 54 is connected to the motor con trol module 41 and to the image capturing module 42. The po sition sensing module 54 comprises a position sensor, such as GPS receiver or any other type of position sensor, an orien tation sensor, and processing means for determining a posi tion and an orientation of the UAV 10. The position sensor can also be provided by an image sensor of the digital cam era .

In particular, the position sensor can be provided by a sen sor that does not rely on a satellite navigation system, which is advantageous in areas in which access to a satellite navigation system is not available or is denied, such as a wireless network positioning sensor or a position sensor that comprises a radar sensor, an infrared sensor, an acoustic sensor, a camera sensor or any combination thereof. In areas, in which access to GPS or other satellite position systems are unavailable or access is denied, it is advanta geous to provide position sensors which can function inde pendently from a satellite navigation system. Similarly, it can be advantageous to provide position sensors which do not depend on the availability of a wireless data link.

The processing means of the image capturing module, of the position sensing module and of the motor control module are not shown in Fig. 4. By way of example, the processing means can be implemented by one or more microprocessors or one or more application specific integrated chips (ASICS) .

The embodiment of Fig. 4 is provided by way of example. The subject matter of the present specification is not limited to the example of Fig. 4. The number of motors is purely given as an example. For example, the UAV may be provided with a multicopter having two, three, four, five or more motors. Furthermore, the functionality of the UAV may be provided by other arrangements of modules which are different from the example of Fig. 4.

In a further embodiment, which is not shown in Fig. 4, the UAV is provided with a transmitter and a corresponding wire less communication module for transfer of data over a wire less communication link. Thereby, image data can be trans ferred from and to the UAV even when the UAV is flying. The wireless communication module may also provide further func tionality such as a remote control of the UAV.

Fig. 5 shows a calibration procedure of the UAV of Fig. 1. In a first step 60, the UAV 10 activates a position sensing means of the position sensing module, such as a GPS receiver.

In a next step 61, the UAV 10 moves to a calibration starting point. In a further step 62, the UAV 10 moves along a cali bration path 11. In the example of Fig. 1, the calibration path 11 is aligned to the boundaries and the corners of the building wall and comprises four segments 12 which are each parallel to the boundaries or edges of the building wall.

In the flow diagram of Fig. 5, the building wall 13 is re ferred to as "building fagade", which takes into account a possible three-dimensional structure, such as windows, col umns, jutties, balconies, piping, etc. Preferentially, the building wall 13 is essentially flat, but it may also com prise protrusions, openings, and recesses.

In the example of Fig. 1, the rectangular calibration path 11 is sufficient to cover a required surface of the building wall 13 with the camera 47 at the selected distance to the building wall 13 and for the given size of the building wall 13. For other configurations, the calibration path 11 may in clude further rectangular paths, which are nested inside the outer rectangular path and which have successively smaller diameters. Depending on the calibration settings, the re quired surface of the building wall 13 may comprise the whole of the building wall 13 or only an outer strip of a predeter mined width.

In a further step 63, which may be executed during and/or af ter the step of moving along the calibration path, the image capturing module 42 determines zero or more no-scan zones 17 on the building wall, which do not need to be scanned for de fects .

In a further step 64, which may be executed during and/or af ter the step of moving along the calibration path 11, the im age capturing module 42 determines the dimensions and the po sition of the building wall 13.

In a further step 65, the image capturing module computes a flight path of the UAV 10 for detecting defects in the build ing wall 13 along the flight path. In a decision step 66, the UAV 10 decides, based on configuration data or on user input over a remote control, whether or not to initiate a fagade scan .

If no subsequent fagade scan is initiated, the UAV 10 cali brates the next flight path or moves to a parking position in step 67. Otherwise, the UAV 10 moves to a starting position of the flight path in step 68.

Fig. 6 shows a scanning process of the UAV 10 of Fig. 1, in which the UAV 10 follows a flight path that has been estab lished on the basis of a preceding calibration process, for example a calibration process according to the flow diagram of Fig. 5.

In a first step 68, the UAV moves to a starting position of the flight path. In a next step 70, the image capturing mod ule 42 sends a control signal to the camera 47 to take or capture a first picture. In a further step 71, the motor con trol module 41 sends control signals to the motors 50 of the UAV 10 to move the UAV 10 in a first direction. In the case of the flight path 20 shown on the front surface 13 of the multi-level building 14 of Fig. 2, the first direc tion is a horizonal direction. In the case of the flight path shown on the side surface of the multi-level building 14 of Fig. 2, the first direction is a vertical direction.

In a decision step 72, the image capturing module determines the size of an overlap between adjacent image capturing areas in the first direction and whether the overlap has reached 5% of the extension of a rectangular image capturing area 29 in the first direction. If this is the case, the motor control module 41 controls the electric motors 50 of the UAV 10 to slow down or halt the UAV 10 at an image capturing position 30. In a step 73, the image capturing module 42 controls the camera 47 to capture the next picture of the building wall 13 by storing image data of the camera 47 in the image data memory 49.

In a decision step 74, the motor control module 41 of the UAV 10 compares the current position of the UAV 10 with the pre determined path and determines whether an end of the current horizontal or vertical path segment has been reached. If not, the UAV 10 is controlled to move further in the first direc tion .

If it is determined that the end of the current path segment is reached, the motor control module 41 determines, in a de cision step 75, whether an end of the flight path is reached. If not, the UAV 10 moves, in a step 76, in the second direc tion by one picture extension, which is one picture height in the case of a horizontal path segment. Else, the UAV 10 selects the next available flight path in a step 77 and the process loops back to step 68 in which the UAV 10 moves to the starting position of the selected flight path .

Fig. 7 shows a system 80 for the acquisition, storage and distribution of defect data which has been acquired with the UAV 10 of Fig. 1. The data acquisition, storage and distribu tion system 80 comprises a data acquisition module 82, which is provided by a first software application or "app" 83, a data processing module 84 and a data compilation module 85, which are provided by a second software application 86 and a client module 87, which is provided by a third software ap plication 88.

The data processing module 84 receives input data from the data acquisition module 82 and from a source 89 of external data and stores processed data using the data compilation module 85. The client module 87 retrieves information select ed by a customer using the data compilation module 85.

In one embodiment, the system 80 comprises a network of ro botic devices that connect to a microservice cloud.

In a further embodiment, a customer can buy a subscription using the client module 87. As a result, the data compilation module 85 updates customer specific configuration data which causes a UAV 10 to inspect a selected building 14 in a pre determined interval, for example monthly, to generate a re quest and to indicate any changes to the results of previous inspections. The inspection process is triggered automatical ly and is performed autonomously without the need of human intervention . Figures 8 to 10 show, by way of example, various different types of defects in outer walls of buildings which can be de tected with a defect detection system according to the pre sent specification.

Figure 8 shows a vertical defect 92 in a wall plaster 93 of a single-family house 94.

Figure 9 shows a vertical defect 95 and a diagonal defect 96 in a concrete wall 97 of a bungalow 98.

Figure 10 shows a vertical defect 100 and a diagonal defect 102 in a masonry wall 103 of a house, which is not shown in Fig. 10. Unlike the diagonal defects of the two preceding ex amples, the diagonal defect of Fig. 10 follows the brick structure and comprises vertical and horizontal segments.

Fig. 11 shows an increase in accuracy as a function of re peated learning steps in a semi-supervised learning procedure of a neural network.

In the example of Fig. 11, the training is performed on false positive results, or in other words those pictures in which a surface feature is wrongly detected as a defect. The false positive can be caused by paint traces, twigs, shadows or other features. In Fig. 11, an accuracy of 0% means that all false positives are still wrongly detected and an accuracy of 100% means that all false positives are rejected.

By way of example, a rate of false negatives, in which a de fect is present but not detected, may be 15%, a rate of false positives, in which a defect is not present but wrongly de tected, may be 20%.

In general, a model training may be performed by first dis carding bad images in a pre-processing stage, train the model with the pre-processed images and, in a post-processing stage, discarding outliers that do not fit into the model.

In one embodiment, a model is represented by numbers that de fine connection weights of a neural network, and which are modified during the training procedure. The neural network can be a network in which connections, a number of layers and a number of nodes are fixed in advance. Or, according to a more complex model, those are variable parameters of the neu ral network. In another embodiment, the model is represented by parameters of a statistical model which are modified dur ing the training procedure. In one embodiment, the statisti cal model comprises a support vector machine, according to another embodiment, the statistical model comprises a nearest neighbour classifier.

According to a further embodiment, the UAV comprises a stere oscopic camera and the picture processing and defect recogni tion comprises an evaluation of depth information.

Figures 12 and 13 show first and second flight paths that are adapted to features of a building. More specifically, the flight paths are adapted to a configuration that is similar to Fig. 1, and in which there is a rectangular no-scan area 17.

The flight path of Fig. 12 comprises a first portion 104 be low a no-scan area 17, in which the flight path extends over substantially the whole width of the wall 13, a second por tion 105, in which the flight path extends from the left of the excluded region to the left boundary of the wall 13, a third portion 106 above the no-scan area 17, in which the flight path extends from the left boundary of the wall to the virtual extension of the right boundary of the no-scan area 17, and a fourth portion 107 in which the flight path between the right boundary of the wall and the virtual extension of the right boundary of the no-scan area 17.

In the flight paths of Fig. 12 and 13 the UAV travels back and forth along one of the horizontal segments of the flight path. In such a case the UAV may take pictures only during the flight in one direction. Or it may take pictures in both directions, which may involve taking the same pictures twice at the same capturing locations or taking the pictures only once at alternate capturing locations .

In one embodiment, a defect detection procedure is similar to a procedure for the optical recognition of characters and comprises a reduction of a picture's colour scale or grey scale to black and white pixels, which may be represented as the binary numbers 0 and 1.

A processing of the captured pictures may be executed on a processing means of the UAV 10, on a ground-based computer or partially on the UAV 10 and partially on the ground-based computer. This processing may involve an image pre-processing stage in which image distortions resulting from a lens shape and/or from an oblique orientation of a camera image plane relative to the scanned wall are removed. In one embodiment, the image processing involves the computa tion of defect specific features such as a general direction or "skeleton" of the defect and a defect width

Fig. 13 shows a flight path which is similar to the flight path of Fig. 12 but in which a third portion 106' above the no-scan region extends over substantially the whole width of the wall. As a consequence, the UAV's flight path has less directional changes and up-and-down movements as compared to the flight path of Fig. 12. As a further consequence, the flight path of Fig. 13 has a vertical connecting segment which extends from the upper boundary of the wall to the vir tual extension of the upper boundary of the no-scan area 17 and which is substantially higher than a height of the cap turing area 29.

Although the above description contains much specificity, these should not be construed as limiting the scope of the embodiments but merely providing illustration of the foresee able embodiments. By way of example, the flight paths are not limited to the ones shown in the examples and the type of UAV that can be used is not limited to a quadcopter. Especially the above stated advantages of the embodiments should not be construed as limiting the scope of the embodiments but merely to explain possible achievements if the described embodiments are put into practise. Thus, the scope of the embodiments should be determined by the claims and their equivalents, ra ther than by the examples given. Reference Numeral s

10 quadcopter/UAV 51 data transfer connector

11 flight path 53 data line

12 path segment 54 position sensing module

13 building wall 60 method step

14 multi-level building 61 method step

15 capturing area 62 method step

16 scan area 63 method step

17 no-scan area 64 method step

20 flight path 65 method step

21 capturing segment 66 decision step

22 connection segment 67 method step

24 flight path 68 method step

25 capturing segment 70 method step

26 connection segment 71 method step

29 capturing area 72 decision step

30 capturing position 73 method step

32 horizontal overlap re 74 decision step

gion 75 decision step

33 vertical overlap region 76 method step

35 wall defect 77 method step

40 battery 80 software application

42 image capturing module 82 data acquisition module

43 charging connector 83 software application

44 charging module 84 data processing module

45 control line 85 data compilation module

46 data line 87 client module

47 camera 88 software application

48 data line 89 external data source

49 image data memory 93 wall plaster

50 motor 94 house vertical defect 104 flight path portion diagonal defect 105 flight path portion concrete wall 106, 106' flight path por bungalow tion

vertical defect 107, 107' flight path por diagonal defect tion

brick wall