Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
APPARATUSES, METHODS AND COMPUTER PROGRAMS FOR CONTROLLING A MICROSCOPE SYSTEM
Document Type and Number:
WIPO Patent Application WO/2021/148346
Kind Code:
A1
Abstract:
Examples relate to apparatuses, methods and computer programs for controlling a microscope system, and to a corresponding microscope system. An apparatus for controlling a microscope system comprises an interface for communicating with a camera module. The camera module is suitable for providing camera image data of a head of a user of the microscope system. The apparatus comprises a processing module configured to obtain the camera image data from the camera module via the interface. The processing module is configured to process the camera image data to determine information on an angular orientation of the head of the user relative to a display of the microscope system. The processing module is configured to provide a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user.

Inventors:
THEMELIS GEORGE (DE)
Application Number:
PCT/EP2021/050932
Publication Date:
July 29, 2021
Filing Date:
January 18, 2021
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
LEICA INSTR SINGAPORE PTE LTD (SG)
THEMELIS GEORGE (DE)
International Classes:
G02B21/00; A61B34/00; A61B34/30; A61B90/50
Foreign References:
EP3125737B12018-12-05
US5345087A1994-09-06
Attorney, Agent or Firm:
2SPL PATENTANWÄLTE PARTG MBB (DE)
Download PDF:
Claims:
Claims

1. An apparatus (110; 420) for controlling a microscope system (100; 400), the appa ratus comprising: an interface (112) for communicating with a camera module (120), the camera mod ule being suitable for providing camera image data of a head of a user of the micro scope system; a processing module (114) configured to: obtain the camera image data from the camera module via the interface, process the camera image data to determine information on an angular orientation of the head of the user relative to a display (140) of the microscope system, provide a control signal for a robotic adjustment system (130) of the microscope sys tem based on the information on the angular orientation of the head of the user.

2. The apparatus according to claim 1, wherein the processing module is configured to determine an angle of view of the user of the microscope system towards the display based on the camera image data, and to provide the control signal based on the angle of view of the user.

3. The apparatus according to claim 2, wherein the processing module is configured to determine the angle of view of the user based on the angular orientation of the head of the user.

4. The apparatus according to claim 3, wherein the processing module is configured to perform eye tracking on one or both eyes of the user, and to determine the angle of view of the user based on the angular orientation of the head of the user and based on the eye tracking.

5. The apparatus according to one of the claims 1 to 4, wherein the processing module is configured to determine the information on the angular orientation using image analysis on the camera image data.

6. The apparatus according to one of the claims 1 to 5, wherein the display of the mi croscope system is an auxiliary display that is arranged in proximity of an eyepiece of a microscope of the microscope system, or wherein the display of the microscope system is an auxiliary display that is ar ranged at a base unit of the microscope system, or wherein the display of the microscope system is an ocular display of an eyepiece of a microscope of the microscope system.

7. The apparatus according to one of the claims 1 to 6, wherein the processing module is configured to determine a fixation of the user on a point based on the camera im age data, wherein the processing module is configured to determine the control sig nal for the robotic adjustment system such that a central point of view of a micro scope (150; 410) of the microscope system remains fixed on the point.

8. The apparatus according to one of the claims 1 to 7, wherein the processing module is configured to determine information on a position of the head of the user and in formation on an angle of the head of the user based on the camera image, and to de termine the information on the angular orientation of the head of the user based on the information on the position of the head of the user and based on the information on the angle of the head of the user.

9. The apparatus according to claim 8, wherein the processing module is configured to generate the control signal such that the robotic adjustment system is triggered to pan a field of view of a microscope of the microscope system if the position of the head of the user changes without a corresponding change in the angle of the head of the user, and/or wherein the processing module is configured to generate the control signal such that the robotic adjustment system is triggered to adjust an angle of view of a microscope of the microscope system if the position of the head of the user changes with a corresponding change in the angle of the head of the user.

10. An apparatus (110; 420) for controlling a microscope system (100; 400), the appa ratus comprising: an interface (112) for communicating with a camera module (120), the camera mod ule being suitable for providing camera image data of a head of a user of the micro scope system; a processing module (114) configured to: obtain the camera image data from the camera module via the interface, process the camera image data to determine information on an angular orientation of the head of the user and to determine a fixation of the user on a point, provide a control signal for a robotic adjustment system (130) of the microscope sys tem based on the information on the angular orientation of the head of the user such that a point of view of a microscope (150; 410) of the microscope system remains fixed on the point.

11. The apparatus according to claim 10, wherein the processing module is configured to determine the information on the angular orientation of the head of the user of the microscope relative to a display (140) of the microscope system.

12. The apparatus according to claim 10, wherein the microscope system is a surgical microscope system, wherein the processing module is configured to determine the information on the angular orientation of the head of the user of the microscope rela tive to a surgical site being observed by a surgical microscope (150; 410) of the sur gical microscope system.

13. A method for controlling a microscope system, the method comprising: obtaining (210) camera image data of a head of a user of a head of a user of the mi croscope system from a camera module of the microscope system; processing (220) the camera image data to determine information on an angular ori entation of the head of the user relative to a display of the microscope system; and providing (230) a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user.

14. A method for controlling a microscope system, the method comprising: obtaining (310) camera image data of a head of a user of a head of a user of the mi croscope system from a camera module of the microscope system; processing (320) the camera image data to determine information on an angular ori entation of the head of the user and to determine a fixation of the user on a point; providing (330) a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user such that a point of view of a microscope (150) of the microscope system remains fixed on the point.

15. A computer program with a program code for performing at least one of the methods according to one of the claims 13 or 14 when the computer program is executed on a processor.

Description:
Apparatuses, Methods and Computer Programs for Controlling a Microscope System

Technical field

Examples relate to apparatuses, methods and computer programs for controlling a micro scope system, and to a corresponding microscope system.

Background

Surgical microscopes systems are often designed so that an optical carrier (i.e. the surgical microscope) can be moved as easy and smoothly as possible, ideally giving the sense of lack of gravity. In some cases, a robotic adjustment system, such as a robotic arm, is used to move the surgical microscope. The movement is often performed using knobs and buttons, which are used to individually control a position of the surgical microscope, an angle of the surgical microscope and other functionality, such as a zoom level, illumination etc. Due to the many factors that can be adjusted in order to obtain the desired field of view of the sur gical microscope, controlling the robotic adjustment system may take up time and concen tration of the surgeon. US patent 5,345,087 shows an optical guide system for spatially posi tioning a surgical microscope, which provides a contactless guidance of the instrument. In the patent, light sources are attached to the head of the surgeon and are sensed to determine the orientation of the head of the surgeon, and correspondingly control a motorized adjust ment mechanism. The head is thereby, in an initialization phase, brought to a pre-defmed position, and the movement of the head is tracked relative to this pre-defmed position. This provides a basic contactless control of the motorized adjustment mechanism that does not consider the mechanics that humans employ when taking different visual perspectives.

There may be a desire for an improved microscope system, that provides an improved con trol of the robotic adjustment system.

Summary

Embodiments of the present disclosure are based on the finding that other systems merely provide a purely mechanical movement equivalent to the head movement of the surgeon, instead of determining the perspective that the surgeon desires to gain when performing a head movement. In embodiments, the angular orientation of the head of the user is deter mined relative to a point - relative to a display, or relative to the surgical site, and this angu lar orientation is used to determine which perspective the operator is trying to take vis-a-vis the point, and how the robotic adjustment system can be adjusted to achieve this desired change in perspective. Embodiments may thus use the head movements, gestures and ex pressions to control the microscope, providing head tracking for robot control in a intuitive manner. One embodiment is the adjustment of a microscope's alignment by robot which follows the head of the user.

Embodiments of the present disclosure provide an apparatus for controlling a microscope system. The apparatus comprises an interface for communicating with a camera module. The camera module is suitable for providing camera image data of a head of a user of the microscope system. The apparatus comprises a processing module configured to obtain the camera image data from the camera module via the interface. The processing module is con figured to process the camera image data to determine information on an angular orientation of the head of the user relative to a display of the microscope system. The processing mod ule is configured to provide a control signal for a robotic adjustment system of the micro scope system based on the information on the angular orientation of the head of the user. By determining the angular orientation of the head of the user relative to a display of the micro scope system, changes of the orientation of the head relative to the display can be translated into a change of perspective of the microscope system in a comprehensible manner, improv ing a control of the microscope system by the user/surgeon of the microscope system.

In at least some embodiments, the processing module is configured to determine an angle of view of the user of the microscope system towards the display based on the camera image data. The processing module may be configured to provide the control signal based on the angle of view of the user. The angle of view may be used to adjust the robotic arm such, that the microscope provides a field of view that is derived from the angle of view.

For example, the processing module may be configured to determine the angle of view of the user based on the angular orientation of the head of the user. The angular orientation may provide a coarse-grained determination of angle of view. In some embodiments, the processing module is configured to perform eye tracking on one or both eyes of the user. The processing module may be configured to determine the angle of view of the user based on the angular orientation of the head of the user and based on the eye tracking. This may provide a more fine-grained determination of the angle of view.

The processing module may be configured to determine the information on the angular ori entation using image analysis on the camera image data. This may enable a determination of the angular orientation without the user having to wear additional tracking markers.

In some embodiments, the display of the microscope system is an auxiliary display that is arranged in proximity of an eyepiece of a microscope of the microscope system. This may enable the surgeon an in-situ adjustment of the robotic arm, without having to leave the sur gical field.

Alternatively, the display of the microscope system may be an auxiliary display that is ar ranged at a base unit of the microscope system. This may enable the use of larger displays, which may be used in preparation of the surgical procedure or by additional surgical staff.

In some embodiments, the display of the microscope system may be an ocular display of an eyepiece of a microscope of the microscope system. Again, this may enable the surgeon an in-situ adjustment of the robotic arm, without having to leave the surgical site.

In various embodiments, the processing module is configured to determine a fixation of the user on a point based on the camera image data. The processing module may be configured to determine the control signal for the robotic adjustment system such that a (central) point of view of a microscope of the microscope system remains fixed on the point. The (central) point (e.g. the point, on which the fixation is focused) may serve as a visual anchor in the adjustment of the robotic arm.

The processing module may be configured to determine information on a position of the head of the user and information on an angle of the head of the user based on the camera image. The processing module may be configured to determine the information on the angu lar orientation of the head of the user based on the information on the position of the head of the user and based on the information on the angle of the head of the user. Both the position and the angle (e.g. absolute or relative to the display) may be factors in the determination of the angular orientation, as the angular orientation is defined relative to the display.

For example, the processing module may be configured to generate the control signal such that the robotic adjustment system is triggered to pan a field of view of a microscope of the microscope system if the position of the head of the user changes without a corresponding change in the angle of the head of the user. This may provide a natural translation between the movement of the position of the head and the orientation of the robotic arm.

Additionally or alternatively, the processing module may be configured to generate the con trol signal such that the robotic adjustment system is triggered to adjust an angle of view of a microscope of the microscope system if the position of the head of the user changes with a corresponding change in the angle of the head of the user. This may provide a comprehensi ble translation between a change of the angle of the head and the orientation of the robotic arm.

In some embodiments, the interface is suitable for obtaining a trigger signal from an input device of the microscope system. The processing module may be configured to provide the control signal in response to the trigger signal of the input device. Thus, the control signal might only be provided if explicitly triggered by the user, avoiding an undesired adjustment of the robotic arm.

For example, the control signal may be configured to adjust at least one of a position in space of a microscope of the microscope system, an observation angle of the microscope, a zoom of the microscope, a working distance of the microscope, and an illumination of the microscope. For example, multiple elements of the above group may be adjusted at the same time, providing a coordinated adjustment of the various properties of the microscope.

Embodiments further provide a further apparatus for controlling a microscope system. The apparatus comprises an interface for communicating with a camera module. The camera module is suitable for providing camera image data of a head of a user of the microscope system. The apparatus comprises a processing module configured to obtain the camera im age data from the camera module via the interface. The processing module is configured to process the camera image data to determine information on an angular orientation of the head of the user and to determine a fixation of the user on a point. The processing module is configured to provide a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user such that a (central) point of view of a microscope of the microscope system remains fixed on the point. By determining the angular orientation of the head of the user, changes of the orienta tion of the head relative to the display can be translated into a change of perspective of the microscope system in an organic manner, improving an interaction between the surgeon and the microscope system. The fixation on the user on the point may provide a visual anchor for the adjustment of the robotic arm.

For example, the processing module may be configured to determine the information on the angular orientation of the head of the user of the microscope relative to a display of the mi croscope system. In other words, the display may be used as a point of reference for deter mining the angular orientation, providing the surgeon with a reference in the adjustment of the robotic arm.

Alternatively, the microscope system may be a surgical microscope system. The processing module may be configured to determine the information on the angular orientation of the head of the user of the microscope relative to a surgical site being observed by a surgical microscope of the surgical microscope system. In this case, the surgeon may control the robotic arm by adjusting their angular orientation towards the surgical site, which is then translated to provide an equivalent orientation of the microscope.

Embodiments of the present disclosure further provide a microscope system comprising the one or both of the above apparatuses.

Embodiments of the present disclosure further provide a method for controlling a micro scope system. The method comprises obtaining camera image data of a head of a user of a head of a user of the microscope system from a camera module of the microscope system. The method comprises processing the camera image data to determine information on an angular orientation of the head of the user relative to a display of the microscope system. The method comprises providing a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user. Embodiments of the present disclosure further provide a method for controlling a micro scope system. The method comprises obtaining camera image data of a head of a user of a head of a user of the microscope system from a camera module of the microscope system. The method comprises processing the camera image data to determine information on an angular orientation of the head of the user and to determine a fixation of the user on a point. The method comprises providing a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user such that a (central) point of view of a microscope of the microscope system remains fixed on the point.

Embodiments of the present disclosure further provide a computer program with a program code for performing at least one of the methods when the computer program is executed on a processor.

Short description of the Figures

Some examples of apparatuses and/or methods will be described in the following by way of example only, and with reference to the accompanying figures, in which

Figs la and lb show a block diagrams of embodiments of an apparatus for controlling a microscope system and of a microscope system comprising such an apparatus, re spectively;

Fig. 2 shows a flow chart of an embodiment of a method for controlling a microscope sys tem;

Fig. 3 shows a flow chart of a further embodiment of a method for controlling a micro scope system; and

Fig. 4 shows a schematic diagram of a system comprising a microscope and a computer system. Detailed Description

Various examples will now be described more fully with reference to the accompanying drawings in which some examples are illustrated. In the figures, the thicknesses of lines, layers and/or regions may be exaggerated for clarity.

Figs la and lb show a block diagram of embodiments of an apparatus 110 for controlling a microscope system 100. The apparatus comprises an interface 112 for communicating with a camera module 120. The camera module is suitable for providing camera image data of a head of a user of the microscope system. The apparatus comprises a processing module 114, which is coupled to the interface 112. The processing module 114 is configured to obtain the camera image data from the camera module via the interface.

The camera image data is subsequently processed to determine information on an angular orientation of the head of the user, either relative to a display or relative to a point on which the gaze of the user is fixated (on the display, or on the surgical site). Accordingly, in some embodiments, the processing module 114 is configured to process the camera image data to determine information on an angular orientation of the head of the user relative to a display 140 of the microscope system. In this case, the processing module is configured to provide a control signal for a robotic adjustment system 130 of the microscope system based on the information on the angular orientation of the head of the user. Alternatively or additionally, the processing module 114 may be configured to process the camera image data to deter mine information on an angular orientation of the head of the user and to determine a fixa tion of the user on a point. In this case, the processing module is configured to provide the control signal for a robotic adjustment system 130 of the microscope system based on the information on the angular orientation of the head of the user such that a central point of view of a microscope 150 of the microscope system remains fixed on the point.

Fig. lb further shows the microscope system 100 comprising the apparatus 110, the camera module 120, the robotic adjustment system 130, optionally the display 140, further optional ly the microscope 150, and further optionally an input device 160. The following description relates to the control apparatus 110 and microscope system 100 of Figs la and lb, and to the corresponding methods of Figs. 2 and 3.

Embodiments of the present disclosure relate to apparatuses, method and computer pro grams for controlling a microscope system 100. In general, a microscope is an optical in strument that is suitable for examining objects that are too small to be examined by the hu man eye (alone). For example, a microscope may provide an optical magnification of an object, such as the sample of organic tissue. A surgical microscope is a microscope that is used during surgery, i.e. a microscope that is suitable for use (by a surgeon) during surgery. Such surgical microscope systems often comprise an arm or a positioning means that is used to position the surgical microscope as desired by the surgeon, e.g. close to the operating space, so the surgical microscope can be used to provide a magnified view of the wound tract or tissue. Such arms or positioning means are usually adapted to provide a wide variety of different positions and/or angles, in order to provide the surgeon with enough space to conduct the surgery, while providing the magnified image of the site of the surgery. Addi tionally, the arms or positioning means of surgical microscope system are often designed so they allow for easy positioning of the surgical microscope, e.g. using weights and counter weights, using pneumatic systems or using motors to support the movement of the arm or positioning means of the surgical microscope system. As has been laid out above, many microscope systems, such as surgical microscope systems, have a robotic adjustment sys tem, such as a robotic arm, for positioning the microscope of the microscope system relative to a sample to be observed using the microscope system. In other words, the robotic adjust ment system may be configured to adjust the arm of the microscope system. For example, the robotic adjustment system may be a robotic arm of the (surgical) microscope system. Such robotic adjustment systems often offer a large number of degrees of freedom regarding the adjustment of the system - in many cases, an angular orientation of the microscope rela tive to the sample can be adjusted as well as a three-dimensional position of the microscope. Accordingly, adjustments to the robotic arms often are not straightforward. In some sys tems, the user/operator, e.g. the surgeon, can grab handles that are attached to the micro scope and manually reposition of the microscope, aided by the robotic adjustment system, which operate as servo motors in this case, facilitating the movement of the machinery, which is often heavy. In embodiments, however, a control of the robotic adjustment system is chosen that is based on the angular orientation of the head of the user, which is derived by processing of camera image data. In other words, the microscope system may be a surgical microscope system having a robotic adjustment system, such as a robotic arm, which may be controlled based on the angular orientation of the head of the user.

The microscope system comprises the camera module 120, which is suitable for providing camera image data of a head of a user of the microscope system. In other words, the camera module 120 may be configured to provide the camera image data of the head of the user of the microscope system. In general, the camera module may comprise an APS (Active Pixel Sensor) - or a CCD (Charge-Coupled-Device)-based imaging sensor module. For example, in APS-based imaging sensor modules, light is recorded at each pixel using a photodetector and an active amplifier of the pixel. APS-based imaging sensor modules are often based on CMOS (Complementary Metal-Oxide-Semiconductor) or S-CMOS (Scientific CMOS) technology. In CCD-based imaging sensor modules, incoming photons are converted into electron charges at a semiconductor-oxide interface, which are subsequently moved between capacitive bins in the imaging sensor modules by a control circuitry of the sensor imaging module to perform the imaging. Alternatively, or additionally, the camera module may be a depth-sensing camera module or comprise a depth sensor, suitable for providing a depth sensing camera module image. Accordingly, the camera image data may be a depth-sensing camera image data or comprise a two-dimensional and a depth-sensing component. For ex ample, the camera module may comprise a depth sensor, e.g. a Time of Flight-based depth sensor or a structured light-based depth sensor. The camera image data may comprise two- dimensional camera image data of the head of the user and/or three-dimensional camera image data of the head of the user.

The processing module 114 is configured to process the camera image data to determine the information on an angular orientation of the head of the user, e.g. relative to the display 140 or relative to the surgical site. In general, the angular orientation of the head of the user may comprise an angle of the head relative to a point of reference, e.g. relative to a resting posi tion of the head, relative to the display 140 or relative to the surgical site. Additionally, the angular orientation may comprise a position of the head relative to a coordinate system, which may be relative to the microscope system, relative to the display or relative to the surgical site. In other words, the processing module may be configured to determine infor mation on a position of the head of the user and information on an angle of the head of the user (e.g. relative to a point of reference) based on the camera image. The processing mod ule may be configured to determine the information on the angular orientation of the head of the user based on the information on the position of the head of the user and based on the information on the angle of the head of the user. The information on the angular orientation may comprise numerical values of the angle of the head (relative to the point of reference) and/or of the position of the head.

The processing of the camera image data may be performed using image analysis. In other words, the processing module may be configured to determine the information on the angu lar orientation using image analysis on the camera image data. For example, the image anal ysis may be used identify one or more features of the head within the camera image data. For example, the processing module may be configured to determine the angular orientation of the head of the user by determining an outline of the head of the user, by identifying eyes and/or a nose within the (two-dimensional or three-dimensional) camera image data, and by deriving the angular orientation of the head of the user based on the outline of the head and based on the position of the eyes and/or the nose relative to the outline of the head. The im age analysis may be based on the (natural) outline of the head. For example, the image anal ysis might not be based on tracking markers being attached to the head of the user. On the contrary - the processing module may be configured to determine the information on the angular orientation of the head of the user without identifying (artificially added) tracking markers.

In various embodiments, the information on the angular orientation of the head of the user is determined relative to a display 140 of the microscope system. For example, the display 140 may be an electronic output device for presentation of information in visual form. For ex ample, the display 140 may be suitable for and/or configured to electronically display an image of a sample to be observed by the microscope, e.g. an image of a surgical site to be observed by the surgical microscope. The processing module 114 may be configured to pro vide the image of the sample to be observed to the display 114. In at least some embodi ments, the image is based on further camera image data of a further camera (not shown) of the microscope 150. For example, the processing module 114 may be configured to obtain the further camera image data from the further camera and to generate the image for the display 114. For example, the display 140 may be a Liquid-Crystal-Display (LCD)-based display or an Organic Light Emitting Diode (OLED)-based display. In some embodiments, the display may be arranged at or within the eyepiece of the microscope 150 of the micro scope system. For example, the display may be arranged above the eyepiece of the micro- scope, or the display may be integrated within the eyepieces of the microscope. In other words, the display of the microscope system may be an auxiliary display 140a that is ar ranged in proximity of an eyepiece of a microscope of the microscope system. For example, the display 140a; b may be attached to the microscope 150 of the microscope system. For example, the display of the microscope system may be an ocular display 140b of an eye piece of a microscope of the microscope system. Alternatively, the display of the micro scope system may be an auxiliary display 140c that is arranged at a base unit (e.g. a stand comprising the apparatus 110) of the microscope system. For example, the display 140b may be attached to the base unit of the microscope system via a further arm 145.

As laid out above, in some embodiments, the angular orientation of the head of the user is determined relative to the display 140 or relative to the surgical site being observed by the surgical microscope. For example, this may be achieved by determine the angle of view of the user towards the display or surgical site. In other words, the processing module may be configured to determine an angle of view of the user of the microscope system towards the display (or towards the surgical site) based on the camera image data. For example, the pro cessing module may be configured to derive the angle of view of the user towards the dis play or surgical site based on the outline of the head and based on the position of the eyes and/or the nose relative to the outline of the head, and based on the position of the head of the user. For example, the processing module may be configured to determine the angle of view of the user based on the angular orientation of the head of the user. For example, the angular orientation of the had of the user may comprise the angle of view of the user to wards the display or surgical site, e.g. if an angular component of the angular orientation is determined relative to the display or surgical site as point of reference. In some embodi ments, this may suffice to determine an angle of view that is precise enough to perform the adjustment of the robotic adjustment system. In some embodiments, however, the pro cessing module may be configured to perform eye-tracking in order to more precisely learn where the user (i.e. the surgeon) is looking. In other words, the processing module may be configured to perform eye tracking on one or both eyes of the user. For example, the pro cessing module may be configured to perform image processing on the camera image data in order to perform eye tracking on one or both eyes of the user. Alternatively, the image processing may be performed on further camera image data of a dedicated eye tracking camera, e.g. in addition to three-dimensional camera image data that is used to determine the angular orientation. The processing module 114 may be configured to determine the angle of view of the user based on the angular orientation of the head of the user and based on the eye tracking. In other words, the processing module 114 may be configured to refine the determined angle of view of the user based on the eye tracking.

The processing module may be configured to provide the control signal based on the angle of view of the user. For example, the processing module 114 may be configured to translate the angle of view of the user into a corresponding orientation of the microscope 150, and to provide the control signal to affect the corresponding orientation of the microscope 150. For example, the control signal may be provided such, that the robotic adjustment system is af fected to arrange the microscope 150 at an angle that is based on the angle of view of the user, e.g. at an angle that corresponds to the angle of view of the user.

When changing the positioning of the microscope, a point (of the sample, or of the surgical site) may be used as “anchor point”, which remains fixed while the angle of view or the distance/zoom level is changed, e.g. so the user can concentrate on a point of interest and view it from different angles. Therefore, a fixation of the user on the point may be deter mined, and the point may serve as anchor point in the adjustment of the microscope (via the robotic adjustment module. In other words, the processing module may be configured to determine a fixation of the user on a point (e.g. a point on the screen, a point of the sam ple/surgical site shown on the screen, or a point of the surgical site) based on the camera image data. The processing module may be configured to process the camera image data to determine the fixation of the user on the point. For example, the processing module may be configured to determine the fixation of the user on the point using eye tracking, e.g. using the angular orientation of the head of the user in conjunction with eye tracking. The pro cessing module may be configured to determine the control signal for the robotic adjustment system such that a central point of view of the microscope 150 of the microscope system remains fixed on the point. In other words, the processing module may be configured to determine the control signal for the robotic adjustment system such, that the point remains at a similar (e.g. the same position) position within the field of view of the microscope 150.

In general, different approaches may be taken depending on whether the user merely chang es the position of the head or whether the user changes the angle of the head. For example, if the user merely changes the position of its head without changing the angle (or with a change of the angle that is lower than a threshold, e.g. lower than 5 degrees), the control signal may be generated such that the field of view of the microscope pans (i.e. scrolls hori zontally and/or vertically without changing the angle of view) over the sample/surgical field. In other words, the processing module may be configured to generate the control sig nal such that the robotic adjustment system is triggered to pan a field of view of the micro scope of the microscope system if the position of the head of the user changes without a corresponding change in the angle of the head of the user. If merely the angle of the head is changed, the angle of view may be adjusted without panning. In other words, the processing module may be configured to generate the control signal such that the robotic adjustment system is triggered to adjust an angle of view of the microscope of the microscope system if the position of the head of the user changes with a corresponding change in the angle of the head of the user. Again, the processing module may be configured to determine the control signal for the robotic adjustment system such that a central point of view of the microscope 150 of the microscope system remains fixed on a point within the field of view, e.g. based on a point on which the user is fixated, or based on a central point of the field of view. In other words, the processing module may be configured to determine the control signal for the robotic adjustment system such, that a point, e.g. a point on which the user is fixated, or a central point of the field of view, remains at a similar (e.g. the same position) position within the field of view of the microscope 150. In this case, the change in the angle of view may also result in a panning of the field of view, e.g. based on the point referenced above. Furthermore, in some embodiments, the processing module may be configured to determine information on a position, posture and/or movement of shoulders of the user based on the camera image data, and use the information on the position, posture and/or movement of the shoulders to generate the control signal.

The processing module is configured to provide the control signal for the robotic adjustment system 130 of the microscope system based on the information on the angular orientation of the head of the user. As mentioned above, the processing module 114 may be configured to translate the angle of view of the user into a corresponding orientation of the microscope 150, and to provide the control signal to affect the corresponding orientation of the micro scope 150. Various properties of the robotic adjustment system may be adjusted in order to affect the orientation of the microscope 150. For example, the control signal may be config ured to adjust at least one of a position in space of the microscope 150 of the microscope system, an observation angle of the microscope, a zoom of the microscope, a working dis tance of the microscope, and an illumination of the microscope. For example, the control signal may be configured to adjust the position in space, and thereby the working distance, of the microscope 150 by controlling the robotic adjustment system such, that the position in space of the microscope is changed, e.g. by controlling robotic joints of the robotic arm in order to obtain the desired position in space. The control signal may be configured to adjust the observation angle of the microscope by controlling the robotic adjustment system such, that the position in space of the microscope is changed and such that an angle of the micro scope towards the sample or surgical site is changed to obtain the observation angle. The control signal may further be configured to control the microscope to change the zoom and/or the illumination of the microscope. Accordingly, the control signal may also be pro vided to the microscope 150 or to an illumination unit of the microscope system 150.

In some cases, it may be desirable to limit the provision of the control signal to situations, in which the user/surgeon wishes to change the position of the microscope, and to refrain from generating or providing the control signal outside these situations. To achieve this, the us er/surgeon may provide a trigger via an input device in order to activate (or deactivate) the generation/provision of the control signal. Accordingly, the interface may be suitable for obtaining a trigger signal from an input device 160 of the microscope system. For example, the input device may be one of a button, a touch-based interface (such as a touchscreen or a capacitive switch), a light-based input device (e.g. by blocking a path between light source and receiver, the input device is actuated), an ultrasound-based input device (e.g. by bring ing a hand or object close to the ultrasound-based input device, the input device is actuated), and a voice-activated input device. The processing module may be configured to provide the control signal in response to the trigger signal of the input device. In other words, the pro cessing module may be configured to provide the control signal (only) if triggered by the input device (and/or as long it is triggered by the input device).

The interface 112 may correspond to one or more inputs and/or outputs for receiving and/or transmitting information, which may be in digital (bit) values according to a specified code, within a module, between modules or between modules of different entities. For example, the interface 112 may comprise interface circuitry configured to receive and/or transmit information. In embodiments the processing module 114 may be implemented using one or more processing units, one or more processing devices, any means for processing, such as a processor, a computer or a programmable hardware component being operable with accord ingly adapted software. In other words, the described function of the processing module 114 may as well be implemented in software, which is then executed on one or more program mable hardware components. Such hardware components may comprise a general-purpose processor, a Digital Signal Processor (DSP), a micro-controller, etc.

More details and aspects of the microscope system or the apparatus for controlling the mi croscope system are mentioned in connection with the proposed concept or one or more examples described above or below (e.g. Fig. 2 to 4). The microscope system or the appa ratus for controlling the microscope system may comprise one or more additional optional features corresponding to one or more aspects of the proposed concept or one or more ex amples described above or below.

Fig. 2 shows a flow chart of an embodiment of a (corresponding) method for controlling a microscope system. For example, the microscope system may be implemented similar to the microscope system of Fig. lb. The method comprises obtaining 210 camera image data of a head of a user of a head of a user of the microscope system from a camera module of the microscope system. The method comprises processing 220 the camera image data to deter mine information on an angular orientation of the head of the user relative to a display 140 of the microscope system. The method comprises providing 230 a control signal for a robot ic adjustment system of the microscope system based on the information on the angular ori entation of the head of the user.

As indicated above, features described in connection with the apparatus 110 and the micro scope system 100 of Figs la and/or lb may be likewise applied to the method of Fig. 2.

More details and aspects of the method are mentioned in connection with the proposed con cept or one or more examples described above or below (e.g. Fig. 1 or 3 to 4). The method may comprise one or more additional optional features corresponding to one or more as pects of the proposed concept or one or more examples described above or below.

Fig. 3 shows a flow chart of a further embodiment of a (corresponding) method for control ling a microscope system. For example, the microscope system may be implemented similar to the microscope system of Fig. lb. The method comprises obtaining 310 camera image data of a head of a user of a head of a user of the microscope system from a camera module of the microscope system. The method comprises processing 320 the camera image data to determine information on an angular orientation of the head of the user and to determine a fixation of the user on a point. The method further comprises providing 330 a control signal for a robotic adjustment system of the microscope system based on the information on the angular orientation of the head of the user such that a central point of view of a microscope of the microscope system remains fixed on the point.

As indicated above, features described in connection with the apparatus 110 and the micro scope system 100 of Figs la and/or lb may be likewise applied to the method of Fig. 3.

More details and aspects of the method are mentioned in connection with the proposed con cept or one or more examples described above or below (e.g. Fig. 1, 2 or 4). The method may comprise one or more additional optional features corresponding to one or more as pects of the proposed concept or one or more examples described above or below.

At least some embodiments relate to the use of an imaging system that is attached to a ro botic arm. The use of an imaging system (e.g. surgical microscope) attached on a robotic arm, which moves the imaging system in space (typically 6 degrees of freedom) allows many adjustments of the imaging settings, such as:

1. Position in space, x,y,z (3 degrees of freedom)

2. Observation angle q,f,w (3 degrees of freedom)

3. Zoom, working distance, illumination/sensitivity (3 degrees of freedom)

Such adjustment may be complex when performed with conventional knob/button adjust ments. Moreover, such adjustments may require time and the attention of the user. Moreo ver, it may be demanding to perform complex adjustments such as point lock (fix the obser vation center, while changing the observation angle), which may require a combined ad justment of parameters (e.g. positioning lies on a sphere, while observation angle adapts accordingly).

Human vision, however, is mechanically/geometrically similar to the robotic imaging sys tem, and humans intuitively adjust the observation geometry in order to perceive the shape of an object. Embodiments provide a control system (e.g. the apparatus 110 of Figs la and/or lb) which allows humans to control a robotic imaging system (e.g. the microscope system 100 of Figs la and/or lb) in an intuitive way which resembles the natural observa- tion motions/gestures. In particular, embodiments may focus on the recognition of the hu man’s posture and movement of head (e.g. the angular orientation of the head) and interpret it as a guide for the arrangement of the robotic imaging system (e.g. by providing the con trol signal). Such a system could be any combination of the following concepts:

1. Replication of the human head observation geometry, i.e. head position (x,y,z) and angle ( q,f,w ) (i.e. the angular orientation). In addition, also eye-tracking information may be included.

2. Interpretation of the movements/gestures of the human head in order to perform cer tain adjustments. For example, when head is moving in space (x,y,z) without turning then it could mean pan of the field of view, while when (x,y,z) movement is com bined with counterbalancing turning of the head {q,f,w), then it may be interpreted as point-lock adjustment. In addition, eye-tracking and shoulder position could also be used for a more robust interaction.

More details and aspects of the concept are mentioned in connection with the proposed con cept or one or more examples described above or below (e.g. Fig. 1 to 3). The concept may comprise one or more additional optional features corresponding to one or more aspects of the proposed concept or one or more examples described above or below.

Some embodiments relate to a microscope comprising a system as described in connection with one or more of the Figs. 1 to 3. Alternatively, a microscope may be part of or connect ed to a system as described in connection with one or more of the Figs. 1 to 3. Fig. 4 shows a schematic illustration of a system 400 configured to perform a method described herein. The system 400 comprises a microscope 410, which may correspond to the microscope 150 of Fig. lb, and a computer system 420, which may correspond to the apparatus 110 of Figs la and/or lb. The microscope 410 is configured to take images and is connected to the com puter system 420. The computer system 420 is configured to execute at least a part of a method described herein. The computer system 420 may be configured to execute a ma chine learning algorithm. The computer system 420 and microscope 410 may be separate entities but can also be integrated together in one common housing. The computer system 420 may be part of a central processing system of the microscope 410 and/or the computer system 420 may be part of a subcomponent of the microscope 410, such as a sensor, an ac tor, a camera or an illumination unit, etc. of the microscope 410. The computer system 420 may be a local computer device (e.g. personal computer, laptop, tablet computer or mobile phone) with one or more processors and one or more storage de vices or may be a distributed computer system (e.g. a cloud computing system with one or more processors and one or more storage devices distributed at various locations, for exam ple, at a local client and/or one or more remote server farms and/or data centers). The com puter system 420 may comprise any circuit or combination of circuits. In one embodiment, the computer system 420 may include one or more processors which can be of any type. As used herein, processor may mean any type of computational circuit, such as but not limited to a microprocessor, a microcontroller, a complex instruction set computing (CISC) micro processor, a reduced instruction set computing (RISC) microprocessor, a very long instruc tion word (VLIW) microprocessor, a graphics processor, a digital signal processor (DSP), multiple core processor, a field programmable gate array (FPGA), for example, of a micro scope or a microscope component (e.g. camera) or any other type of processor or processing circuit. Other types of circuits that may be included in the computer system 420 may be a custom circuit, an application-specific integrated circuit (ASIC), or the like, such as, for ex ample, one or more circuits (such as a communication circuit) for use in wireless devices like mobile telephones, tablet computers, laptop computers, two-way radios, and similar electronic systems. The computer system 420 may include one or more storage devices, which may include one or more memory elements suitable to the particular application, such as a main memory in the form of random access memory (RAM), one or more hard drives, and/or one or more drives that handle removable media such as compact disks (CD), flash memory cards, digital video disk (DVD), and the like. The computer system 420 may also include a display device, one or more speakers, and a keyboard and/or controller, which can include a mouse, trackball, touch screen, voice-recognition device, or any other device that permits a system user to input information into and receive information from the computer system 420.

Some or all of the method steps may be executed by (or using) a hardware apparatus, like for example, a processor, a microprocessor, a programmable computer or an electronic cir cuit. In some embodiments, some one or more of the most important method steps may be executed by such an apparatus.

Depending on certain implementation requirements, embodiments of the invention can be implemented in hardware or in software. The implementation can be performed using a non- transitory storage medium such as a digital storage medium, for example a floppy disc, a DVD, a Blu-Ray, a CD, a ROM, a PROM, and EPROM, an EEPROM or a FLASH memory, having electronically readable control signals stored thereon, which cooperate (or are capable of cooperating) with a programmable computer system such that the respective method is performed. Therefore, the digital storage medium may be computer readable.

Some embodiments according to the invention comprise a data carrier having electronically readable control signals, which are capable of cooperating with a programmable computer system, such that one of the methods described herein is performed.

Generally, embodiments of the present invention can be implemented as a computer pro gram product with a program code, the program code being operative for performing one of the methods when the computer program product runs on a computer. The program code may, for example, be stored on a machine readable carrier.

Other embodiments comprise the computer program for performing one of the methods de scribed herein, stored on a machine readable carrier.

In other words, an embodiment of the present invention is, therefore, a computer program having a program code for performing one of the methods described herein, when the com puter program runs on a computer.

A further embodiment of the present invention is, therefore, a storage medium (or a data carrier, or a computer-readable medium) comprising, stored thereon, the computer program for performing one of the methods described herein when it is performed by a processor. The data carrier, the digital storage medium or the recorded medium are typically tangible and/or non-transitionary. A further embodiment of the present invention is an apparatus as described herein comprising a processor and the storage medium.

A further embodiment of the invention is, therefore, a data stream or a sequence of signals representing the computer program for performing one of the methods described herein. The data stream or the sequence of signals may, for example, be configured to be transferred via a data communication connection, for example, via the internet. A further embodiment comprises a processing means, for example, a computer or a pro grammable logic device, configured to, or adapted to, perform one of the methods described herein.

A further embodiment comprises a computer having installed thereon the computer program for performing one of the methods described herein.

A further embodiment according to the invention comprises an apparatus or a system con figured to transfer (for example, electronically or optically) a computer program for per forming one of the methods described herein to a receiver. The receiver may, for example, be a computer, a mobile device, a memory device or the like. The apparatus or system may, for example, comprise a file server for transferring the computer program to the receiver.

In some embodiments, a programmable logic device (for example, a field programmable gate array) may be used to perform some or all of the functionalities of the methods de scribed herein. In some embodiments, a field programmable gate array may cooperate with a microprocessor in order to perform one of the methods described herein. Generally, the methods are preferably performed by any hardware apparatus.

As used herein the term “and/or” includes any and all combinations of one or more of the associated listed items and may be abbreviated as ‘7”.

Although some aspects have been described in the context of an apparatus, it is clear that these aspects also represent a description of the corresponding method, where a block or device corresponds to a method step or a feature of a method step. Analogously, aspects described in the context of a method step also represent a description of a corresponding block or item or feature of a corresponding apparatus. List of reference Signs

100 Microscope system

110 Apparatus for controlling a microscope system

112 Interface

114 Processing module

120 Camera module

130 Robotic adjustment system

140 Display (suffixes a-c)

145 Arm

150 Microscope

155 Arm

160 Input device

210 Obtaining camera image data

220 Processing the camera image data

230 Providing a control signal

310 Obtaining camera image data

320 Processing the camera image data

330 Providing a control signal

400 System

410 Microscope

420 Computer system