Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
VACUUM CLEANER
Document Type and Number:
WIPO Patent Application WO/2022/008872
Kind Code:
A1
Abstract:
A vacuum cleaner comprises: a sensor configured to generate sensor signals based on sensed motion and orientation of the vacuum cleaner; a human-computer interface, HCI; and a controller configured to: process the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner: and control the HCI to provide a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

Inventors:
CAMPLANI MASSIMO (GB)
WATSON ANDREW (GB)
MILLINGTON DAVID (GB)
LAWSON MCLEAN NATHAN (GB)
Application Number:
PCT/GB2021/051376
Publication Date:
January 13, 2022
Filing Date:
June 03, 2021
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
DYSON TECHNOLOGY LTD (GB)
International Classes:
A47L5/22; A47L5/26; A47L5/30; A47L9/28
Foreign References:
GB2578872A2020-06-03
DE102015108464A12016-12-01
US20190387943A12019-12-26
Attorney, Agent or Firm:
MITCHELL, Joshua et al. (GB)
Download PDF:
Claims:
Claims

1. A vacuum cleaner comprising: a sensor configured to generate sensor signals based on sensed motion and orientation of the vacuum cleaner; a human-computer interface, HCI; and a controller configured to: process the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and control the HCI to provide a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

2. The vacuum cleaner of claim 1, wherein the HCI comprises a visual display unit, and wherein the recommendation comprises a visual recommendation.

3. The vacuum cleaner of claim 1 or 2, wherein the HCI comprises an audio output device, and wherein the recommendation comprises an audible recommendation.

4. The vacuum cleaner of any preceding claim, wherein the recommendation comprises information relating to a cleaning tool suitable for the determined type of cleaning activity.

5. The vacuum cleaner of any preceding claim, wherein the recommendation comprises information relating to one or more of: a stroke rate; a dwell time; and an applied pressure, each being suitable for the determined type of cleaning activity'.

6. The vacuum cleaner of any preceding claim, wherein the sensor signals are based only on sensed motion of the vacuum cleaner or only on sensed orientation of the vacuum cleaner.

7. The vacuum cleaner of any preceding claim, wherein the sensor comprises an inertial measurement unit, IMU.

8. The vacuum cleaner of any preceding claim, further comprising: a cleaner head comprising an agitator; and one or more diagnostic sensors configured to generate further sensor signals based on sensed parameters of the cleaner head, wherein the controller is configured to process the generated further sensor signals to determine the type of cleaning activity being performed by the user using the vacuum cleaner.

9. The vacuum cleaner of claim 8, wherein the cleaner head further comprises an agitator motor arranged to rotate the agitator, and wherein the sensed parameters of the cleaner head comprise the agitator motor current.

10. The vacuum cleaner of claim 8 or 9, wherein the sensed parameters of the cleaner head comprise the pressure applied to the cleaner head.

11. The vacuum cleaner of any one of claims 8 to 10, wherein the cleaner head further comprises an adjustable gate, and wherein the recommendation comprises information relating to a setting of the adjustable gate suitable for the determined type of cleaning activity.

12. The vacuum cleaner of any preceding claim, wherein the controller is configured to process the sensor signals by performing a pre-processing step and a classification step.

13. The vacuum cleaner of claim 12, wherein the pre-processing step comprises extracting features from time portions of the sensor signals.

14. The vacuum cleaner of claim 12 or 13, wherein the pre-processing step comprises filtering the sensor signals.

15. The vacuum cleaner of claim 13 or 14, wherein the classification step comprises processing the extracted features using a machine learning classifier.

16. The vacuum cleaner of claim 15, wherein the machine learning classifier comprises one or more of: an artificial neural network, a random forest and a support-vector machine.

17. A method of facilitating the use of a vacuum cleaner, the method comprising: generating sensor signals based on sensed motion and orientation of the vacuum cleaner; processing the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and providing a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

18. A computer program comprising a set of instructions, which, when executed by a computerised device, cause the computerised device to perform a method of facilitating the use of a vacuum cleaner, the method comprising: generating sensor signals based on sensed motion and orientation of the vacuum cleaner; processing the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and providing a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

Description:
Vacuum Cleaner

Technical Field

The present disclosure relates to a vacuum cleaner. In particular, but not exclusively, the present disclosure concerns measures, including methods, apparatus and computer programs, for operating a vacuum cleaner.

Background

Broadly speaking, there are four types of vacuum cleaner: ‘upright’ vacuum cleaners, ‘cylinder’ vacuum cleaners (also referred to as ‘canister’ vacuum cleaners), ‘handheld’ vacuum cleaners and ‘stick’ vacuum cleaners.

Upright vacuum cleaners and cylinder vacuum cleaners tend to be mains-power- operated.

Handheld vacuum cleaners are relatively small, highly portable vacuum cleaners, suited particularly to relatively low duty applications such as spot cleaning floors and upholstery in the home, interior cleaning of cars and boats etc. Unlike upright cleaners and cylinder cleaners, they are designed to be carried in the hand dining use, and tend to be powered by battery.

Stick vacuum cleaners may comprise a handheld vacuum cleaner in combination with a rigid, elongate suction wand which effectively reaches down to the floor so that the user may remain standing while cleaning a floor surface. A floor tool is typically attached to the end of the rigid, elongate suction wand, or alternatively may be integrated with the bottom end of the wand.

Stick vacuum cleaners can be used with a wide variety of detachable tools to facilitate different types of cleaning. Furthermore, the vacuum cleaner and some of the associated tools may have different settings, which can be varied by a user to suit different cleaning scenarios. However, users may at times use sub-optimal tools and/or settings for a particular type of cleaning activity being undertaken.

It is an object of the present disclosure to mitigate or obviate the above disadvantages, and/or to provide an improved or alternative vacuum cleaner.

Summary

According to an aspect of the present disclosure, there is provided a vacuum cleaner comprising: a sensor configured to generate sensor signals based on sensed motion and orientation of the vacuum cleaner; a human-computer interface, HCI; and a controller configured to: process the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and control the HCI to provide a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

Advantageously, the HCI provides the user of the vacuum cleaner with a recommendation which is dependent on the type of cleaning activity being performed, as determined by the controller. In this manner, if the user is operating the vacuum cleaner in a sub-optimal manner for the cleaning task being undertaken (e.g. using an incorrect or sub- optimal tool, tool setting or technique), the user will receive feedback, via the HCI, which will prompt the user to re-configure the vacuum cleaner in order to optimize cleaning performance and/or battery performance. This facilitates the user to learn the features of the vacuum cleaner overtime.

In embodiments, the HCI comprises a visual display unit and the recommendation comprises a visual recommendation.

In embodiments, the HCI comprises an audio output device and the recommendation comprises an audible recommendation.

In embodiments, the recommendation comprises information relating to a cleaning tool suitable for the determined type of cleaning activity.

In embodiments, the recommendation comprises information relating to one or more of: a stroke rate, a dwell time, and an applied pressure, each being suitable for the determined type of cleaning activity.

In embodiments, the sensor signals are based only on sensed motion of the vacuum cleaner or only on sensed orientation of the vacuum cleaner.

In embodiments, the sensor comprises an inertial measurement unit, IMU.

In embodiments, the vacuum cleaner further comprises a cleaner head comprising an agitator and one or more diagnostic sensors configured to generate further sensor signals based on sensed parameters of the cleaner head.

In embodiments, the controller is configured to process the generated further sensor signals to determine the type of cleaning activity being performed by the user using the vacuum cleaner. In this manner, when additional sensors are available, the additional sensor data are used by the controller to determine the cleaning activity being undertaken. This may improve the accuracy and/or speed at which the current cleaning activity is determined.

In embodiments, the cleaner head further comprises an agitator motor arranged to rotate the agitator and the sensed parameters of the cleaner head comprise the agitator motor current. In embodiments, the sensed parameters of the cleaner head comprise the pressure applied to the cleaner head.

In embodiments, the cleaner head further comprises an adjustable gate and the recommendation comprises information relating to a setting of the adjustable gate suitable for the determined type of cleaning activity.

In embodiments, the controller is configured to process the sensor signals by performing a pre-processing step and a classification step.

In embodiments, the pre-processing step comprises extracting features from time portions of the sensor signals.

In embodiments, the pre-processing step comprises filtering the sensor signals.

In embodiments, the classification step comprises processing the extracted features using a machine learning classifier. Advantageously, a machine learning classifier can be pretrained, for example at the factor)', by subjecting the vacuum cleaner to a multitude of different cleaning activities/scenarios and defining how the vacuum cleaner should respond in each case. Furthermore, the machine learning classifier may be capable of further learning in the user’s home environment.

In embodiments, the machine learning classifier comprises one or more of: an artificial neural network, a random forest and a support-vector machine.

According to an aspect of the present disclosure, there is provided a method of facilitating the use of a vacuum cleaner, the method comprising: generating sensor signals based on sensed motion and orientation of the vacuum cleaner; processing the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and providing a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

According to an aspect of the present disclosure, there is provided a computer program comprising a set of instructions, which, when executed by a computerised device, cause the computerised device to perform a method of facilitating the use of a vacuum cleaner, the method comprising: generating sensor signals based on sensed motion and orientation of the vacuum cleaner, processing the generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner; and providing a recommendation to the user of the vacuum cleaner in dependence on the determined type of cleaning activity.

The present disclosure is not limited to any particular type of vacuum cleaner. For example, the aspects of the disclosure may be utilised on upright vacuum cleaners, cylinder vacuum cleaners or handheld or 'stick 1 vacuum cleaners. It should be appreciated that features described in relation to one aspect of the present disclosure may be incorporated into other aspects of the present disclosure. For example, a method aspect may incorporate any of the features described with reference to an apparatus aspect and vice versa.

Description of the Drawings

Embodiments of the present disclosure will now be described by way of example only with reference to the accompanying schematic drawings of which:

Figure 1 is a perspective view of a stick vacuum cleaner according to an embodiment of the present disclosure;

Figure 2 is a view of a cleaner head of the vacuum cleaner of Figure 1, shown from underneath;

Figure 3 is a schematic illustration of electrical components of the vacuum cleaner of

Figure 1;

Figure 4 is a perspective view of a main body of the stick vacuum cleaner of Figure 1; Figures 5a and 5b illustrate sensor signals corresponding to linear and angular acceleration generated by an inertial measurement unit of a vacuum cleaner according to embodiments of the present disclosure;

Figures 6 and 7 illustrates further sensor signals corresponding to orientation generated by the inertial measurement unit of a vacuum cleaner according to embodiments of the present disclosure;

Figure 8 is a simplified schematic illustration of electrical components of the vacuum cleaner of Figure 3, showing electrical connections between sensors, a human-computer interface, motors and the controller according to embodiments of the present disclosure;

Figure 9 is a block diagram illustrating example sensor signal processing performed by the controller according to various embodiments of the present disclosure;

Figure 10 is a flow diagram showing a method of facilitating the use of a vacuum cleaner according to an embodiment of the present disclosure; and

Figures 1 la and 1 lb illustrate schematically the operation of the human computer interface of the vacuum cleaner of Figure 1 according to an embodiment of the present disclosure.

Detailed Description Figures 1 to 4 illustrate a vacuum cleaner 2 according to embodiments of the present disclosure. The vacuum cleaner 2 is a ‘stick’ vacuum cleaner comprising a cleaner head 4 connected to a main body 6 by a generally tubular elongate wand 8. The cleaner head 4 is also connectable directly to the main body 6 to transform the vacuum cleaner 2 into a handheld vacuum cleaner. Other removable tools, such as a crevice tool 3, a dusting brush 7 and a miniature motorized cleaner head 5 may be attached directly to the main body 6, or to the end of the elongate wand 8, to suit different cleaning tasks.

The main body 6 comprises a dirt separator 10 which in this case is a cyclonic separator. The cyclonic separator has a first cyclone stage 12 comprising a single cyclone, and a second cyclone stage 14 comprising a plurality of cyclones 16 arranged in parallel. The main body 6 also has a removable filter assembly 18 provided with vents 20 through which air can be exhausted from the vacuum cleaner 2. The main body 6 of the vacuum cleaner 2 has a pistol grip 22 positioned to be held by the user. At an upper end of the pistol grip 22 is a user input device in the form of a trigger switch 24, which is usually depressed in order to switch on the vacuum cleaner 2. However, in some embodiments the physical trigger switch 24 is optional. Positioned beneath a lower end of the pistol grip 22 is a battery pack 26 which comprises a plurality of rechargeable cells 27. A controller 50 and a vacuum motor 52, comprising a fan driven by an electric motor, are provided in the main body 6 behind the dirt separator 10.

The cleaner head 4 is shown from underneath in Figure 2. The cleaner head 4 has a casing 30 which defines a suction chamber 32 and a soleplate 34. The soleplate 34 has a suction opening 36 through which air can enter the suction chamber 32, and wheels 37 for engaging a floor surface. The casing 30 defines an outlet 38 through which air can pass from the suction chamber 32 into the wand 8. Positioned inside the suction chamber 32 is an agitator 40 in the form of a brush bar. The agitator 40 can be driven to rotate inside the suction chamber 32 by an agitator motor 54. The agitator motor 54 of this embodiment is received inside the agitator 40. The agitator 40 has helical arrays of bristles 43 projecting from grooves 42, and is positioned in tire suction chamber such that the bristles 43 project out of the suction chamber 34 through the suction opening 36.

Figure 3 is a schematic representation of the electrical components of the vacuum cleaner 2. The controller 50 manages the supply of electrical power from the cells 27 of the battery pack 26 to the vacuum motor 52. When the vacuum motor 52 is powered on, this creates a flow of air so as to generate suction. Air with dirt entrained therein is sucked into the cleaner head 4 (or, when attached, one of the other tools such as the crevice tool 3, the mini motorised cleaner head 5, or the dusting bmsh 7), into the suction chamber 32 through the suction opening 36. From there, the air is sucked through the outlet 38 of the cleaner head 4, along the wand 8 and into the dirt separator 10. Entrained dirt is removed by the dirt separator 10 and then relatively clean air is drawn through the vacuum motor 52, through the filter assembly 18 and out of the vacuum cleaner 2 through the vents 20. In addition, the controller 50 also supplies electrical power from the battery pack 26 to the agitator motor 54 of the cleaner head 4, through wires 56 running along the inside of the wand, so as to rotate the agitator 40. When the cleaner head 4 is on a hard floor, it is supported by the wheels 37 and the soleplate 34 and agitator 40 are spaced apart from the floor surface. When the cleaner head 4 is resting on a carpeted surface, the wheels 37 sink into the pile of the carpet and the soleplate 34 (along with the rest of the cleaner head 4) is therefore positioned further down. This allows carpet fibres to protrude towards (and potentially through) the suction opening 36, whereupon they are disturbed by bristles 43 of the rotating agitator 40 so as to loosen dirt and dust therefrom.

Vacuum cleaners 2 according to embodiments of the present disclosure comprise additional components, which are visible in Figures 3 and 4. These include one or more of: a current sensor 58 for sensing the electrical current drawn by the agitator motor 54 of the cleaner head 4, a pressure sensor 60 for sensing the pressure applied to the soleplate 34 of the cleaner head 4, an inertial measurement unit (IMU) 62 which is sensitive to motion and orientation of the main body 6 of the vacuum cleaner 2, a human computer interface (HCI) 64, one or more proximity sensors, typically in the form of time of flight (TOF) sensors 72, a tool switch sensor 74 and a capacitive sensor 76 located in the pistol grip 22. Although the current sensor 58 is shown as being situated in the cleaner head 4, it could alternatively be located in the main body 6. For example, the current sensor 58 could be integrated as part of the controller 50, provided it is operable to sense electrical current supplied to the agitator motor 54 from the battery 26 via the wires 56. In the illustrated embodiment, one TOF sensor 72 is located at the end of the detachable wand 8, close to where the cleaner head 4, or one of the other tools 3, 5, 7, is attached. Further TOF sensors 72 may be provided on the removable tools 3, 5, 7 themselves. Each TOF sensor 72 generates a sensor signal dependent on the proximity of objects to the TOF sensor 72. Suitable TOF sensors 72 include radar or laser devices. The tool switch sensor 74 is located on the main body 6 of the vacuum cleaner 2 and generates signals dependent on whether a tool 3, 4, 5, 7 or the wand 8 is attached to the main body 6. In embodiments, the tool switch sensor 74 generates signals dependent on the type of tool 3, 4, 5, 7 attached to main body 6 or the wand 8. The capacitive sensor 76 is located in the pistol grip 22 and generates signals dependent on whether a user is gripping the pistol grip. In embodiments, the vacuum cleaner 2 may comprise one or more additional IMUs. For example, the cleaner head 4 may comprise an IMU which is sensitive to motion and orientation of the cleaner head 4 and which generates further sensor signals to supplement those generated by the IMU 62 of the main body 6. The IMU 62 may comprise one or more accelerometers, one or more gyroscopes and/or one or more magnetometers.

As shown in more detail in Figure 4, the main body 6 of the vacuum cleaner 2 defines a longitudinal axis 70 which runs from a front end 9 to a rear end 11 of the main body 6. When it is attached to the front end 9 of the main body 6, the wand 8 is parallel to (and in this case collinear with) the longitudinal axis 70. In the illustrated embodiment, the HCI 64 comprises a visual display unit 65, more particularly a planar, full colour, backlit thin-film transistor (TFT) screen. The screen 65 is controlled by the controller 50 and receives power from the battery 26. The screen displays information to the user, such as an error message, an indication of a mode the vacuum cleaner 2 is operating in, or an indication of remaining battery 26 life. The screen 65 faces substantially rearwards (i.e. its plane is orientated substantially normal to the longitudinal axis 70). Positioned beneath the screen 65 (in the vertical direction defined by the pistol grip 22) is a pair of control members 66, also forming part of the HCI 64 and each of which is positioned adjacent to the screen 65 and is configured to receive a control input from the user. In embodiments, the control members are configured to change the mode of the vacuum cleaner, for example to manually increase or decrease the power of the vacuum motor 52. In embodiments, the HCI 64 also comprises an audio output device such as a speaker 67 which can provide audible feedback to the user.

The IMU 62 generates sensor signals dependent on the motion and orientation of the main body 6 of the vacuum cleaner 2 in three spatial dimensions (x, y, and z). The motion includes the Unear acceleration and angular acceleration of the main body 6. Figure 5a illustrates exemplary generated IMU 62 sensor data corresponding to the linear acceleration of the main body 6 before, during and after a cleaning operation. The time scale shows the index of samples which were gathered at a sampling rate of 25Hz. The vertical scale is in units of acceleration due to gravity. Traces 91a, 91b and 91c correspond to the linear acceleration of the main body 6 in the x, y and z directions respectively. Figure 5b illustrates exemplary generated IMU 62 sensor data corresponding to the angular acceleration of the main body 6 before, during and after the same cleaning operation as represented in Figure 5a. Traces 92a, 92b and 92c correspond to the angular acceleration about the x, y and z axes respectively. In both Figures 5a and 5b, the vacuum cleaner 2 is initially static (at rest). This is followed by a cleaning session comprising cleaning strokes, giving rise to oscillatory behaviour in some of the generated sensor data. Finally, the vacuum cleaner 2 is again retured to rest. The data shown in Figures 5a and 5b have been smoothed, for example by means of a band-pass filter or a low-pass filter. Figure 6 illustrates example generated IMU 62 sensor data corresponding to of the orientation of the main body 6 about the y axis during different hand-held cleaning operations. Specifically, interval 93a corresponds to cleaning of a low-level surface, e.g. a skirting board, interval 93b corresponds to a period dining which the main body 6 is at rest on a table and interval 93c corresponds to cleaning of an elevated surface, for example a ceiling, blind, curtain, or the top of a cupboard. Figure 7 illustrates further exemplary generated IMU 62 sensor data corresponding to orientation of the main body 6 about the y axis during different cleaning operations using the motorized cleaner heads 4, 5. Trace 94a corresponds to cleaning under furiture using the main cleaner head 4 attached to the wand 8. Trace 94b corresponds to stair cleaning using the miniature motorized cleaner head 5 attached directly to the main body 6, without using the wand 8. Trace 94c corresponds to normal upright vacuum cleaning using the cleaner head 4 attached to the wand 8. It should be appreciated that the different cleaning activities give rise to different signatures in the sensor data generated by the IMU 62. In this manner, it should be appreciated that the IMU 62 sensor data can be processed to infer information about the cleaning activity being performed by a user using the vacuum cleaner, or about the environment in which the vacuum cleaner is being operated.

Figure 8 illustrates schematically the electrical layout of the vacuum cleaner 2 according to embodiments. In embodiments, the controller 50 receives and processes signals generated by one or more of the trigger 24, the current sensor 58, the pressure sensor 60, the IMU 62, the one or more TOF sensors 72, the tool switch sensor 74 and the capacitive sensor 76. The controller 50 has a memory 51 on which are stored instructions according to which the controller 50 processes the sensor signals. Based on the processing of the sensor signals, the controller 50 controls one or more of the vacuum motor 52, the agitator motor 54 and the HCI 64 in order to enhance operation of the vacuum cleaner 2 and thereby improve the user experience. Example enhancements include improved pickup of dirt and improved battery life, amongst others.

Figure 9 is a block diagram which illustrates example sensor signal processing performed by the controller 50 according to various embodiments of the present disclosure. Unfiltered sensor signals 88 are received at the controller 50 from one or more of the available sensors. Different embodiments utilize sensor signals from different sensors. Some embodiments utilize sensor signals from only one sensor, such as the IMU 62, for example. A band-pass filter or low-pass filter 82 filters the raw sensor signals 88 to generate smoothed sensor signals 90 which are more suitable for further processing. At block 84, pre-determined features Fj, F2... F„ are extracted from the smoothed sensor signals and subsequently analysed by a classifier 86. In embodiments, the classifier 86 determines, from the extracted features, a particular cleaning activity being performed by a user using the vacuum cleaner 2. In other embodiments, the classifier 86 determines, from the extracted features a particular surface type on which the vacuum cleaner 2 is being operated. In other embodiments, the classifier 86 determines, from the extracted features, whether the vacuum cleaner 2 is actively being used, to assist in providing a trigger-less vacuum cleaner 2. Having determined the above, the controller 50 causes an action or actions to be performed involving one or more of the vacuum motor 52, agitator motor 54 and HCI 64, which are configured in dependence on the classifier 86 output, and optionally on the status of the trigger 24. It should be appreciated that the filter

82, feature extraction block 84 and classifier 86 are in general implemented as software modules which are executed on or under the control of the controller 50. The controller memory 51 stores sets of instructions defining the operation of the filter 82, feature extraction 84, classifier 86 and resultant action. In embodiments, the classifier is based on a machine learning classifier such as an artificial neural network, a random forest, a support-vector machine or any other appropriate trained model. The model could have been pre-trained, for example at the factory, using a supervised learning approach. A sliding window approach is generally used to span the filtered sensor signals and extract features corresponding to that particular time portion of the signal. Consecutive frames usually overlap to some degree but are usually processed separately. It should be appreciated that it is not always necessary to receive and process sensor data from all of the available sensors. For example, in embodiments the controller 50 may process only IMU 62 sensor data to obtain a classifier output. Furthermore, in the case of IMU 62 sensor data, the controller 50 may for example take account only of IMU 62 sensor data relating to orientation of the vacuum cleaner 2, or only IMU 62 sensor data relating to acceleration of the vacuum cleaner 2.

Figure 10 is a flow diagram showing a method 200 of facilitating the use of a vacuum cleaner 2 according to embodiments. In step 202, sensor signals based on sensed motion and orientation of the vacuum cleaner are generated, for example by the IMU 62. In step 204, the controller 50 processes generated sensor signals to determine a type of cleaning activity being performed by a user using the vacuum cleaner 2. In step 206, a recommendation is provided to the user of the vacuum cleaner 2 in dependence on the determined type of cleaning activity. The recommendation can take the form of a visual recommendation produced on the screen 65 and/or an audible recommendation produced on the speaker 67. The recommendation may be cancelled or acknowledged by the user by pressing one of the control members 66. In embodiments, the controller 50 processes die sensor signals in accordance with the example sensor signal processing described above with reference to Figure 9.

With reference to Figures 1 la and 1 lb, in embodiments, the recommendation comprises information relating to a recommended cleaning tool which is deemed suitable for the type of cleaning activity determined by the controller 50. For example, if the controller 50 determines that the user is manoeuvring the vacuum cleaner in a manner indicative of crevice cleaning, a visual representation of the crevice tool 3 is displayed on the screen 65 accompanied, optionally, by an audible recommendation 80 to use the crevice tool 3 produced on the speaker 67. In Figure 1 lb, the type of cleaning activity determined by the controller 50 corresponds to stair cleaning or upholstery cleaning. Accordingly, a visual representation of the miniature motorized tool 5 is displayed on tire screen 65 accompanied by an audible recommendation 80 to use the miniature motorized tool 5. In embodiments, the recommendation is only displayed when the user is not using the optimal tool for the cleaning activity currently being undertaken. For example, if the user is performing crevice cleaning using a dusting brush 7 rather than the crevice tool 3, the recommendation to use the crevice tool 3 is provided to the user. However, if the user is already using the crevice tool 3 then no recommendation need be provided.

In embodiments, the recommendation comprises information relating to a recommended cleaning technique, such as tire cleaning stroke rate, dwell time (the length of time for which the vacuum cleaner is held over a certain area), applied pressure, or a configuration of a cleaning tool (for example, the position of a suction gate on the cleaner head

4), each being suitable for the type of cleaning activity determined by the controller 50. When the vacuum cleaner 2 is being used with the cleaner head 4, the controller 50 may additionally process signals based on sensed parameters of the cleaner head 4 in determining the type of cleaning activity being performed. Example parameters include the agitator motor 54 current, sensed by current sensor 58, and the pressure on the cleaner head 4 sensed by the pressure sensor 60.

In embodiments of the present disclosure, the vacuum cleaner 2 comprises a controller 50. The controller 50 is configured to perform various methods described herein. In embodiments, the controller comprises a processing system. Such a processing system may comprise one or more processors and/or memory. Each device, component, or function as described in relation to any of the examples described herein, for example the IMU 62 and/or HCI 64 may similarly comprise a processor or may be comprised in apparatus comprising a processor. One or more aspects of the embodiments described herein comprise processes performed by apparatus. In some examples, the apparatus comprises one or more processors configured to carry out these processes. In this regard, embodiments may be implemented at least in part by computer software stored in (non-transitory) memory and executable by the processor, or by hardware, or by a combination of tangibly stored software and hardware (and tangibly stored firmware). Embodiments also extend to computer programs, particularly computer programs on or in a carrier, adapted for putting the above described embodiments into practice. The program may be in the form of non-transitory source code, object code, or in any other non-transitory form suitable for use in the implementation of processes according to embodiments. The carrier may be any entity or device capable of carrying the program, such as a RAM, a ROM, or an optical memory device, etc.

The one or more processors of processing systems may comprise a central processing unit (CPU). The one or more processors may comprise a graphics processing unit (GPU). The one or more processors may comprise one or more of a field programmable gate array (FPGA), a programmable logic device (PLD), or a complex programmable logic device (CPLD). The one or more processors may comprise an application specific integrated circuit (ASIC). It will be appreciated by the skilled person that many other types of device, in addition to the examples provided, may be used to provide the one or more processors. The one or more processors may comprise multiple co-located processors or multiple disparately located processors. Operations performed by the one or more processors may be carried out by one or more of hardware, firmware, and software. It will be appreciated that processing systems may comprise more, fewer and/or different components from those described.

The techniques described herein may be implemented in software or hardware, or may be implemented using a combination of software and hardware. They may include configuring an apparatus to cany out and/or support any or all of techniques described herein. Although at least some aspects of the examples described herein with reference to the drawings comprise computer processes performed in processing systems or processors, examples described herein also extend to computer programs, for example computer programs on or in a carrier, adapted for putting the examples into practice. The carrier may be any entity or device capable of carrying the program. The carrier may comprise a computer readable storage media. Examples of tangible computer-readable storage media include, but are not limited to, an optical medium (e.g., CD-ROM, DVD-ROM or Blu-ray), flash memory card, floppy or hard disk or any other medium capable of storing computer-readable instructions such as firmware or microcode in at least one ROM or RAM or Programmable ROM (PROM) chips.

Where in the foregoing description, integers or elements are mentioned which have known, obvious or foreseeable equivalents, then such equivalents are herein incorporated as if individually set forth. Reference should be made to the claims for determining the true scope of the present disclosure, which should be construed so as to encompass any such equivalents. It will also be appreciated by the reader that integers or filatures of the present disclosure that are described as preferable, advantageous, convenient or the like are optional and do not limit the scope of the independent claims. Moreover, it is to be understood that such optional integers or features, whilst of possible benefit in some embodiments of the present disclosure, may not be desirable, and may therefore be absent, in other embodiments.




 
Previous Patent: VACUUM CLEANER

Next Patent: VACUUM CLEANER