Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
LIP ACTIVITY DETECTION
Document Type and Number:
WIPO Patent Application WO/2012/127477
Kind Code:
A1
Abstract:
Provided is a method of detecting lip activity. The method determines magnitude of optical flow in lip region and at least one non-lip region of a detected face. The ratio of magnitude of optical flow in lip region and at least one non-lip region is compared against a threshold. If the ratio is found to be greater than the threshold, lip activity of the detected face is recognized.

Inventors:
DEY PRASENJIT (IN)
MADHVANATH SRIGANESH (IN)
Application Number:
PCT/IN2011/000183
Publication Date:
September 27, 2012
Filing Date:
March 21, 2011
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
HEWLETT PACKARD DEVELOPMENT CO (US)
DEY PRASENJIT (IN)
MADHVANATH SRIGANESH (IN)
International Classes:
G06K9/00
Foreign References:
US6754373B12004-06-22
CN101937268A2011-01-05
EP0863474A21998-09-09
Attorney, Agent or Firm:
NAMA, Prakash (198F 27th Cross,,3rd Block Jayanagar,1 Bangalore, Karnataka, IN)
Download PDF:
Claims:
Claims:

1. A computer-implemented method of detecting lip activity, comprising:

determining magnitude of optical flow in lip region of a detected face;

determining magnitude of optical flow in at least one non-lip region of the detected face;

comparing ratio of magnitude of optical flow in lip region of the detected face and magnitude of optical flow in at least one non-lip region of the detected face against a threshold; and

if the ratio is greater than the threshold, recognizing lip activity of the detected face.

2. A method according to claim 1, wherein the detected face is present in a video data stream or an audio-visual data stream.

3. A method according to claim 1, wherein if more than one non-lip region of the detected face is selected for determination of magnitude of optical flow in non-lip region, a mean of the magnitude of optical flow in selected non-lip regions is taken as magnitude of optical flow in the non-lip region.

4. A method according to claim 1, wherein the non-lip region includes a forehead region or cheek region of the detected face.

5. A computer-implemented method of detecting lip activity, comprising:

determining direction and magnitude of optical flow vectors in lip region of the detected face; and

if the direction of optical flow vectors in the lip region is both in upward and downward direction and comparable in magnitude, recognizing lip activity of the detected face.

6. A method according to claim 5, further comprising:

determining whether magnitude of optical flow along upward direction of lip region of a detected face is greater than a first threshold;

determining whether magnitude of optical flow along downward direction of lip region of the detected face is greater than a second threshold; and

if the magnitude of optical flow along upward direction and downward direction of lip region is greater than their respective thresholds, recognizing lip activity of the detected face.

7. A method according to claim 6, wherein the first threshold and the second threshold is same.

8. A system, comprising:

a memory containing a module; and

a processor to execute the module, the module comprising machine executable instructions that determine magnitude of optical flow in lip region of a detected face, determine magnitude of optical flow in at least one non-lip region of the detected face, compare ratio of magnitude of optical flow in lip region of the detected face and magnitude of optical flow in at least one non-lip region of the detected face against a threshold, and if the ratio is greater than the threshold, recognize lip activity of the detected face.

9. A system according to claim 9, further comprising a face detection module to detect a face in a captured image.

10. A system according to claim 9, wherein the detected face is present in a video data stream or an audio-visual data stream.

11. A system according to claim 8, further comprising an imaging device to capture an image.

12. A system according to claim 11, wherein the imaging device is removable or nonremovable.

13. A system according to claim 9, wherein the module further determines direction and magnitude of optical flow vectors in lip region of the detected face, and if the direction of optical flow vectors in the lip region is both in upward and downward direction and comparable in magnitude, recognizes lip activity of the detected face.

14. A computer program product adapted to cause a computer to perform the method of claim 1.

Description:
LIP ACTIVITY DETECTION

Background

Human-computer interaction has evolved considerably over the years. From good old command line based interfaces to present day multimodal interactive systems, developers are redefining the way people engage with computing devices. Intuitive interfaces are gaining importance. For example, touch based input is being increasingly preferred on handheld devices, such as, mobile phones, smart phones, PDAs (Personal Digital Assistant), touchpad, etc.

Speech is another modality that has been used either alone or in conjunction with other input modalities, such as, traditional keyboard, mouse input, gestures, gaze, etc. to interact with a computing device. Speech input may be used to convert spoken words into text, to provide commands or to control a computer application. For example, speech recognition has been used for voice dialing, performing a search, providing an interactive voice response, playing video games and controlling a robotic device.

Brief Description of the Drawings

For a better understanding of the solution, embodiments will now be described, purely by way of example, with reference to the accompanying drawings, in which:

FIG. 1 shows a flow chart of a computer-implemented method of detecting lip activity according to an embodiment. FIGS. 2A and 2B illustrate some aspects of the method of FIG. 1 according to an embodiment.

FIG. 3 shows a flow chart of a computer-implemented method of detecting lip activity according to an embodiment.

FIG. 4 shows a block diagram of a computing system according to an embodiment.

Detailed Description of the Invention

As mentioned earlier, speech may be used as an input to perform a variety of tasks on a computing device. However, one of the difficulties with recognition of speech as an input is the presence of spurious speech in the surrounding environment. For example, while providing a speech input to a computing system, there may be other people or users speaking in the background, or there may be ambient noise due to vehicular traffic, a procession, a festival event, a party, and the like. Under such circumstances, it becomes difficult for a computing system to identify speech of an actual user from other people's voices. This may lead to improper command execution and undesirable results.

Considering that a speech input may constitute an integral mode of a multimodal user interface, identification and recognition of "genuine" speech is important for the robustness of a multimodal interaction system as a whole. A weak speech recognition system may mar a user's experience with a multimodal user interface, even if other modalities (such as, touch, gesture, gaze, etc.) function effectively.

The proposed solution provides a mechanism to infer speech activity of a user of a computing device. It detects lip activity using optical flow to determine speech activity of a user. The solution enables rejection of spurious speech originating from other users in the surrounding. Detection of lip activity of a user along with audio activity from the same direction can be used to infer a "genuine" speech command. Absence of either of these is taken as an indication of an ambient speech activity or other lip movements apart from speech (such as, chewing of gum, a smile, etc.).

Embodiments of the present solution provide a method and system for detecting lip activity.

For the sake of clarity, the term "user", in this document, may include a "consumer", an "individual", a "person", or the like.

It is also clarified that the term "module", as used herein, may mean to include a software component, a hardware component or a combination thereof. A module may include, by way of example, components, such as software components, processes, functions, attributes, procedures, drivers, firmware, data, databases, and data structures. The module may reside on a volatile or non-volatile storage medium and configured to interact with a processor of a computing device (system).

FIG. 1 shows a flow chart of a computer-implemented method of detecting lip activity according to an embodiment.

The method may be implemented on a computing device (system), such as, but not limited to, a personal computer, a desktop computer, a laptop computer, a notebook computer, a network computer, a personal digital assistant (PDA), a mobile device, a hand-held device, or the like. A typical computing device that may be used is described further in detail subsequently with reference to FIG. 4. Additionally, the computing device may be connected to another computing device or a plurality of computing devices via a network, such as, but not limited to, a Local Area Network (LAN), a Wide Area Network, the Internet, or the like.

Referring to FIG. 1, block 110 involves face detection of a user of a computing device. In an example, the computing device may comprise an imaging device for capturing an image of a user. The imaging device may include a camera, which may capture an image of the user as a still image, a video data stream or an audio video (AV) data stream. Once captured, the image may be processed by a face detection module residing in the memory of the computing device for detecting the user's face. In an example, Viola-Jones face detection method may be used for face detection. However, other face detection methods, such as, but not by way of limitation, Schneiderman & Kanade's and Rowley, Baluja & Kanade's methods, may also be used. The face detection module may, amongst other features, recognize the size and location of the user's face in the captured image. It also detects the lip region of a user using face geometry, thereby distinguishing the lip region from a non-lip region(s).

In another example, multiple people may be using a computing device. In such case, the computing device detects faces of all users.

Also, the imaging device may be removably or non-removably attachable to the computing device.

Block 120 involves determination of magnitude of optical flow in at least one non-lip region of a detected face. Optical flow is the apparent motion of brightness patterns in an image. Generally, optical flow corresponds to the motion field. In an example, the magnitude of optical flow in at least one non-lip region of a face detected at block 110 is determined. In the present example, a 3 rd level Pyramidal Lucas-Kanade optical flow method may be used for determining magnitude of optical flow in at least one non-lip region of a detected face. However, optical flow detection methods, such as, but not by way of limitation, phase correlation method, block-phase methods, Horn- Schunk method, Buxton-Buxton method, Black-Jepson method, discrete optimization methods may also be used.

Also, in the present example, the cheek region of a detected face may be used for determination of magnitude of optical flow in a non-lip region. However, in other examples, different non-lip regions (such as, forehead) of the face may also be used.

Furthermore, in another example, more than one non-lip region of a detected face may be used for determination of magnitude of optical flow in the non-lip region. For example, both the cheek region and the forehead region may be used together for determination of magnitude of optical flow in the non-lip region of a detected face. In such case, a mean of the magnitude of optical flow in the cheek region and the magnitude of optical flow in the forehead region of a detected face is taken.

Block 130 involves determination of magnitude of optical flow in lip or mouth region of a detected face. In an example, the magnitude of optical flow in the lip region of a face detected at block 110 is determined. In the present example, a 3 rd level Pyramidal Lucas-Kanade optical flow method may be used for determining magnitude of optical flow in the lip region of a detected face. However, optical flow detection methods, such as, but not by way of limitation, phase correlation method, block-phase methods, Horn-Schunk method, Buxton-Buxton method, Black-Jepson method, discrete optimization methods may also be used.

Block 140 involves a comparison between a ratio of magnitude of optical flow in lip region of a detected face and magnitude of optical flow in at least one non-lip region of a detected face with a threshold. In an example, a ratio of the magnitude of optical flow computed in the lip region of a detected face (block 130) and the magnitude of optical flow computed for at least one non-lip region of a detected face (block 120) is determined. The ratio is then compared against a threshold which may be decided based on experimentation. To illustrate, a ratio of the magnitude of optical flow in the lip region and the magnitude of optical flow in the forehead region may be compared against a pre-determined threshold.

In another example, if more than one non-lip region (for example, cheek and forehead regions) of a detected face is selected for determination of magnitude of optical flow in the non-lip region, a mean of the magnitude of optical flow in selected non-lip regions is taken as magnitude of optical flow in the non-lip region. The mean is then used for the computation of the ratio between the magnitude of optical flow in lip region and selected non-lip regions of a detected face.

Upon comparison (block 150), if it is found, that the ratio of the magnitude of optical flow computed in the lip region of a detected face and the magnitude of optical flow computed for at least one non-lip region of a detected face is greater than the threshold, the lip activity of a detected face is recognized (block 160) i.e. it indicates lip movement or lip motion in a detected face. On the other hand, if the ratio is smaller than the threshold, it implies no lip activity or movement in a detected face (block 170).

In case there are multiple detected faces, the method distinguishes faces that are showing lip activity from the ones where no lip activity is indicated. For example, in a multi-user interaction one can determine which user is speaking and hence execute commands based on that user's context. If the identity of the user is recognized through face recognition, a command, for instance, "show my photos" from a user among multiple users can be resolved by correlating the identity and lip activity of the identified user.

FIGS. 2A and 2B illustrate some aspects of the method of FIG. 1 according to an embodiment. FIG. 2A illustrates face detection of a user of a computing device 200. In the present example, a computing device 200 has captured an image 210 of a user. The image 210 may be a still image, a video data stream or an audio video (AV) data stream. The image is processed by a face detection module residing on the computing device 200 for detecting the user's face 220.

FIG. 2B illustrates identification and subsequent determination of magnitude of optical flow in the lip region and at least one non-lip region of detected face 220. The non-lip regions of face used for determination of magnitude of optical flow include a left patch 230 below the left eye and a right patch 240 below the right eye. For determination of magnitude of optical flow in the non-lip region of detected face 220, a mean of magnitude of optical flow in the left patch 220 and magnitude of optical flow in the right patch 240 may be taken. Lip region 250 may be used for determining amount of optical flow in the lip region of detected face 220.

FIG. 3 shows a flow chart of a computer-implemented method of detecting lip activity according to an embodiment.

FIG. 3 illustrates further optional or additional steps that may be carried for detecting lip activity of a user or multiple users of a computing device.

Block 310 involves determination of direction of optical flow vectors in lip region of a detected face (such as the one detected at block 110). Optical flow arises from relative motion of an object and a viewer. In an example, while determining magnitude of optical flow in the lip region of a detected face, the direction of optical flow vectors along X and Y axis (at the lip region) is determined. If it is found that the direction of optical flow vectors in the lip region is both in upward and downward direction (i.e. only along the Y axis) and of comparable magnitude, the lip activity of a detected face is recognized (block 330). In other words, bidirectional movement of optical flow vectors of comparable magnitude along Y-axis in the lip region indicates lip activity in a detected face. On the other hand if it is found that that the direction of optical flow vectors in the lip region is either in upward or downward direction (unidirectional) only, or along the X-axis, it indicates absence of lip activity in the detected face (block 340).

Determination of direction of optical flow vectors removes spurious lip activity detections due to head motion in the vertical direction or due to forward backward motion. In both cases, optical flow along opposite directions and of comparable magnitude is not observed. For example, in case of head movement in the vertical direction, there may be optical flow vectors present along the positive vertical (Y) axis, but none may be observed down the negative vertical (Y) axis.

Block 320 involves determination of magnitude of optical flow along upward and downward direction of the lip region of a detected face. The magnitude of optical flow along positive vertical (Y) axis and negative vertical (Y) axis is determined. It is observed that the vertical component of optical flow along both positive and negative vertical axis is large in magnitude and almost equal in proportion.

In an example, the magnitude of optical flow along upward direction of the lip region of a detected face is compared against a first threshold (previously determined). Similarly, the magnitude of optical flow along downward direction of the lip region is compared against a second threshold (again, previously determined). If upon comparison it is found that the magnitude of optical flow along upward direction of the lip region of a detected face is greater than the first threshold and the magnitude of optical flow along downward direction of lip region is greater than the second threshold, and the magnitude of the flows in the opposite directions are comparable, lip activity of a detected face is recognized. If the magnitude of optical flow along upward and downward direction of the lip region is greater than their respective thresholds, it indicates that there is lip movement in the detected face (block 330). The presence of a threshold on the individual optical flow magnitudes in upward and downward direction ensures that spurious lip movements do not result in detections even though they may be of comparable magnitude. Conversely, an upward and downward optical flow above the threshold may not indicate genuine lip activity unless they are bi-directional and of comparable magnitude.

In an example, the first and second threshold may be of different value. In another example, however, the first and second threshold may be of same value.

FIG. 4 shows a block diagram of a computing system according to an embodiment.

The system 400 may be a computing device, such as, but not limited to, a personal computer, a desktop computer, a laptop computer, a notebook computer, a network computer, a personal digital assistant (PDA), a mobile device, a hand-held device, or the like.

System 400 may include a processor 410, for executing machine readable instructions, a memory 412, for storing machine readable instructions (such as, a module 414), and an imaging device 416. These components may be coupled together through a system bus 418.

Processor 410 is arranged to execute machine readable instructions. The machine readable instructions may comprise a module that determines magnitude of optical flow in lip region of a detected face, determines magnitude of optical flow in at least one non-lip region of the detected face, compares ratio of magnitude of optical flow in lip region of the detected face and magnitude of optical flow in at least one non- lip region of the detected face against a threshold, and if the ratio is greater than the threshold, recognizes lip activity of the detected face. Processor 410 may also execute modules related to face detection of a user or multiple users. The memory 412 may include computer system memory such as, but not limited to, SDRAM (Synchronous DRAM), DDR (Double Data Rate SDRAM), Rambus DRAM (RDRAM), Rambus RAM, etc. or storage memory media, such as, a floppy disk, a hard disk, a CD-ROM, a DVD, a pen drive, etc. The memory 412 may include a module 414. In an example, the module 414 may be a face detection module that includes machine executable instructions for detecting a user's face in an image. The image may be a still photograph, a video data stream or an audio-video data stream.

Imaging device 416 is used to capture an image or a series of images of a user or multiple users. The imaging device may be a separate device, which may be attachable to the computing system 400, or it may be integrated with the computing system 400. In an example, the imaging device 416 may be a camera, which may be a still camera, a video camera, a digital camera, and the like.

It would be appreciated that the system components depicted in FIG. 4 are for the purpose of illustration only and the actual components may vary depending on the computing system and architecture deployed for implementation of the present solution. The various components described above may be hosted on a single computing system or multiple computer systems, including servers, connected together through suitable means.

The examples described detect the magnitude (amount) of optical flow in the lip (mouth) region to detect lip activity. The method requires no training and lip detection may take place as soon as a user comes in the range of vision of an imaging device. The proposed solution also works well in less illuminated environments. It does not require any history to be maintained and optical flow is calculated over a very small part of a detected face which makes the mechanism less calculation intensive.

It will be appreciated that the embodiments within the scope of the present solution may be implemented in the form of a computer program product including computer-executable instructions, such as program code, which may be run on any suitable computing environment in conjunction with a suitable operating system, such as Microsoft Windows, Linux or UNIX operating system. Embodiments within the scope of the present solution may also include program products comprising computer- readable media for carrying or having computer-executable instructions or data structures stored thereon. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer. By way of example, such computer-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM, magnetic disk storage or other storage devices, or any other medium which can be used to carry or store desired program code in the form of computer-executable instructions and which can be accessed by a general purpose or special purpose computer.

It should be noted that the above-described embodiment of the present solution is for the purpose of illustration only. Although the solution has been described in conjunction with a specific embodiment thereof, those skilled in the art will appreciate that numerous modifications are possible without materially departing from the teachings and advantages of the subject matter described herein. Other substitutions, modifications and changes may be made without departing from the spirit of the present solution.