Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
METHOD AND SYSTEM FOR MATCHING 2D HUMAN POSES FROM MULTIPLE VIEWS
Document Type and Number:
WIPO Patent Application WO/2021/186225
Kind Code:
A1
Abstract:
This disclosure is directed to a method and system for matching human pose data in the form of 2D skeletons for the purposes of 3D reconstruction. The system may comprise a scoring module that assigns an affinity score to each pair of cross-view 2D skeletons, a matching module that assigns optimal pairwise matches based on the affinity scores, a grouping module that assigns each 2D skeleton to a group such that each group corresponds to a unique person, based on the pairwise matches; and a temporal consistency module that assigns each group an ID that maintains correspondence to the same person over the multi-video sequence.

Inventors:
ZHANG WENXIN (CA)
Application Number:
PCT/IB2020/052609
Publication Date:
September 23, 2021
Filing Date:
March 20, 2020
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
WRNCH INC (CA)
International Classes:
G06T7/292; G06K9/78; G06T7/246
Foreign References:
GB2573170A2019-10-30
Other References:
SHIH-MING CHANG ; YI-SHENG TSAI ; HUI-HUANG HSU ; KUAN-CHING LI: "3D Skeleton Construction by Multi-view 2D Images and 3D Model Segmentation", UBI-MEDIA COMPUTING (U-MEDIA), 2011 4TH INTERNATIONAL CONFERENCE ON, IEEE, 3 July 2011 (2011-07-03), pages 168 - 173, XP032037350, ISBN: 978-1-4577-1174-9, DOI: 10.1109/U-MEDIA.2011.48
WANG ZIREN; LIU GUOLIANG; TIAN GUOHUI: "Human skeleton tracking using information weighted consensus filter in distributed camera networks", 2017 CHINESE AUTOMATION CONGRESS (CAC), IEEE, 20 October 2017 (2017-10-20), pages 4640 - 4644, XP033290186, DOI: 10.1109/CAC.2017.8243598
See also references of EP 4121940A4
Attorney, Agent or Firm:
DLA PIPER (CANADA) LLP et al. (CA)
Download PDF:
Claims:
CLAIMS:

1. A method of identifying humans between two or more camera views from 2d skeletons of the humans of each view comprising: a) for each skeleton in each of the two or more camera views, performing a pairwise scoring with each of the skeletons in another of the two or more camera views and assign an affinity score to each pair; b) identifying a best match of a skeleton in a first camera view to a skeleton in a second camera view by maximizing the affinity score of the pair; and c) grouping skeletons by identifying a set of skeletons in a first camera view, the set relating to the humans in the first camera view, with a set of skeletons in a second camera view using the best match.

2. The method of identifying humans between two or more camera views of claim 1 further comprising assigning an identifier to each skeleton in the grouped skeletons in a frame of the camera views, and assigning the same identifier to each skeleton in the grouped skeletons a subsequent frame of the camera views that match.

3. The method of identifying humans between two or more camera views of claim 1 or 2 wherein the pairwise scoring of a pair of skeletons from a pair of camera views comprises modelling a ray from each camera view to an element of the 2d skeleton associated with the camera view and determining the minimum distance between the two rays.

4. The method of claim 3 wherein if the rays are divergent, the pair is not included in the affinity score.

5. The method of claim 3 wherein the pairwise scoring of a pair of skeletons from a pair of camera views further comprises excluding elements where the minimum distance between the two rays exceeds a threshold.

6. The method of claim 3 wherein the pairwise scoring of a pair of skeletons from a pair of camera views further comprises determining a deviation of attributes of a putative 3d skeleton formed from the 2d skeletons from a typical human.

7. The method of any one of claims 1 to 6 further comprising calibrating each camera view by determining the position and angle of the camera, and synchronizing the camera view by aligning frames taken at the same time from the one or more camera views.

8. The method of any one of claims 1 to 7 wherein identifying a best match of a skeleton in a first camera view to a skeleton in a second camera view includes not identify any match.

9. A motion capture system for two or more humans comprising: two or more calibrated cameras generating synchronized video streams, each camera having an overlapping field of views that include the two or more humans; a 2D pose estimator module associated with each of the two or more calibrated cameras for generating a 2D skeletons for each human in the field of view of the camera for a frame of the video stream; a scoring module for perform a pairwise scoring for each of the 2D skeletons associated with a first camera with each 2D skeleton of another of the two or more cameras and assigning an affinity score to each pair; a matching module that matches a 2D skeleton in a first camera view to a 2D skeleton in a second camera view by maximizing the affinity score of the pair; a grouping module that groups 2D skeletons by identifying a set of 2D skeletons in a first camera view, the set relating to the humans in the first camera view, with a set of 2D skeletons in a second camera view using the best match; a temporal matching module that assigns an identifier to each 2D skeleton group that remains consistent across a sequence of frames of the video streams; and a 3D reconstruction module that combines the grouped 2D skeleton across a sequence of frames for a human to create a 3D skeleton of the human, capturing the position of the human.

10. The system of claim 9 wherein the scoring module comprises a model of a ray from each camera view to an element of the 2d skeleton associated with the camera view and determining the minimum distance between the two rays.

Description:
METHOD AND SYSTEM FOR MATCHING 2D HUMAN POSES FROM

MULTIPLE VIEWS

FIELD

[0001] This disclosure relates to identifying and tracking 2D joint skeletons in video segments. More particularly, this disclosure relates to matching 2D skeletal data corresponding to the same person where the 2D data is extracted from frames of video segments taken from multiple viewpoints.

BACKGROUND

[0002] Reconstruction of 3D human poses from synchronized 2D video sequences may be accomplished in two stages. The first stage, 2D human pose estimation, detects keypoints in each frame of each video sequence. The second stage fuses the 2D keypoints, along with the camera calibration parameters, into 3D skeletons.

[0003] 2D human pose estimators may rely on deep neural networks to detect keypoints, which may correspond to anatomical joints, in each video frame of a video sequence. A group of keypoints belonging to a single person may be connected to form a 2D skeleton. For scenes containing multiple persons, multiple 2D skeletons may be detected in each frame, and each is assigned an index or unique ID. Multi-person pose estimation may be accomplished by performing keypoint detection on multiple regions of interest, or it may be accomplished by detecting all keypoints in a single image frame jointly in “one shot” and then grouping them into individual 2D skeletons.

[0004] For each person in the scene, 2D skeletons that correspond to the specific person are grouped together and the 3D skeleton is estimated through a data fusion technique. For instance, each 3D joint position may be independently estimated by triangulation of 2 or more keypoints. Alternatively, 3D joint positions may be estimated by Kalman Filters that model the motion of the joints over time.

[0005] For scenes containing multiple persons, it may be important that 2D skeletons be grouped such that each group corresponds to a single person. Because the 2D skeletons in each view may be extracted independently, their indices or IDs are not correlated across views. Accordingly, a matching step is typically used to identify the 2D groups that get fused in order to recover the 3D skeletons.

SUMMARY

[0006] This disclosure relates in an aspect to a method of identifying humans between two or more camera views from 2d skeletons of the humans of each view. The method includes for each skeleton in each of the two or more camera views, performing a pairwise scoring with each of the skeletons in another of the two or more camera views and assign an affinity score to each pair. The method also includes identifying a best match of a skeleton in a first camera view to a skeleton in a second camera view by maximizing the affinity score of the pair. The method includes grouping skeletons by identifying a set of skeletons in a first camera view, the set relating to the humans in the first camera view, with a set of skeletons in a second camera view using the best match.

[0007] In an aspect, this disclosure relates to a motion capture system for two or more humans comprising two or more calibrated cameras generating synchronized video streams, each camera having an overlapping field of views that include the two or more humans. The system has a 2D pose estimator module associated with each of the two or more calibrated cameras for generating a 2D skeletons for each human in the field of view of the camera for a frame of the video stream and a scoring module for perform a pairwise scoring for each of the 2D skeletons associated with a first camera with each 2D skeleton of another of the two or more cameras and assigning an affinity score to each pair. The system also has a matching module that matches a 2D skeleton in a first camera view to a 2D skeleton in a second camera view by maximizing the affinity score of the pair and a grouping module that groups 2D skeletons by identifying a set of 2D skeletons for each person, respectively, in the captured scene such that each 2D skeleton in a group corresponds to a view of the respective person in a given camera view. The system also includes a temporal matching module that assigns an identifier to each 2D skeleton group that remains consistent across a sequence of frames of the video streams and a 3D reconstruction module that combines the grouped 2D skeleton across a sequence of frames for a human to create a 3D skeleton of the human, capturing the position of the human.

BRIEF DESCRIPTION OF THE DRAWINGS

[0008] In drawings which illustrate by way of example only an embodiment of the disclosure,

[0009] Figure 1 is an exemplary pictorial representation of 2D skeleton data derived from three video sequences, in accordance with an embodiment.

[0010] Figure 2 is a block diagram of a system for matching 2D human poses, in accordance with an embodiment.

[0011] Figure 3 is an exemplary table of affinity scores for a pair of views, and the matching pairs produced by the pairwise matching module, in accordance with an embodiment.

[0012] Figure 4 is an exemplary graph of pairwise matches, and the connected components or cycles that represent groups that each correspond to a unique person.

DETAIFED DESCRIPTION

[0013] This disclosure is directed to a method and system for matching human pose data in the form of 2D skeletons for the purposes of 3D reconstruction. The system may comprise a scoring module 20 that assigns an affinity score to each pair of cross-view 2D skeletons, a matching module 30 that assigns optimal pairwise matches based on the affinity scores, a grouping module 50 that assigns each 2D skeleton to a group such that each group corresponds to a unique person, based on the pairwise matches; and a temporal consistency module 60 that assigns each group an ID that maintains correspondence to the same person over the multi-video sequence. [0014] With reference to Figure 1, 2D skeleton data 10 is extracted from two or more video sequences, taken from calibrated cameras. To perform 3D reconstruction, the 2D skeletons may be matched across views. A calibrated camera is preferably a camera for which field of view, angle and location information is known. The two or more video sequences are preferably synchronized so that each of the video sequences include the same period of time and include at least some of the same humans/skeletons. In some instances, one or more humans/skeletons may leave the field of view of one or more of the cameras.

[0015] A 2D human pose estimator may generate 2D skeletons for each human in each of the two or more video sequences. This may be done using known techniques, such as using a convolutional neural network (CNN), including such as by Wrnch. AI. A sequence of 2D skeletons may be provided corresponding to the video sequences for each camera.

[0016] With reference to Figure 2, the 2D matching system may comprise the following modules: the pairwise scoring module 20, the pairwise matching module 30, the grouping module 40, and the temporal consistency module 50. The pairwise-score module 20 may assign an affinity score to each possible combination of cross-view pairs of 2D skeletons. A cross-view pair of 2D skeleton is any pair of skeletons where one skeleton is from a first video sequence and the second is from a second video sequence. The affinity score of a given pair of 2D skeletons correlates to the likelihood that the pair belong to the same person. In a preferred embodiment, the affinity score may be a weighted sum of several metrics based on the concept of “approximate triangulations” of cross-view keypoint pairs, as described below.

[0017] An approximate triangulation is computed by projecting a ray through each of the two keypoints. A keypoint of a 2D skeleton may be one particular element such as the centre of the head, centre of the pelvis, right or left wrist. Assuming a pinhole camera model, each ray is modelled as originating at the respective camera’s optical center, based on the parameters known of the camera such as its location, angle and field of view, and proceeding in the direction that passes through the keypoint on the virtual image plane. This is done for the same keypoint, for example the centre of the head, for the two skeletons being compared, one arising from a first camera and video sequence and one arising from the second camera and video sequence. The triangulation point is the point in 3 -space with a minimum Euclidean distance between the two rays. The triangulation error may be the minimum distance between the two rays. If the triangulation point is determined to be behind the cameras, the rays are diverging and this point may not be considered in the score calculations. In some embodiments, this may be done for more than one keypoint pairs.

[0018] One affinity score metric may be the total count of “inlier” keypoint pairs for the set of approximate triangulations for the given pair of 2D skeletons, where in inlier pair may be defined as a keypoint pair with a triangulation error below a certain threshold. For instance, a pair of 2D skeletons {A, B} may have a total of 7 inlier pairs out of a possible 8 (the pair corresponding to the left wrist joint is not considered an inlier because of high triangulation error), and another pair of skeletons {A, C} may have a total of 6 inlier pairs out of a possible 8 (the pairs corresponding to the right ankle and head joints respectively are not considered inliers). In this instance, {A, B} may score higher on the inlier metric of the weighted affinity score than {A, C}. Another metric may be the average triangulation error of all the pairs of keypoints belongs to the two skeletons. Another metric may be the “human-ness” of a putative 3D skeleton reconstruction consisting of all inlier triangulation points. The human-ness metric may be inversely proportional to the deviation of the limb lengths of the putative skeleton from those of an average person, based on anthropometric data. For instance, a putative 3D skeleton derived from a mismatched pair of 2D skeletons may have limbs that may be double the length of an average person, and thus may have a lower human-ness metric than a pair of correctly matched skeletons.

[0019] With reference to Figure 3, the pairwise-matching module 30 may examine in turn all the cross-view affinity scores 60. In other words, the affinity score for each pair of skeletons, a first skeleton from a first camera and a second skeleton from a second camera. The module may find a set of one-to-one matches between the 2D skeletons in the two views that maximizes the affinity score 70. This may be solved by using an assignment method such as the Hungarian algorithm, the primal simplex algorithm, or the auction algorithm. To handle the case where no matches are made (for instance, when the two views capture disjoint sets of persons), an embodiment may suppress matches whose affinity scores fall below a threshold. This process may be repeated for all pairs of camera views.

[0020] The grouping module 50 may take the set of pairwise matches and outputs N sets of 2D skeletons, where N is the number of distinct people in the scene and each set corresponds to a distinct person in the scene. With reference to Figure 4, the procedure for this grouping may be as follows. An undirected graph 80 may be first constructed where each 2D skeleton is associated with a vertex, and each pairwise match is an edge. Next, the graph is partitioned into subgraphs 90 such that each subgraph’s vertices comprise 2D skeletons that belong to the same person. The subgraphs may be connected components or biconnected components, and these subgraphs may be extracted using a standard depth-first search method.

[0021] The temporal matching module 60 may assign an ID to each 2D skeleton group, such that each person’s ID remains consistent over the video sequences. An embodiment may achieve this by reprojecting the 3D skeletons from a previous timestep according to the camera parameters to create a set of predicted 2D skeletons in a current timestep. The pixel distance to each 2D skeleton group from the 2D skeleton projections of the previous timestep may be computed, and a matching method such as Hungarian algorithm is used to generate a one-to-one correspondence between the set of extant 3D skeletons and the 2D skeleton groups such that the pixel distances are minimized. The 2D groups may then be assigned IDs that correspond to the indices of the extant 3D skeletons. This may be continued for each timestep of the video sequence.

[0022] The system modules described may be separate software modules, separate hardware modules, or portions or one or more hardware components. The functionality of the modules described above may be implemented in a single system or provided in separate modules similar to or different from the modules described. [0023] The software modules may consist of instructions written in a computer language such as C++ or assembly code and run on computer hardware such as a CPU, or they may be implemented on an FPGA. The software may utilize storage, such as RAM or magnetic storage, such as one or more hard drives. The system may run on a desktop computer, mobile phone or another platform that includes suitable memory for holding the software, data and skeletons parameters.

[0024] In an embodiment, the human matching system may comprise part of a motion capture system which digitizes the 3D poses of two or more humans subjects, such as in real time or post processing. This digitized pose data may be used for such applications as performance capture for digital media, or for sport analytics. Two or more calibrated cameras may be synchronized and their video streams captured and processed by 2D pose estimator systems, such as one for each video stream. The matching system may receive the output 2D skeletons from the 2D pose estimators, such as through a network interface or computer bus. The matched 2D skeleton groups may then be provided to a 3D reconstruction module, which fuses the 2D keypoints for each person in the scene to obtain the 3D pose data for each skeleton.

[0025] Various embodiments of the present disclosure having been thus described in detail by way of example, it will be apparent to those skilled in the art that variations and modifications may be made without departing from the disclosure. The disclosure includes all such variations and modifications as fall within the scope of the appended claims.