Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
OBJECT DETECTION AND IMAGE SEGMENTATION FOR EUKARYOTIC CELLS
Document Type and Number:
WIPO Patent Application WO/2024/020085
Kind Code:
A1
Abstract:
A method of generating artificial images for medical evaluation of eukaryotic cells can include extracting each instance of single-cells in an image of eukaryotic cells; extracting each instance of multi-cells in the image of eukaryotic cells; generating a background image from the image of eukaryotic cells; selecting a set of cells from the extracted single-cells and the extracted multi-cells; applying at least one augmentation technique to each cell in the set of cells to generate augmented cells; and generating an artificial image of eukaryotic cells using the augmented cells and the background image.

Inventors:
LIU DONGFANG (US)
NAGHIZADEH ALIREZA (US)
METAXAS DIMITRIS (US)
Application Number:
PCT/US2023/028125
Publication Date:
January 25, 2024
Filing Date:
July 19, 2023
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
UNIV RUTGERS (US)
International Classes:
G06T7/00; G06V10/25; G06V20/69; G06V10/00
Foreign References:
US20010041347A12001-11-15
US20070038384A12007-02-15
US20210325308A12021-10-21
Attorney, Agent or Firm:
KNIGHT, Sarah J. (US)
Download PDF:
Claims:
CLAIMS

What is claimed is:

1. A method of generating artificial images for medical evaluation of eukaryotic cells, comprising: extracting each instance of single-cells in an image of eukaryotic cells; extracting each instance of multi-cells in the image of eukaryotic cells; generating a background image from the image of eukaryotic cells; selecting a set of cells from the extracted single-cells and the extracted multi-cells; applying at least one augmentation technique to each cell in the set of cells to generate augmented cells; and generating an artificial image of eukaryotic cells using the augmented cells and the background image.

2. The method of claim 1, wherein generating the background image from the image of eukaryotic cells comprises: generating a bounding box around at least one instance of a eukaryotic cell in the image of eukaryotic cells, wherein pixels of the image of eukaryotic cells inside the bounding box are inside pixels; and replacing values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box to remove the at least one instance of the eukaryotic cell in the image of eukaryotic cells, wherein the certain pixels found outside of the bounding box are outside pixels.

3. The method of claim 2, wherein generating the background image from the image of eukaryotic cells further comprises: performing the generating of the bounding box around the at least one instance of the eukaryotic cell in the image of eukaryotic cells and the replacing of the values of the inside pixels until all instances of eukaryotic cells are removed.

4. The method of claim 2, wherein replacing the values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box comprises: for each inside pixel: selecting a predetermined number of outside pixels, comparing each of the predetermined number of outside pixels to pixels at comers of the bounding box to identify an outside pixel of the predetermined number of outside pixels that is most similar to the pixels at the comers of the bounding box, and replacing that inside pixel with the identified outside pixel.

5. The method of claim 4, wherein the predetermined number of outside pixels are selected randomly.

6. The method of claim 4, wherein comparing each of the predetermined number of outside pixels to pixels at comers of the bounding box to identify the outside pixel of the predetermined number of outside pixels that is most similar to the pixels at the comers of the bounding box comprises evaluating similarity using a mean Euclidean distance.

7. The method of claim 2, wherein generating the background image from the image of eukaryotic cells further comprises: creating a masking image, wherein the masking image is an image with a black-colored background having a smaller, white area in a middle, wherein the masking image is a same size as the bounding box; applying a Gaussian filter to the masking image to generate a filtered masking image; inserting the filtered masking image in a same position as the bounding box; and smoothing an area of the image of eukaryotic cells where the at least one instance of eukaryotic cells was removed using the filtered masking image.

8. The method of claim 1, wherein generating the artificial image of eukaryotic cells using the augmented cells and the background image comprises: placing a selection of the augmented cells on the background image.

9. The method of claim 8, wherein the selection of the augmented cells mimics a distribution of the eukaryotic cells in the image of eukaryotic cells.

10. The method of claim 8, wherein the selected augmented cells are placed randomly on the background image.

11. The method of claim 8, wherein placing the selection of the augmented cells on the background image comprises: generating a bounding box around at least one cell from the selection of the augmented cells; comparing pixels from four comers of the bounding box to pixels in the background image to determine a target location, wherein the target location is an area of the background image where the pixels in the background image have a highest degree of similarity to the pixels in the four comers of the bounding box; placing the at least one of the selection of the augmented single-cells and the augmented multi-cells in the target location; creating a cell masking image, wherein the cell masking image includes a black color background and a white area in the middle, wherein the white area is substantially the same shape as the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; and applying a Gaussian filter to smooth a transition between texture of the background image and texture of the at least one cell from the selection of the augmented single-cells and the augmented multi-cells.

12. The method of claim 11, wherein comparing pixels from four comers of the bounding box to pixels in the background image to determine the target location comprises using a mean Euclidean distance for determining similarity.

13. The method of claim 11, wherein the bounding box further comprises a height.

14. The method of claim 1, wherein the at least one augmentation technique applied to each cell corresponds to a particular augmentation policy of a pre-determined selection of a set of augmentation techniques.

15. The method of claim 14, wherein augmentation techniques of the pre- determined selection of the set of augmentation techniques are selected from FlipLR, FlipUD, AutoContrast, Equalize, Rotate, Posterize, Contrast, Brightness, Sharpness, Smooth, and Resize.

Description:
OBJECT DETECTION AND IMAGE SEGMENTATION FOR EUKARYOTIC CELLS

BACKGROUND

[0001] There exist many powerful architectures for object detection and instance segmentation of both biomedical and natural images. However, a difficulty arises in the ability to create training datasets that are large and well -varied. The importance of this subject is nested in the amount of training data that artificial neural networks need to accurately identify and segment objects in images and the infeasibility of acquiring a sufficient dataset within the biomedical field.

[0002] In particular, the ability to image, extract and study cells is essential to various research areas within the medical field. Advancements in high-resolution fluorescent microscopy have allowed medical professionals access to a more detailed visualization of cells and their interactions. A prime example is within immunotherapy, where there exists great importance to assess the efficacy of different treatments for fatal illnesses such as cancer, or more contemporarily, HIV/AIDS. Automated analysis of cellular images enables medical researchers to avoid time-consuming quantification, vastly improving the speed at which it takes to quantify reproducible mass data. The development of strong artificial neural networks for instance segmentation and object detection should naturally accompany the automation process.

[0003] However, the large datasets required to train models of Artificial Neural Networks (ANNs) greatly slows down researchers due to the naturally lengthy process of acquiring ground truth images. In the translational and basic biomedical research, the well- annotated datasets are critical for machine learning algorithms. Particularly, the detection and segmentation of eukaryotic cells used for intensity quantification of these cells require large manually annotated microscopical datasets for artificial neural networks (ANNs). Creating training data from microscopical images requires expertise and precision to separate nuclei cells.

[0004] Unfortunately, current methods in machine learning which deal with artificially augmenting training data such as different forms of Generative Adversarial Networks (GANs) do not deal with generating instance segmentation masks for the objects. They also have other problems that make them unsuitable for generating training data in medical fields. To properly train GAN networks, there is usually need for a large set of training data, which is not always available in microscopical nuclei imaging. [0005] Another issue that current models of data augmentation may have with nuclei cell images originates from the nature of this type of data. Given that medical imaging deals with highly sensitive data, having uncontrolled, noisy, and unreliable artificial images can potentially lead to disastrous outcomes. For instance, in natural images, creating unidentified objects that look real is always exciting. However, for immunotherapists that are treating cancer patients, using such training sets may result in not correctly evaluating an immune response within a cancer patient. For example, physicians use the images of chimeric antigen receptor (CAR) immunological synapse (IS) quality using ANNs to predict the efficacy of certain CAR- T cells in clinics. The existence of unidentified, synthesized CAR IS objects in the training images can result in incorrectly evaluating an immune response for a cancer patient.

[0006] Thus, there is a need for systems and methods to generate artificially augmented images of eukaryotic cells in a manner that is also generates accurate masks of the objects within the images.

BRIEF SUMMARY

[0007] Improved object detection and image segmentation for eukaryotic cells is described, which is suitable for generating training data, particularly of eukaryotic cells, for medical-related machine learning systems.

[0008] A method of generating artificial images for medical evaluation of eukaryotic cells can include extracting each instance of single-cells in an image of eukaryotic cells; extracting each instance of multi-cells in the image of eukaryotic cells; selecting a set of cells from the extracted single-cells and the extracted multi-cells; applying at least one augmentation technique to each cell in the set of cells to generate augmented cells; and generating an artificial image of eukaryotic cells using the augmented cells. The method can further include generating a background image from the image of eukaryotic cells, the artificial image being generated using the augmented cells and the background image.

[0009] This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.

BRIEF DESCRIPTION OF THE DRAWINGS

[0010] Figure 1 illustrates a method of image creation for generating training data for image segmentation of eukaryotic cells. [0011] Figure 2 illustrates a process pipeline of a method of image creation.

[0012] Figure 3 illustrates an autoencoder pipeline.

[0013] Figures 4A-4C illustrate a process pipeline of a method of image creation including creation of background images for the artificial images.

[0014] Figures 5A-5F show removal of cells from a real image to create an artificially generated background.

[0015] Figures 6A-6F show placement of cells on an artificially generated background.

[0016] Figures 7A-7C show single-cell image generation of CAR-T and Kaggle datasets.

[0017] Figures 8A-8C show multi-cells image generation on CAR-T and Kaggle datasets.

[0018] Figures 9A-9G show cell nuclei image generation on CAR-T and Kaggle datasets.

[0019] Figures 10A-10C show cell nuclei image generation on CAR-T, Kaggle, and Neural datasets.

[0020] Figure 11 shows a sample of an artificially generated Neural dataset (first row) with its respective instance segmentation (second row).

[0021] Figure 12 shows a real sample from the Neural dataset (first row) with its respective instance segmentation (second row).

[0022] Figure 13 shows a sample of an artificially generated CAR-T dataset (first row) with its respective instance segmentation (second row).

[0023] Figure 14 shows a real sample from the CAR-T dataset (first row) with its respective instance segmentation (second row).

[0024] Figure 15 shows a sample of an artificially generated Kaggle dataset (first row) with its respective instance segmentation (second row).

[0025] Figure 16 shows a real sample from the Kaggle dataset (first row) with its respective instance segmentation (second row).

[0026] Figure 17 shows single-cell real images on CAR-T, Neural, and Kaggle datasets.

[0027] Figure 18 shows single-cell image generation on CAR-T, Neural, and Kaggle datasets.

[0028] Figure 19 shows multi-cell real images on CAR-T, Neural and Kaggle datasets. [0029] Figure 20 shows multi-cell image generation on CAR-T, Neural and Kaggle datasets.

[0030] Figure 21 shows a comparison between complete image generation of Neural dataset.

[0031] Figure 22 shows images of the original CAR-T dataset and the original Kaggle dataset.

[0032] Figure 23 shows a sample using different forms of augmentation techniques on the cells.

[0033] Figure 24 shows a sample using different forms of augmentation techniques on the cells.

[0034] Figure 25 shows a sample of the final images created from the proposed method

(128 x 128 for each sample) using the Neural dataset.

[0035] Figure 26 shows a sample of the final images created from the real images (128 x 128 for each sample) using the Neural dataset.

[0036] Figure 27 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the Neural dataset.

[0037] Figure 28 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the CAR-T dataset.

[0038] Figure 29 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the Kaggle dataset.

[0039] Figure 30 shows a sample of the final images created from the real images (128 x 128 for each sample) using the CAR-T dataset.

[0040] Figure 31 shows a sample of the final images created from the real images (128 x 128 for each sample) using the Kaggle dataset.

DETAILED DESCRIPTION

[0041] Improved object detection and image segmentation for eukaryotic cells is described, which is suitable for generating training data, particularly of eukaryotic cells, for medical-related machine learning systems.

[0042] One of the most exciting, recent advancements in medical sciences is automating the analysis of microscopical images with machine learning tools. Specifically, many of these images are related to detection of nuclei cells. Creating well-varied and accurate datasets that are needed to train machine learning tools is time-consuming, costly, and labor- intensive. Through techniques described herein, it is possible to generate artificial cell nuclei microscopical images along with their correct instance segmentation masks. The resulting images can be used to assist in accessing higher generalization capabilities of artificial neural networks.

[0043] The described techniques can be used to reliably automate the ground truth generation of nuclei cells, which can prevent common human errors. In addition, by being able to generate a huge number of training data, it is possible to help the learning of artificial neural networks for related biological fields in a way that is not possible with current technologies.

[0044] Figure 1 illustrates a method of image creation for generating training data for image segmentation of eukaryotic cells and Figure 2 illustrates a process pipeline of a method of image creation.

[0045] Referring to Figure 1, a method 100 of image creation includes extracting (110) each instance of single cells in an image of eukaryotic cells and extracting (120) each instance of multi-cells in the image of eukaryotic cells. The extracting (110) of each instance of single- cells can include identifying each instance of single-cells in the image of eukaryotic cells, labelling each instance of single-cells as a single-cell, and storing each instance of single-cells in a single-cell resource. The extracting (120) of each instance of multi-cells can include identifying each instance of multi-cell groupings in the image of eukaryotic cells, labelling each instance of multi-cells as a multi-cell, and storing each instance of multi-cells in a multi- cell resource.

[0046] For example, as can be seen in Figure 2, an initial set of original images 200 can be segmented to extract cells and then sorted between two categories, the first being singular cells that can be easily distinguished from other cells and the second type being multi- cells, resulting in a set of single cells 210 and corresponding single-cell masks 220 and a set of multi-cells 230 and corresponding multi-cell masks 240. The original images 200 include instance segmentation masks, which may have been labeled by experts within the biomedical field or labeled by automated methods.

[0047] A multi-cell is a segmentation object consisting of multiple cells with intersecting bounding boxes. Multi-cells are identified as joint cells that more commonly suffer from the impurities such as low contrast of cell boundaries, background noise, adhesion, and cell clustering. Advantageously, by separately extracting (and using) multi-cells instead of generating multi-cells from single cells, it is possible to generate cells that have the natural properties of joint cells found in the data. Similarly, by separately extracting (and using) single- cells instead of generating single-cells from joint cells, errors such as unnatural imagery from using ambiguous borders of joint cells in creating single cells can be avoided. In some cases, when extracting the single-cells and multi-cells objects from the original images, those that are too close to the boundaries of the original images are ignored.

[0048] Returning to Figure 1, method 100 includes selecting (130) a set of cells from the extracted single-cells and the extracted multi-cells; and applying (140) at least one augmentation technique to each cell in the set of cells to generate augmented cells.

[0049] In operation 130, the set of cells selected from the extracted single-cells and the extracted multi-cells can be selected according to a distribution policy. The distribution policy can be based on mimicking the distribution of the number of cells (and noise) within the training data and the ratio of multi-cells to singular cells. In some cases, all the extracted single- cells and/or extracted multi-cells are selected for augmentation. In some cases, a subset of the extracted single-cells and/or a subset of extracted multi-cells are selected. In some cases, the selection of cells for the set of cells is random.

[0050] In operation 140, the at least one augmentation technique can be a pre- determined selection of a set of augmentation techniques, referred to as an augmentation policy. That is, the at least one augmentation technique applied to each cell in the set of cells corresponds to a particular augmentation policy of pre-determined selection of a set of augmentation techniques. Data augmentation is a technique used to extrapolate a dataset through the application of standard transformations onto image data. In this method, normal transformations are selected and applied to images through a pre-processing step. Augmentation utilizes policies, where each policy contains information on the different standard augmentation techniques that will be utilized, the probabilities of applying the technique, and the magnitude of the operation.

[0051] The set of augmentation techniques of a particular augmentation policy can be selected from the following 11 types of augmentation techniques.

[0052] FlipLR - where the image is flipped along the vertical axis.

[0053] FlipUD - where the image is flipped along the horizontal axis.

[0054] AutoContrast - where the contrast of the image is increased.

[0055] Equalize - where the histogram of the image is equalized.

[0056] Rotate - where the image is rotated by a certain number of degrees.

[0057] Posterize - where the number of bits for each pixel in the image is reduced.

[0058] Contrast - where the contrast of the image is changed.

[0059] Brightness - where the brightness of the image is adjusted.

[0060] Sharpness - where the sharpness of the image is adjusted. [0061] Smooth - where the image is smoothed (e.g., low-pass filtering).

[0062] Resize - where the image resolution is changed.

[0063] Of course, it should be understood that more or fewer techniques may be available as part of the types of augmentation techniques from which the set is selected. In addition, a set may contain a single augmentation technique or any number of augmentation techniques (e.g., up to 11 different types of techniques when selecting from the 11 types of augmentation techniques) such that the augmentation policies can include one or more augmentation techniques, depending on the corresponding set. In addition, certain policies may contain the same one or more types of techniques, but with different magnitudes. In some cases, a subset of the augmentation techniques is used (i.e., fewer than the 11 different types) in order to maintain the integrity of the original cells.

[0064] To identify the particular augmentation policy of a pre-determined selection of a set of augmentation techniques that is applied to each extracted single-cell and each extracted multi-cell, potential policies of combinations of augmentation techniques are scored and ranked. A specified number of potential policies are then made available for use to generate augmented cells. In some cases, the highest ranking potential policy is applied to generate the augmented cells in operation 140. In some cases, two or more of the highest ranking potential policies are applied to generate the augmented cells in operation 140 such that there will be more augmented cells available for generating the artificial image than original cells in the set of cells.

[0065] Automated augmentation finds the best sets of augmentation policies through an efficient search algorithm that aids in the determination of the strongest sets of transformations that improve accuracy in computer vision tasks. Search algorithms, such as Greedy AutoAugment (such as in synchronous mode), can be used to find the strongest augmentation policies in the search space of policies using a specified scoring criterion. A high- scoring augmentation policy transforms an image while maintaining its authentic properties, avoiding obscure augmentations.

[0066] To identify highest ranked policies, an autoencoder such as described with respect to Figure 3 can be used. An autoencoder neural network is an unsupervised learning method, which sets the target values to be equal to their given inputs, estimating an identity function. The identify function learns to compress and encode data for the purpose of reconstructing the data through the reduced encoded representation. Given real image data, this function learns the distribution of the data and gains the ability to estimate the realisticify of artificial data. It is important when an augmentation takes place, the policy does not alter the given nuclei cell to an unrecognizable state. The AutoEncoder attempts to leam an identity function that maps an object to itself, minimizing reconstruction error between inputs and its respective outputs. The features found through this neural network helps to retain a level of authenticity when manipulating an object. The best-performing sub policies can then be used as final policies applied in operation 140.

[0067] In a specific implementation, a Wasserstein Autoencoder is used. A Wasserstein Autoencoder chooses to minimize a Wasserstein distance, which is a distance function between two probability functions within a metric space, and provides a suitable scoring system for Greedy AutoAugment (described in more detail below, which ranks different combinations of sub-policies.

[0068] In some implementations, instead of using an autoencoder, diffusion methods may be used to identity' highest ranked policies of a set of augmentation techniques that can be applied to each extracted single-cell and each extracted multi-cell. Of course, other neural network algorithms may be used.

[0069] Method 100 further includes generating (150) an artificial image of eukaryotic cells using the augmented cells. The resulting artificial image of eukaryotic cells includes artificial cell nuclei microscopical images along with their correct instance segmentation masks.

[0070] For example, referring again to Figure 2, an image generation process 250 can involve selecting a set of cells from the single-cells 210 and the multi-cells 230 and then applying augmentation policies 260 to the set of cells. A set of noise 270 can be obtained for use in the image generation process 250. Adding noise helps to avoid overfitting within the created datasets. Each of the cell and mask objects (e.g., 210, 220, 230, 240) along with the policies 260 and noise 270 can be stored in a suitable storage resource 275. Storage resource 245 can be a single storage device or a plurality' of storage devices. In addition, storage resource 245 can conceptually be separate resources for the various object types, noise, and policies, or be a single resource with a suitable data structure or data structures for maintaining the different information.

[0071] When creating the artificial images 280 from the image generation process 250, the distribution of the number of cells (and noise 270) within the training data and the ratio of multi-cells to singular cells can be mimicked. For example, let j denote the mode of the distribution representing the number of cells found within the original images 200, then j policies (of the n available policies 260) are selected from the Greedy AutoAugment search algorithm and applied to j original cells. These newly augmented cells are then applied in a controlled manner onto a blank image, avoiding overlaps while maintaining the ratio between multi and singular cells.

[0072] This image generation process 250 (with application of augmentation policies 260 and noise 270) is also applied to the corresponding masks of the cell objects (e.g., from the single-cell masks 220 and the multi-cell masks 240). Thus, creating an artificial image with natural properties and complete segmentation masks (e.g., complete ground truth information). The image generation process can be performed until r images are generated (e.g., the artificial images 280 and artificial masks 290) with a matching distribution to the number of cells found within the original data (e.g., the original images 200).

[0073] As previously mentioned, multi-cells are used to generate augmented multi- cells and single-cells are used to generate augmented single-cells such that when a multi-cell is placed inside a frame of an artificial image, that multi-cell is selected from the separated collection of multi-cells 230 images and when a single-cell is placed inside the frame of the artificial image, the single cell is selected from the separated collection of single-cell 210 images.

[0074] Returning again to Figure 1, generating (150) the artificial image of eukaryotic cells using the augmented cells can include placing a selection of the augmented cells on a background image. The selection of the augmented cells may be all the augmented cells or a subset of the augmented cells. In some cases, the selected augmented cells are placed randomly (or pseudorandomly) on the background image. In some cases, such as described with respect to Figures 4A and 4B, the method 100 can further include creating the background image from the image of eukaryotic cells.

[0075] Figure 3 illustrates an autoencoder pipeline. Referring to Figure 3, the overall distribution of the individual cells can be determined using an AutoEncoder 300 trained to estimate an identity function, ƒ(S i ) « S i for i G [0, n], where a set of instance segmented objects S = (x 0 , ... , x n ) is defined as an initial set of single-cells and multi-cells, with an infinitely large search space, P, containing different augmentation policies. A goal is to find the best augmentation techniques that will output the most naturalistic nuclei images while maintaining complete segmentation masks.

[0076] The AutoEncoder 300 attempts to leam the function ƒ(S i ) where S i is a set of segmented cells that maps S i to itself. Here, S i contains all of the segmented obj ects 305 from the single-cells (e.g., single-cells 210 of Figure 2) and multi-cells (e.g., multi-cells 230 of Figure 2). The segmented objects 305 (the single-cells and the multi-cells) are obtained from the original training data 310 similar to that described with respect to Figure 2 (for extracting single-cells and multi-cells). These extracted cells are separated as sampled objects (S i ) (the segmented objects 305). The AutoEncoder 300 is used to map S i (represented in segmented objects 305) to (represented in segmented objects 315) to find the distribution of S i .

[0077] After training for a specific number of epochs, the next step is to use the trained AutoEncoder to create a criterion for the purpose of ranking different augmentation policies as described above with respect to the augmented cell generation process 140.

[0078] For example, to determine the quality of a single policy, the policy is applied to all members of S i . When a cell is augmented and passed through the AutoEncoder 300 and maps to the same exact cell without applying the policy, its output can evaluate the quality of that policy. With this mechanism, it is possible to make sure that the original cells can be changed but maintain their integrity. The purpose of this is when the augmentation takes place, the policy should not alter the cell to an unrecognizable state. In the searching phase, the accuracy of f(S i ) is passed as the scoring criterion for the Greedy Search to rank every explored policy.

[0079] The AutoEncoder gives a high-level evaluation of the quality of the augmented policies and creates the scoring criterion for Greedy AutoAugment.

[0080] Greedy AutoAugment is an efficient search algorithm that finds the best augmentation policies within an arbitrarily large sample space. After finding the best policies for single-cells and multi-cells objects, the cells accept policies and can be used to generate the artificial image as described with respect to operation 150 of Figure 1 and operation 250 of Figure 2

[0081] As mentioned above, a policy is used to perform data augmentation on an image. Each policy P (e.g., of policies 260 of Figure 2) includes the augmentation technique, the magnitude of the operation, and the probability. Therefore, a search mechanism for finding the best augmentation techniques is a search space that should consider all possible combinations of these three elements. Example augmentation techniques include those described above. The magnitude is the degree to which an operation is applied. For instance, in the rotation augment, the magnitude specifies how much an image is rotated. Finally, the third element specifies the probability of applying the augmentation to the image.

[0082] Search Space. The search space, S, consists of m sequential image operations. Each image operation can be defined as a sub-policy that contains information of two hyperparameters, the probability of applying the operation, and its magnitude. The range of probabilities and magnitude include two discrete variables within uniform spaces, n P , and n m . In this manner, a discrete search algorithm may be utilized to find the best sub-policies. The size of the search space for n o operations can be denoted as n s = (n o x n p x nm) l .

[0083] Score. A score is given to each sub-policy through the utilization of the autoencoder 300 that compares how well an augmented cell maps to its original variant. The score is a metric that measures the performance of a policy by passing a given cell through a trained Wasserstein AutoEncoder. The degree of accuracy provided from this method measures the closeness of an image to the original cell dataset and is passed to the search algorithm to select the strongest augmentation policies.

[0084] Search Algorithm. The size of the search space of potential augmentation policies grows exponentially due to the different combinations of varying transformations, their probability of applying a policy, and magnitude. This explains the impracticality of a brute force approach. In order to make the search process feasible, a greedy search is utilized. A reduced search space is traversed where each augmentation policy contains only a single sub- policy in the beginning, I = 1. In this reduced space, the best probability and magnitude for each of the varying image operations is identified through the scoring criterion mentioned earlier. Similarly, the best variables are identified for a second sub-policy, I = 2. For every found policy within the first stage, the best combinations of image operations are identified with their associated probabilities and magnitudes. This process is repeated until I = l max , the maximum number of sub-policies. All policies are then sorted via their scores, and the top b augmentations are selected to create the artificial training set.

[0085] In particular, before searching for the best policy, the space of probabilities and the magnitudes are discretized. The discretization of probabilities uses eleven values within uniform space (based on the eleven augmentation techniques described above), and the discretization of magnitudes uses ten values within uniform space. With this setting, the search space is simply (20 x 10 x 11) . This search space is defined to consider all possible combinations of elements in one sub-policy (the first search layer). If the search space is expanded in order to find all the possible combinations for two sub-policies, the search space increases its size to (20 x 10 x 11) 2 (two search layers). Continuously, it is possible to expand the search space for more layers infinitely. In general, the search space is defined as (20 x 10 x 11) l . where / is the number of layers within the search space.

[0086] To reduce the search space, Greedy AutoAugment considers the probability value of one, for all augmentation techniques. Additionally, instead of considering all possible combinations, only the best combinations from the strongest sub policies are considered. In other words, a greedy search algorithm is used, and expanding the search layers occurs only when it is required. Accordingly, the new number of the possibilities is defined as follows

[0087]

[0088] In this notation, k is an arbitrary integer number, which indicates the number of iterations the algorithm is allowed to perform within the search. The t and m variables represent the augmentation technique and magnitude. Respectively, t n and m n are the maximum values for t and m. To augment the data, a selection of the policies are used among all of the searched policies based on a scoring criterion. With the selected policies, the training data can then be expanded with the new augmented data as much as is required.

[0089] Figures 4A-4C illustrate a process pipeline of a method of image creation including creation of background images for the artificial images. Referring to Figure 4A, background images 400 can be obtained from original images 200 and used in image generation 410. Similar to that described with respect to Figure 2 using data (e.g., the cell and mask objects, 210, 220, 230, 240 obtained from original images 200, along with the policies 260 and noise 270) on the storage resource 275 (and operation 150 of Figure 1), the image generation process 410 for generating the artificial image 420 (and associated artificial mask 430) of eukaryotic cells using the augmented cells can include placing a selection of the augmented cells on a background image.

[0090] As mentioned above, the background image upon which the artificial image is formed can be created from the image of eukaryotic cells. This supports using color (and grayscale) backgrounds similar to that found in the original image (as opposed to a solid single- color background).

[0091] Referring to Figure 4B, a process 440 of creating empty backgrounds 400 from original images 200-A is shown. To create natural background images, it is possible to take advantage of the images acquired from the existing training datasets. Here, removal patches 450 are created and applied (460) to an original image to generate the empty background image. The removal patches 450 can include a bounding box for each segmented object (which may be single-cell object or multi-cell object), a masking image, and filter.

[0092] The process of creating empty backgrounds 400 from original images 200-A involves removing at least one instance of eukary otic cells from the image of eukaryotic cell by generating a bounding box around at least one instance of a eukaryotic cell in the image of eukaryotic cells, wherein pixels of the image of eukaryotic cells inside the bounding box are inside pixels; and replacing values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box to remove the at least one instance of the eukaryotic cell in the image of eukaryotic cells, wherein the certain pixels found outside of the bounding box are outside pixels.

[0093] For example, the process can include performing a method that includes generating a bounding box around the at least one instance of eukaryotic cells; selecting outside pixels, where the outside pixels are selected from pixels in the image of eukaryotic cells that are outside of the limits of the bounding box; comparing the outside pixels to at least one pixel in each of the bounding box comers; determining which of the outside pixels has the highest degree of similarity to each comer of the bounding box comers; selecting an inside pixel, where the inside pixel is selected from pixels within the at least one instance of eukaryotic cells; determining a closest comer, wherein the closest comer is the comer of the bounding box comers that is closest to the inside pixel; and replacing the inside pixel with the outside pixel having the highest degree of similarity to the closest comer. Replacing values of inside pixels with outside pixels can include for each inside pixel: selecting a predetermined number of outside pixels, comparing each of the predetermined number of outside pixels to pixels at comers of the bounding box to identify an outside pixel of the predetermined number of outside pixels that is most similar to the pixels at the comers of the bounding box, and replacing that inside pixel with the identified outside pixel.

[0094] The method can further include creating a masking image, where the masking image is an image with a black-colored background having a smaller, white area in the middle, wherein the masking image is the same size as the bounding box; and applying a Gaussian filter to the masking image to generate a filtered masking image. The filtered masking image is inserted in the same position as the bounding box; and an area of the image of eukaryotic cells is smoothed where the at least one instance of eukaryotic cells was removed.

[0095] Figures 5A-5F show removal of cells from a real image to create an artificially generated background. Figure 5A shows a sample of the real dataset from the Neural dataset. Figure 5B shows the process of removing the first cell from the image. Figures 5C, 5D, 5E, and 5F show the results of removing four other cells.

[0096] In the original images, the existing cells need to be removed and then replaced with similar pixels that replicate the textures of the images. For example, Figure 5A shows a sample from the Neural dataset in its original form with five cells that need to be removed. To remove the cells, segmented areas are used to create bounding boxes around the cells. If the bounding boxes are interconnected, a bigger bounding box is created to contain all the connected boxes. Next, the pixels inside each bounding box (inside pixels) need to be replaced by other pixels found outside of the bounding box (outside pixels) that are similar to the bordering pixels of the bounding box. For each inside pixel, n outside pixels are selected randomly and compared to the pixels of the upper-left comer, upper-right comer, lower-left comer, and lower-right comer of the bounding box. Among the selected pixels, the outside pixel that is most similar to the mentioned comers of the bounding box (evaluated using the mean Euclidean distance) is used to replace the inside pixel. This process is repeated until all inside pixels are replaced with outside pixels.

[0097] Through this method, cells are removed from the bounding boxes with a similar texture to the original background. However, there would still be notable discrepancies between the background and transformed bounding boxes. To impede these problems and create smoother textures, a masking technique is used that utilizes a Gaussian filter to replace the previous bounding box with a new one. In this method, an image with similar size to the bounding box is created with a black-colored background and a white rectangular area in the middle that acts as the mask. Next, a Gaussian filter is applied to disperse the white rectangle based on Gaussian distribution. The filtering process aids in a smooth transition between the textures of the background and the newly created bounding box. The new mask is finally used to copy the newly created bounding box onto the image. By repeating this process for each segmented cell, all of the cells are removed from the image. Figure 5B shows the process of the removal of the first cell. At the top of each image in Figures 5B-5E, there are small rectangles that show the process needed to remove the cells and replace them with a smooth texture. From left to right, the steps include a newly created bounding box, the preliminary mask, a filtered version of the mask using a Gaussian Filter and, an applied filtered mask on the new bounding box. As can be seen in Figure 5F, the pixel replacement process is effective and the removed areas are not distinguishable from the other background areas. The newly generated background can now be used to place augmented cells to create artificial images.

[0098] Referring to Figure 4C, cells can be placed on the empty backgrounds during the image generation process 410 to generate the artificial images 420. The placement of cells on a background shares similar problems to the removal of cells. The brighter color spectrum found in the background and the cells signifies the differences in textures between them, which results in unnatural looking images. Accordingly, similar processes as performed to remove original cells can be carried out to create smooth transitions for the augmented cells. [0099] Indeed, cell patches 470 (which can involve the augmented single-cells and augmented multi-cells created as described with respect to Figures 1 and 2) can be applied (480) to the empty backgrounds 400 to create the artificial images 420. The method can include placing at least one cell from the selection of the augmented single-cells and the augmented multi-cells onto the background image by generating a bounding box around the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; comparing pixels from four comers of the bounding box to pixels in the background image to determine a target location, where the target location is an area of the background image where the pixels in the background image have the highest degree of similarity to the pixels in the four comers of the bounding box; placing the at least one of the selection of the augmented single-cells and the augmented multi-cells in the target location; creating a cell masking image, where the cell masking image includes a black color background and a white area in the middle, wherein the white area is substantially the same shape as the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; and applying a Gaussian filter to smooth the transition between texture of the background image and texture of the at least one cell from the selection of the augmented single-cells and the augmented multi-cells. An illustrative example is shown in Figures 6A-6F.

[0100] Figures 6A-6F show placement of cells on an artificially generated background. Figure 6A shows an empty background (the artificially generated background shown in Figure 5F). Figure 6B shows the process of adding the first cell to the background. Figures 6C-6F show the process of adding four other cells. For Figures 6B-6F, each small rectangle at the top of the image shows the process performed to add a cell with a smooth transition. From left to right, process involves the original bounding box of the cell, the new bounding box taken from the new location of the cell, a filtered version of the manually segmented mask using a Gaussian filter, and an applied filtered mask on a searched bounding box from the background.

[0101] In detail for this example implementation, manually segmented areas can be used to create bounding boxes around each cell. If the bounding boxes are interconnected, a larger bounding box is generated to contain all the connected boxes. Next, the background image is searched for areas in the image that most closely resemble the pixel colors of the generated bounding boxes. For each bounding box, n number of x, y points are randomly selected. In order to select the best x, y pair, the upper-left comer, upper-right comer, lower left comer, and lower-right comer of the bounding box are compared with x, y, x + w, y + h pixels, in which w, h are the width and height of the bounding box. Among the selected x, y pairs, the x, y pair that is the most similar to the mentioned comers of the bounding box (evaluated using the mean Euclidean distance) is used to position the bounding box.

[0102] Finding a place where two textures have similar colors is helpful, but it is often not enough to avoid notable discrepancies between the backgrounds and the newly placed cells. To solve this problem and create a smooth transition between the cell and the background, a Gaussian Filter technique is applied. In order to place a cell in a background image, first, an image with similar size to the bounding box is created with a black color background and a white area in the middle that acts as a mask. Manual segmentation of the cell is then used to determine the shape of the white area. Next, a Gaussian filter is applied to disperse the white segmented area through a Gaussian distribution. The filtering process aids in a smooth transition between the texture of the background and the segmented cell. The new mask is then used to copy the separated bounding box from the cell onto the image. By repeating this process for all the segmented cells, any desired number of cells can potentially be placed on the image.

[0103] As can be seen in Figures 6B-6F, the adding process is effective, and the added area is not distinguishable from the original image sample.

[0104] The newly generated images can be used to train artificial neural networks. Note that for the segmentation images, the original unfiltered mask is used instead of Gaussian filter masks due to the Gaussian filter being less deterministic and may include areas other than the cells, which is not desired.

[0105] Experiments and Evaluation of Results

[0106] The inventors use the proposed method to generate artificially augmented images on four prominent GAN models. The DCGAN is specifically designed to generate synthetic images for large-scale datasets. The small dataset Big- GAN (BigGAN-SD) is one of the first serious attempts to introduce GAN for small datasets. Data-Efficient GAN, which is more optimized for smaller datasets, was selected for the experiments. There are two variants of Data-Efficient GAN for two state-of-the-art methods, one is StyleGAN2, and the other is BigGAN. In the paper, the inventors call these two variants BigGAN-Diff and StyleGAN2- Diflf. While these models are not designed to maintain instance segmentation masks for their image objects, they can still be used to compare the quality of the images created from our method.

[0107] Evaluation metrics. The inventors use two known evaluation metrics that are used for the GAN networks. Frechet Inception Distance (FID) calculates the Frechet distance between two multidimensional Gaussian distributions. This technique compares the distribution of the generated images, along with the mean and variances of the Gaussian distributions between artificial and original images. Kernel Inception Distance (KID) compares the two probability distributions by drawing samples independently from each distribution. This method improves FID and acts as a more reliable and unbiased estimator.

[0108] Experiment setup. All of the experiments were conducted on K80 NVIDIA graphic cards. As for the embedded deep neural networks for our method, the inventors used Pytorch (Paszke, A., et al. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pages 8024-8035. Curran Associates, Inc., 2019). For Greedy Auto Augment, the inventors used the implementation described in Naghizadeh, A., et al. Greedy AutoAugment, 2020 (arXiv: 1908.00704 (cs.LG]) and for the AutoEncoder the inventors used the Wasserstein variation (see Tolstikhin, I., et al. Wasserstein auto-encoders, 2019 (arXiv: 1711.01558 [stat.ML]) and Gulrajani, I., et al. Improved training of wasserstein gans, 2017 (NIPS’ 17 Proceedings of the 31 st International Converence on Neural Information Processing Systems, December 2017, pages 5769-5779; arXiv: 1704.00028 [cs.LG]). For the experiments, all of the models are compared with their baseline released codes from the authors. The inventors did not alter the suggested configuration from the official papers. To reduce computations, the inventors searched the search space of augmentation policies within the maximum number of 20000 policies, in which the first 1000 policies are used. All of the GAN models were trained for 150000 epochs across all experiments.

[0109] The inventors mainly use two cell nuclei datasets containing microscopical images obtained from two different biological fields. The CAR-T cell dataset consists of 156 images at a resolution of 1024 x 1024. This dataset was collected by Rutgers Cancer Institute of New Jersey using microscope system AIR HD25 (Nikon, Japan), which provides a 25 mm field of view. The second dataset is the Kaggle 2018 Data Science Bowl (Bowl- 18), which contains around 589 images that were acquired under a variety of conditions and vary in the cell type, magnification, and imaging modality (brightfield vs. fluorescence). A subset of 193 images with consistent attributes is used.

[0110] Single-Cell Image Generation

[0111] Single-cell images are the most basic elements in the creation of complete artificial images. To create artificial single-cells, the inventors use extracted segmentations from experts. The cells should then transform from its original state to a new state using an augmented policy. The outcome is an artificial cell that can be used to create artificial images. In this section, the inventors measure the accuracy of these generated cells through the use of the evaluation methods discussed above (FID and KID). In other words, the inventors answer how much applying different augmentation policies affects the quality of single-cells. The results are shown in Table l.The average scores for two datasets are reported. The numerical values apply to the average scores and are computed over three random sets of sample generation. In the table, the proposed method is referred to as Instance Aware Automatic Augmentation (IAAA).

[0112] Table I - FID and KID scores for single cells

1 represents a score using the CAR-T dataset, 2 represents a score using the Kaggle dataset

[0113] F rechet Inception Distance. The results show that the proposed method performs competitively compared to existing models. The best performing GAN model is StyleGAN2-Diff, which is outperformed by 17.6 with the proposed method using the CAR-T data. When using the Kaggle dataset, StyleGAN2-Diff scores 10.36, which is marginally better than the proposed method.

[0114] Kernel Inception Distance. The KID scores show that the proposed method has a higher score than StyleGAN2-Diff by 0.01105 using the CAR-T data. When using the Kaggle dataset, StyleGAN2-Diff scores .0067, which is better than the proposed method.

[0115] Due to the simplified nature of single-cells, most methods performed well in producing strong artificial single-cells. The results for the image generation are presented in Figures 7A-7C. Figures 7A-7C show single-cell image generation of CAR-T and Kaggle datasets. In Figure 7A, the first two rows and last two rows represent the single-cell images and their masks from DAR-T dataset and Kaggle dataset, respectively. Figure 7B shows the single-cell images and their associated masks that were generated by the described method. The first two rows are the single-cell images generated from the CAR-T dataset and their masks. The last two rows are the single-cell images generated from the Kaggle dataset and their masks. Figure 7C shows the single-cell images generated by Sl leGan2-Diff. In the generation of single-cell images, StyleGan2-Diff had the best performance among selected GAN models. In particular, StyleGAN2- Diff had the best FID/KID scores within the GAN comparisons. As mentioned above, the GAN methods do not produce segmentation results. The proposed method gives competitive, if not better results compared to selected GAN methods and outputs images with a great likeness to the original single-cells. This shows that the controlled environment was successful in the generation of single-cells and preserves the original traits of each natural cell.

[0116] Multi-Cells Image Generation

[0117] Multi-cells images are another basic element for cell nuclei microscopical images. The artificial generation of multi-cells is more challenging for GAN models because their numbers are generally much lower than single-cells. This should not affect the model, which has minimal dependence on the number of images to produce high-quality cells. Similar to the previous section, the inventors investigate how the proposed method compares to the GAN models used in the creation of realistic images. Respectively, the inventors perform experiments on CAR-T and Kaggle datasets. The scores are reported in Table 2.

[0118] Table 2 - FID and KID scores for multi-cells

1 represents a score using the CAR-T dataset, 2 represents a score using the Kaggle dataset

[0119] Frechet Inception Distance. The results show that the proposed method greatly outperforms the existing models across all parameters. The FID score of the proposed method, surpasses the best performing GAN model StyleGAN2-Diff by 68.41 within the CAR-T dataset and 59.11 within the Kaggle dataset.

[0120] Kernel Inception Distance. Similar to the FID scores, the data shows the proposed method has a better generation quality for multi-cells. For KID, the best performing GAN model is again StyleGAN2-Diff, which is outperformed by 0.04365 by the proposed method within the CAR-T dataset and 0.0267 within the Kaggle dataset. [0121] The evaluation scores show the difficulty of generating naturalistic multi-cells due to the many variables that result in cell clustering and adhesion. The results to produce multi-cells using the CAR-T dataset and Kaggle dataset are presented in Figures 8A-8C. Figures 8A-8C show multi-cells image generation on CAR-T and Kaggle datasets. In Figure 8A, the first two rows and last two rows represent the multi-cells images and their masks from CAR-T dataset and Kaggle dataset, respectively. Figure 8B shows the multi-cells images and their associated masks that were generated by the described method. The first two rows are the multi-cells images generated from the CAR-T dataset and their masks. The last two rows are the multi-cells images generated from the Kaggle dataset and their masks. Figure 8C shows the multi-cells images generated by StyleGan2-Diff In the generation of multi-cells images, StyleGan2-Diff had the best performance among selected GAN models. Similar to the single- cells, samples are provided for observation. As can be seen, the proposed method consistently provides better results compared to GAN for multi-cells. This shows that the controlled environment was again successful in generating new multi-cells while simultaneously preserving the traits found within the original cells.

[0122] Cell Nuclei Image Generation

[0123] The proposed method utilizes two sets of images (single-cells and multi-cells), along with recognized noise background, to place them into single frames. The quantity of single-cells and multi-cells are important for creating diverse images. Since the inventors are utilizing small sets of training data, generating artificial data seems more difficult for GANs. The results are shown in Table 3.

[0124] Table 3

1 represents a score using the CAR-T dataset, 2 represents a score using the Kaggle dataset [0125] F rechet Inception Distance. The proposed method was able to consistently outperform the best performing GAN method, StyleGAN2, by 29.73 with the CAR-T dataset and 93.93 with the Kaggle dataset.

[0126] Kernel Inception Distance. The results remained consistent with the FID values, with the proposed method outperforming St leGAN2 by 0.97 using the CAR-T dataset and 0.0176 using the Kaggle dataset.

[0127] The final results of the creation of complete CAR-T cell images are presented in Figures 9A-9G. Figures 9A-9G show cell nuclei image generation on CAR-T and Kaggle datasets. In Figure 9A, the first two rows show original CAR-T dataset, and the last two rows show original Kaggle dataset (real images and masks, respectively). Figures 9B and 9C are the CAR-T nuclei images generated by our method and their associated masks. Figures 9D and 9E are the Kaggle nuclei images generated by our method and their associated masks. Figures 9F and 9G represent CAR-T images and Kaggle images generated by StylGAN2- Diff. The results show the precise segmentation information that the described method provides, as well as demonstrating the clear advantage in the creation of high-quality images over alternative GAN solutions.

[0128] Performance on Real Applications

[0129] In Table 4 Part 1 , the inventors tested the augmentation algorithm with accuracy (AP score) of 11 state-of-the-art detection algorithms. The values on the left side indicate the AP scores from original algorithms, and the values on the right side indicate the AP scores after applying 100 augmented data to the original training sets. The results show that the inventors could consistently improve the results of these algorithms (30 out of 33).

[0130] Complex Datasets: While the dark background was used to create synthetic photo-realistic images, datasets with higher spectral color can still benefit from the method for certain tasks. In Table 5 Part 1, the inventors report the results for bounding box detection for the Neural cells dataset.

[0131] Table 4 Part 1

[0132] Conditional GAN The GAN methods in their pure form cannot be used to improve the bounding box detection because they do not produce the respective labels which are required for comparisons. In Table 4 Part 2, the inventors use the methodology described in Mahmood, F., et al. Deep adversarial training for multi-organ nuclei segmentation in histopathology images. IEEE transactions on medical imaging, 39(1 1):3257-3267, 2019 for comparing the method with conditional GAN. By applying 100 extra augmentation data to the original training, the inventors could increase the accuracy by 10.79% and 25.67% for CART and Kaggle. The accuracy here is the mean of the Jaccard scores between pairs of generated and original masks for all test points (see Kumar, N., et al. A dataset and a technique for generalized nuclear segmentation for computational pathology. IEEE transactions on medical imaging, 36(7): 1550-1560, 2017).

[0133] Table 4 Part 2

[0134] Instance Segmentation. For instance segmentation, the results of the mixed CAR-T dataset are at least 0.39 % and up to 2.86% higher than results of the real data on AP. It also shows improvement of at least 0.65% and up to 1.18% when compared to the reference CAR-T dataset at lOUs. The details of these experiments are presented in supplementary. The results show that the method could easily improve the accuracy of instance segmentation of cells.

[0135] As can be seen from the first set of experiments, the inventors presented Instance Aware Automatic Augmentation for the generation of artificial cell nuclei microscopical images along with their correct instance segmentation masks. An initial set of segmentation objects are used with Greedy AutoAugment to find the best performing policies. The found policies and initial set of segmentation objects are then used to create the final artificial images. The images are compared with state-of-the-art data augmentation methods. The results show that the quality of the proposed method in different stages is on par with the original data and surpasses different GAN models. These observations are confirmed by FID and KID scores. The proposed method is effectively ready to generate artificial cell nuclei microscopical images.

[0136] Experiments and Evaluation of Results

[0137] The next set of experiments and evaluation results again utilize the GAN models described in the first experiments and evaluation of results section.

[0138] Evaluation metrics. To measure the improvement in bounding box detection, the inventors use variants of average precision (AP) to evaluate and compare how different object detectors perform given artificial data. Average precision computes the average precision value for recall values over 0 to 1. AP50 and AP75 are metrics used to evaluate the intersection over a union of objects (loU). A perfect prediction yields an loU value of 1, while a completely wrong detection yields a value of 0. A degree of overlap gives a value between the two. A decision can be made on how much overlap can potentially be considered a correct prediction. Hence, AP50 and AP75 gives the average precision with loU thresholds of 0.5 and 0.75, respectively. The APs represents how well the model performs with small objects. Since APs is used to evaluate the identification of small objects, it is specifically important for nuclei cell detection.

[0139] Experiment setup. All of the experiments were conducted on K80 NVIDIA graphic cards. As for the embedded deep neural networks for our method, the inventors used PyTorch. For Greedy AutoAugment and Wasserstein variation the inventors used the official implementations. For the experiments, all of the models were compared with their baseline released codes from the authors. The inventors did not alter the suggested configuration from the official papers. To reduce computations, the inventors search the search space of augmentation policies within the maximum number of 20000 policies, in which the top 1000 performing policies are used. All of the GAN models were trained for 150000 epochs across all the experiments.

[0140] The inventors use three cell nuclei datasets containing microscopical images obtained from two different biological fields. The CAR-T cell dataset consists of 156 images at a resolution of 1024 x 1024, divided into train, test, and validation sets with 93, 31, and 32 images. The second dataset is from the Kaggle 2018 Data Science Bowl, which contains around 589 images that were acquired under a variety of conditions and vary in the cell ty pe, magnification, and imaging modality (brightfield vs. fluorescence). A subset of 193 images, which share consistent attributes, and is divided into train, test, and validation sets with 117:38; and 38 images. To test bright color images the inventors use the Neural dataset consisting of 644 images at a resolution of 640 x 512, also divided into train, test, and validation sets with 386, 129, and 129 images.

[0141] Single-Cell Image Generation

[0142] Due to the simplified nature of single-cells, most methods perform well in producing strong artificial single-cells. The results for the image generation are presented in Figures 10A-10C. Figures 10A-10C show cell nuclei image generation on CAR-T (Figure 10A), Kaggle (Figure 10B), and Neural datasets (Figure 10C). The first column represents the reference (real) images, and the second and third columns are images generated by the described method. The images are accompanied by their associated masks. For reference, a zoomed in image of five regions of the images for single cells in the second column and multi- cells in the third column are shown. The first column zooms into five regions that include both single-cell images and multi-cell nuclei images. The second column of Figure 10A (artificial single-cell samples) shows the single-cell images and their associated masks that were generated by our method for the CAR-T dataset. The second column of Figure 10B shows the single-cell images and their associated masks that were generated by the proposed method for the Kaggle dataset. Respectively, the first three images of the first column (real single-cell and multi-cell samples) show' the original single-cell images with their labels. In all the images, the first row shows the original images, and the second row is for zoomed in regions. From the images, the proposed method produces outputs with a great likeness to the original single-cells. This shows that the controlled environment was successful in the generation of single-cells and preserves the original traits of the natural cells found in the raw data.

[0143] Multi-Cell Image Generation

[0144] The results to produce multi-cells using the CAR-T dataset and Kaggle dataset are also shown in Figures 10A-10C. Similar to the single-cells, samples are provided for observation. The third column of Figure 10A (artificial multi-cell samples) shows the multi- cell images and their associated masks that were generated by the proposed method for CAR- T dataset. The third column of Figure 10B shows the multi-cell images and their associated masks that were generated by the proposed method for Kaggle dataset. Respectively, the last two images of the first column (real single-cell and multi-cell samples) show the original multi - cell images with their labels. The results show that the controlled environment was again successful in generating new multi-cells while simultaneously preserving the traits found within the original cells.

[0145] Performance of Real-World Bounding Box Detection

[0146] In this section, the inventors show the performance improvements of bounding box detection when the proposed augmentation method is applied. For training, the inventors follow the standard practice for bounding box detection algorithms and train for 12 epochs. The experiments involve both multi-stage and one-stage bounding box detection algorithms. The inventors use the following state-of-the-art studies for bounding box detection, Faster R- CNN w/ FPN 21 , Cascade R-CNN 22 , Grid R-CNN 23 , Libra R-CNN 24 , RepPoints 25 , FreeAnchor 26 , FSAF 27 , ATSS 28 , PAA 29 , GFL 20 . These networks are implemented on MMDetection software, based on ResNet-101 (R-101) and ResNeXt-101 (X-101) backbones.

[0147] Table 5 shows the Ap, AP50, AP75, and APS scores of bounding box (bbox) detection on CAR-T, Kaggle, and Neural datasets. The values for the datasets represent the original complete cell nuclei images and the values for AUG+Datasets represent the original complete cell nuclei images with the addition of augmented images.

[0148] Table 5

[0149] CAR-T Dataset The inventors first investigate the effectiveness of the method on the CAR-T dataset (the left-most numbers for each column in Table 5). According to the results, the augmented data consistently improves the results. For the AP scores, 11 out of 11 methods are improved. The AP50 shows that 10 out of 11 methods show significant improvements, the AP75 scores show that 11 out of 11 methods improve, and the APs indicates that 9 out of 11 methods are improved. The improvement of the methods is consistent, with an average 5.94% increase in overall AP evaluation across all the experiments.

[0150] Kaggle Dataset. The inventors investigated the effectiveness of the proposed augmentation method on the Kaggle dataset (the middle numbers in Table 5). According to the results, the inventors additionally observe that the augmented data consistently improves the results when using the Kaggle dataset. For the AP, 10 out of the 11 methods show improvements. The AP50 shows that 6 out of the 11 methods are improved, the AP75 shows that 10 out of the 11 methods are improved, and the APs also shows a performance increase across 10 out of the 11 methods. Overall, the improvements to the 11 methods are consistent. The average performance gain is 1.77% across the experiments.

[0151] Experiments for Bright Color Microscopical Images

[0152] To test the proposed method for images in the high spectrum color, the inventors use the Neural dataset. In the following, the average scores for FID and KID scores for this dataset are reported.

[0153] Frechet Inception Distance. The FID scores for DCGAN, BigGAN-SD, BigGAN-Diff, StyleGAN2-Diff, and Sem-Aware are in order, 224.33278.4576.20.37.73, and 52.49. The proposed method was able to create natural images and produce the second-best FID score when creating artificial images in 128 x 128 resolution.

[0154] Kernel Inception Distance. The results remained consistent with the FID values. The KID scores for DCGAN, BigGAN-SD, BigGAN-Diff, StyleGAN2-Diff, and Sem- Aware are in order, .1803, .2361, .0619, .0147, and .0310. The proposed method was able to create natural images and produce the second-best KID score when creating artificial images in 128 X 128 resolutions.

[0155] As mentioned above, the final results of the creation of complete Neural cell images are presented in Figure 10C. The second column of Figure 10C shows the single-cell images and their associated masks that were generated by the proposed method for the Neural dataset. Similar to the CAR-T and Kaggle results, the Figure 10C represents zoomed in regions for single-cells and multi-cells along with samples from the original dataset.

[0156] The inventors investigated the effectiveness of the proposed augmentation method on the Neural dataset (also shown in Table 5). According to the results, the inventors observe that the augmented data consistently improves the results. The AR evaluation shows improvements for 10 out of the 11 methods. The AP50 shows that 7 out of the 11 methods are improved, the AP75 shows that 10 out of the 11 methods are improved, and finally the APs shows that 10 out of the 11 methods are improved. The average performance increase of AP evaluation is 2.15% across the experiments. Although, this dataset contains images with brighter colors, the performance gam is consistent with the gams observed within the CAR-T and Kaggle datasets.

[0157] The inventors presented Instance Aware Automatic Augmentation for the generation of artificial cell nuclei microscopical images along with their correct instance segmentation labels. An initial set of segmentation objects are used with Greedy AutoAugment to find the best performing policies. The found policies and initial set of segmentation objects are then used to create the final artificial images. The images are compared with state-of-the- art data augmentation methods. The results show that the quality of the proposed method in different stages is on par with the original data and surpasses different GAN models. These observations are confirmed by FID and KID scores. The experiments demonstrate that the proposed augmentation technique consistently improves the detection of cells which is confirmed by variants of AP scores. The proposed method is effectively ready to generate artificial cell nuclei microscopical images. In the future, this method can help to better train microscopical images with artificial neural networks. [0158] Additional Experiments

[0159] For the additional experiments, the inventors use data generated by the proposed method to improve cell segmentation accuracy and the robustness of model training results. All of our experiments in supplementary are conducted on RTX Titan NVIDIA graphic cards. As for the embedded deep neural networks for our method, the inventors used Pytorch. For WAE, the inventors used the implementation in and for the KG Instance Segmentation the inventors used the implementation by the author.

[0160] Data Preparation

[0161] Original Datasets The inventors used three cell nuclei datasets containing microscopical images obtained from two different biological fields. The CAR-T cell dataset consists of 156 images at a resolution of 1024 X 1024, divided into train, test, and validation sets with 93;31; and 32 images. The second dataset is from the Kaggle 2018 Data Science Bowl, which contains around 589 images that were acquired under a variety of conditions and vary in the cell type, magnification, and imaging modality (brightfield vs. fluorescence). A subset of 193 images, which share consistent attributes, and is divided into train, test, and validation sets with 117;38; and 38 images. To test bright color images the inventors use the Neural dataset consisting of 644 images at a resolution of 640 x 512, also divided into train, test, and validation sets with 386;129; and 129 images.

[0162] Mixed Datasets The inventors mixed the images generated by the proposed method with the original images. The mixture is only for the training sets and not validation and test sets. After combining, the training set for CAR-T cell dataset contains 200 images which is a combination of 93 original images and 107 generated images. The training set for Kaggle 2018 Data Science Bowl dataset contains 200 images which is a combination of 117 original images and 83 generated images. The training set of the Neural dataset contains 800 images which is a combination of 386 original images and 414 generated images.

[0163] Evaluation In order to verify the effect of our method on real data, the training results of the two datasets (Original Datasets and Mixed Datasets) are both evaluated on the real test sets for all three datasets.

[0164] As shown in Table 6 and Table 7, for the BBox detection performances, the training results of the mixed dataset is at least 0.15 % and at most 7.63% better than the results of the real dataset on the AP. For instance segmentation, the results of the mixed dataset is at least 0.49 % and up to 16.85% higher than results of the real data on AP. It also shows improvement of at least 0.04% and up to 1.87% when compared to the reference datasets at lOUs [0165] Table 6

[0167] The results show that the method could easily improve the accuracy of object detection and instance segmentation of cells. The reason is that augmentation generates a lot of new data that benefits the generalization of the training process, which reduces overfitting and improves the robustness of the model. The experiments show that there should be a balance between real and artificial samples. The inventors observed that if the number of artificial samples is around the same number of real training data, it can generally increase the accuracy. For higher orders of artificial data, if the real data can provide high accuracy, overshadowing the data with artificial samples can have negative results. However, if the original data cannot provide high accuracy, higher artificial samples generally lead to better results.

[0168] The following Figures show additional generated image samples

[0169] Figure 11 shows sample of an artificially generated Neural dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the generated cells and their respective masks.

[0170] Figure 12 shows a real sample from the Neural dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the real cells and their respective masks.

[0171] Figure 13 show's a sample of an artificially generated CAR-T dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the generated cells and their respective masks.

[0172] Figure 14 shows a real sample from the CAR-T dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the real cells and their respective masks.

[0173] Figure 15 shows a sample of an artificially generated Kaggle dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the generated cells and their respective masks. [0174] Figure 16 shows a real sample from the Kaggle dataset (first row) with its respective instance segmentation (second row). In six different locations, the zoom-in functionality is used to provide better details of the real cells and their respective masks.

[0175] Figure 17 shows single-cell real images on CAR-T, Neural, and Kaggle datasets; The first two rows represent the single-cell images and their masks from CAR-T dataset. The third and fourth rows represent the single-cell images and their masks from Neural dataset. The fifth and sixth rows represent the single-cell images and their masks from Kaggle dataset.

[0176] Figure 18 shows single-cell image generation on CAR-T, Neural, and Kaggle datasets; The first two rows represent the single-cell images and their masks from CAR-T dataset. The third and fourth rows represent the single-cell images and their masks from Neural dataset. The fifth and sixth rows represent the single-cell images and their masks from Kaggle dataset.

[0177] Figure 19 shows multi-cell real images on CAR-T, Neural and Kaggle datasets; The first two rows represent the multi-cell images and their masks from CAR-T dataset. The third and fourth rows represent the multi-cell images and their masks from Neural dataset. The fifth and sixth rows represent the multi-cell images and their masks from Kaggle dataset.

[0178] Figure 20 shows multi-cell image generation on CAR-T, Neural and Kaggle datasets; The first two rows represent the multi-cell images and their masks from CAR-T dataset. The third and fourth rows represent the multi-cell images and their masks from Neural dataset. The fifth and sixth rows represent the multi-cell images and their masks from Kaggle dataset.

[0179] Figure 21 shows a comparison between complete image generation of Neural dataset; The first row represents 16 images from StyleGAN2-Diff The second and third rows represent images that are generated from the proposed method. The fourth and fifth rows represent real images and their respective masks.

[0180] Figure 22 shows images of the original CAR-T dataset and the original Kaggle dataset. The first two rows show original CAR-T dataset, and the last two rows show original Kaggle dataset.

[0181] Figure 23 shows a sample using different forms of augmentation techniques on the cells. Each row represents a new augmentation technique applied on a multi-cell within the CAR-T dataset. The augmentations used in the order of rows are: ‘Rotate’, ‘Posterize’, ‘CropBilinear’, ‘Solarize’, ‘Color’, ‘Contrast’, ‘Brightness’, ‘ShearX’, ‘ShearY’, ’TranslateX’, ‘TranslateY’, ‘Cutout’, ‘Equalize’, ‘Invert’, ‘AutoContrast’. [0182] Figure 24 shows a sample using different forms of augmentation techniques on the cells. Each row represents a new augmentation technique applied on a multi-cell within the Kaggle dataset. The augmentations used in the order of rows are: ‘Rotate’, ‘Posterize’, ‘CropBilinear’, ‘Solarize’, ‘Color’, ‘Contrast’, ‘Brightness’, ‘ShearX’, ‘ShearY’, ‘TranslateX’, ‘TranslateY’, ‘Cutout’, ‘Equalize’, ‘Invert’, ‘AutoContrast’.

[0183] Figure 25 shows a sample of the final images created from the proposed method (128 x 128 for each sample) using the Neural dataset.

[0184] Figure 26 shows a sample of the final images created from the real images (128 x 128 for each sample) using the Neural dataset.

[0185] Figure 27 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the Neural dataset.

[0186] Figure 28 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the CAR-T dataset.

[0187] Figure 29 shows a sample of the final images created from the StyleGAN2-Diff method (128 x 128 for each sample) using the Kaggle dataset.

[0188] Figure 30 shows a sample of the final images created from the real images (128 x 128 for each sample) using the CAR-T dataset.

[0189] Figure 31 shows a sample of the final images created from the real images (128 x 128 for each sample) using the Kaggle dataset.

[0190] The following provides some example methods of generating artificial images for medical evaluation of eukaryotic cells. These methods can be implemented as instructions stored on and executed by a computing system.

[0191] Clause 1. A method of generating artificial images for medical evaluation of eukaryotic cells, comprising: extracting each instance of single-cells in an image of eukaryotic cells; extracting each instance of multi-cells in the image of eukaryotic cells; generating a background image from the image of eukaryotic cells; selecting a set of cells from the extracted single-cells and the extracted multi-cells; applying at least one augmentation technique to each cell in the set of cells to generate augmented cells; and generating an artificial image of eukaryotic cells using the augmented cells and the background image.

[0192] Clause 2. The method of clause 1, wherein generating the background image from the image of eukaryotic cells comprises: generating a bounding box around at least one instance of a eukaryotic cell in the image of eukarvotic cells, wherein pixels of the image of eukaryotic cells inside the bounding box are inside pixels; and replacing values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box to remove the at least one instance of the eukaryotic cell in the image of eukaryotic cells, wherein the certain pixels found outside of the bounding box are outside pixels.

[0193] Clause 3. The method of clause 2, wherein generating the background image from the image of eukaryotic cells further comprises: performing the generating of the bounding box around the at least one instance of the eukaryotic cell in the image of eukaryotic cells and the replacing of the values of the inside pixels until all instances of eukaryotic cells are removed.

[0194] Clause 4. The method of clause 2 or 3, wherein replacing the values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box comprises: for each inside pixel: selecting a predetermined number of outside pixels, comparing each of the predetermined number of outside pixels to pixels at comers of the bounding box to identify an outside pixel of the predetermined number of outside pixels that is most similar to the pixels at the comers of the bounding box, and replacing that inside pixel with the identified outside pixel.

[0195] Clause 5. The method of clause 4, wherein the predetermined number of outside pixels are selected randomly.

[0196] Clause 6. The method of clause 4, wherein comparing each of the predetermined number of outside pixels to pixels at comers of the bounding box to identify the outside pixel of the predetermined number of outside pixels that is most similar to the pixels at the comers of the bounding box comprises evaluating similarity using a mean Euclidean distance.

[0197] Clause 7. The method of any of clauses 2-6, wherein generating the background image from the image of eukaryotic cells further comprises: creating a masking image, wherein the masking image is an image with a black-colored background having a smaller, white area in a middle, wherein the masking image is a same size as the bounding box; applying a Gaussian filter to the masking image to generate a filtered masking image; inserting the filtered masking image in a same position as the bounding box; and smoothing an area of the image of eukaryotic cells where the at least one instance of eukaryotic cells was removed using the filtered masking image.

[0198] Clause 8. The method of any preceding clause, wherein generating the artificial image of eukaryotic cells using the augmented cells and the background image comprises: placing a selection of the augmented cells on the background image. [0199] Clause 9. The method of clause 8, wherein the selection of the augmented cells mimics a distribution of the eukaryotic cells in the image of eukaryotic cells.

[0200] Clause 10. The method of clause 8, wherein the selected augmented cells are placed randomly on the background image.

[0201] Clause 11. The method of clause 8, wherein placing the selection of the augmented cells on the background image comprises: generating a bounding box around at least one cell from the selection of the augmented cells; comparing pixels from four comers of the bounding box to pixels in the background image to determine a target location, wherein the target location is an area of the background image where the pixels in the background image have a highest degree of similarity to the pixels in the four comers of the bounding box; placing the at least one of the selection of the augmented single-cells and the augmented multi-cells in the target location; creating a cell masking image, wherein the cell masking image includes a black color background and a white area in the middle, wherein the white area is substantially the same shape as the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; and applying a Gaussian filter to smooth a transition between texture of the background image and texture of the at least one cell from the selection of the augmented single-cells and the augmented multi-cells.

[0202] Clause 12. The method of clause 11, wherein comparing pixels from four comers of the bounding box to pixels in the background image to determine the target location comprises using a mean Euclidean distance for determining similarity.

[0203] Clause 13. The method of clause 1 1 or 12, wherein the bounding box further comprises a height.

[0204] Clause 14. The method of any preceding clause, wherein the at least one augmentation technique applied to each cell corresponds to a particular augmentation policy of a pre-determined selection of a set of augmentation techniques.

[0205] Clause 15. The method of clause 14, wherein augmentation techniques of the pre-determined selection of the set of augmentation techniques are selected from FlipLR, FlipUD, AutoContrast, Equalize, Rotate, Posterize, Contrast, Brightness, Sharpness, Smooth, and Resize.

[0206] Clause 16. A method of generating artificial images for medical evaluation of eukaryotic cells, comprising: extracting each instance of single-cells in an image of eukaryotic cells; extracting each instance of multi-cells in the image of eukaryotic cells; selecting a set of cells from the extracted single-cells and the extracted multi-cells; applying at least one augmentation technique to each cell in the set of cells to generate augmented cells; and generating an artificial image of eukaryotic cells using the augmented cells.

[0207] Clause 17. The method of clause 16, wherein extracting each instance of single- cells includes identifying each instance of single-cells in the image of eukaryotic cells, labelling each instance of single-cells as a single-cell, and storing each instance of single-cells in a single-cell resource.

[0208] Clause 18. The method of clause 16, wherein extracting each instance of multi- cells includes identifying each instance of multi-cell groupings in the image of eukaryotic cells, labelling each instance of multi-cells as a multi-cell, and storing each instance of multi-cells in a multi-cell resource.

[0209] Clause 19. The method of clause 16, wherein selecting the set of cells from the extracted single-cells and the extracted multi-cells comprises randomly selecting the cells.

[0210] Clause 20. The method of clause 16, wherein selecting the set of cells from the extracted single-cells and the extracted multi-cells comprises mimicking a distribution of the eukaryotic cells in the original image of eukaryotic cells.

[0211] Clause 21. The method of clause 16, wherein the at least one augmentation technique applied to each cell corresponds to a particular augmentation policy of a pre- determined selection of a set of augmentation techniques.

[0212] Clause 22. The method of clause 21, further comprising identifying the particular augmentation policy used for the at least one augmentation technique applied to each cell, wherein identifying the particular augmentation policy used for the at least one augmentation technique applied to each cell comprises: applying potential policies of combinations of augmentation techniques to a cell; scoring results of the application of the potential policies; ranking the scored results; and selecting a highest ranking one or more potential policies as the particular augmentation policy.

[0213] Clause 23. The method of clause 21, wherein the augmentation techniques of the pre-determined selection of the set of augmentation techniques are selected from FlipLR, FlipUD, AutoContrast, Equalize, Rotate, Posterize, Contrast, Brightness, Sharpness, Smooth, and Resize.

[0214] Clause 24. The method of clause 16, wherein generating the artificial image of eukaryotic cells using the augmented cells comprises placing a selection of the augmented cells on a background image.

[0215] Clause 25. The method of clause 24, wherein the selected augmented cells are placed randomly on the background image. [0216] Clause 26. The method of clause 24, further comprising generating the background image from the image of eukaryotic cells.

[0217] Clause 27. The method of clause 26, wherein generating the background image from the image of eukaryotic cells comprises generating a bounding box around at least one instance of a eukaryotic cell in the image of eukaryotic cells, wherein pixels of the image of eukaryotic cells inside the bounding box are inside pixels; and replacing values of the inside pixels with values corresponding to certain pixels found outside of the bounding box that are similar in values as bordering pixels of the bounding box to remove the at least one instance of the eukaryotic cell in the image of eukaryotic cells, wherein the certain pixels found outside of the bounding box are outside pixels.

[0218] Clause 28. The method of clause 27, wherein creating the background image further comprises: creating a masking image, wherein the masking image is an image with a black-colored background having a smaller, white area in the middle, wherein the masking image is the same size as the bounding box; applying a Gaussian filter to the masking image to generate a filtered masking image; inserting the filtered masking image in the same position as the bounding box; and smoothing an area of the image of eukaryotic cells where the at least one instance of eukaryotic cells was removed.

[0219] Clause 29. The method of clause 26, further comprising placing at least one cell from the selection of the augmented single-cells and the augmented multi-cells onto the background image via the method comprising: generating a bounding box around the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; comparing pixels from four comers of the bounding box to pixels in the background image to determine a target location, wherein the target location is an area of the background image where the pixels in the background image have the highest degree of similarity to the pixels in the four comers of the bounding box; placing the at least one of the selection of the augmented single-cells and the augmented multi-cells in the target location; creating a cell masking image, wherein the cell masking image includes a black color background and a white area in the middle, wherein the white area is substantially the same shape as the at least one cell from the selection of the augmented single-cells and the augmented multi-cells; and applying a Gaussian filter to smooth the transition between texture of the background image and texture of the at least one cell from the selection of the augmented single-cells and the augmented multi-cells.

[0220] Clause 30. The method of clause 29, wherein the bounding box further comprises a height. [0221] Although the subject matter has been described in language specific to structural features and/or acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as examples of implementing the claims, and other equivalent features and acts are intended to be within the scope of the claims.