Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
COPY DETECTION PATTERN
Document Type and Number:
WIPO Patent Application WO/2022/219049
Kind Code:
A9
Abstract:
Method of generating a copy detection pattern (CDP), or a portion of a CDP, for printing on a substrate, comprising: - generating a plurality of digital files, each of an image comprising an at least partially random two dimensional (2D) distribution of dark and light pixels, and - applying an optimization process configured to increase a copy detection performance of a resultant digital file output by the method, said optimization process comprising a) comparing a copy detection performance of a first of said plurality of digital files, which constitutes a test digital file, with a copy detection performance of another of said plurality of digital files modified with respect to said test digital file, which constitutes a modified digital file, b) replacing the test digital file with said modified digital file, if the copy detection performance of said modified digital file is greater than the copy detection performance of the test digital file, in which case the modified digital file becomes the test digital file, and c) repeating steps a) and b) until a termination condition is met.

Inventors:
PICARD JUSTIN (CH)
Application Number:
PCT/EP2022/059889
Publication Date:
April 06, 2023
Filing Date:
April 13, 2022
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
SCANTRUST B V (NL)
International Classes:
H04N1/32; G03G21/04; G06K19/06; H04N1/00
Attorney, Agent or Firm:
REUTELER & CIE SA (CH)
Download PDF:
Claims:
CLAIMS

1. Method of generating a copy detection pattern (CDP), or a portion of a CDP, for printing on a substrate, comprising:

- generating a plurality of digital files, each of an image comprising an at least partially random two dimensional (2D) distribution of dark and light pixels, and

- applying an optimization process configured to increase a copy detection performance of a resultant digital file output by the method, said optimization process comprising a) comparing a copy detection performance of a first of said plurality of digital files, which constitutes a test digital file, with a copy detection performance of another of said plurality of digital files modified with respect to said test digital file, which constitutes a modified digital file, b) replacing the test digital file with said modified digital file, if the copy detection performance of said modified digital file is greater than the copy detection performance of the test digital file, in which case the modified digital file becomes the test digital file, and c) repeating steps a) and b) until a termination condition is met, d) after step c), forming said CDP or said portion of CDP, from the test digital file, wherein said step of comparing a copy detection performance includes, for each of said test and modified digital files:

- applying a first print simulation filter configured to generate a first print image file simulating an original print of said image on a substrate,

- applying a second print simulation filter to said first print image file to generate a second print image file simulating a printed copy of said original print, and computing a copy detection performance for the digital file indicative of a measure of the differences between the printed copy and the original print.

2. Method according to the preceding claim, wherein computing a copy detection performance includes computing a score for each of the first and second print image files representative of a difference between the digital file and each of said first and second print image files.

3. Method according to the preceding claim wherein computing the score comprises measuring a Pearson correlation coefficient.

4. Method according to any preceding claim wherein said first print simulation filter is a gaussian filter.

5. Method according to any preceding claim wherein said second print simulation filter is identical to said first print simulation filter.

6. Method according to any preceding claim 1-4 wherein said second print simulation filter is different from said first print simulation filter

7. Method according to the preceding claim wherein said second print simulation filter includes a Gaussian unsharp filter.

8. Method according to any preceding claim 1-3 wherein the first and said second print simulation filters are neural networks.

9. Method according to any preceding claim wherein said step of generating a plurality of digital files, includes modifying a current one of said test digital files by changing one or more pixels from dark to light or light to dark to form a current one of said modified digital files.

10. Method according to the preceding claim wherein modifying said digital file by changing one or more pixels from dark to light or light to dark is performed randomly.

11 . Method according to any of the two directly preceding claims wherein modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP of not more than 10%, preferably not more than 5%, preferably between 0,5 and 2 %.

12. Method according to the preceding claim wherein the generation of a plurality of digital files is iterative.

13. Method according to any preceding claim wherein the dark pixels are black pixels and/or the light pixels are white pixels.

14. Method according to any preceding claim wherein said at least partially random two dimensional (2D) distribution of dark and light pixels are generated using a cryptographic key, and optionally the cryptographic key is randomly generated.

15. Method of generating a copy detection pattern (CDP), comprising generating a plurality of portions of a CDP, each using the method according to any preceding claim, and assembling said plurality of portions to form said CDP.

16. Method of generating a copy detection pattern (CDP), or a portion of a CDP, for printing on a substrate, comprising: a) generating a test digital file of an image comprising an at least partially random two dimensional (2D) distribution of dark and light pixels, b) applying a first print simulation filter to said digital file configured to generate a first print image file simulating an original print of said image on a substrate, c) applying a second print simulation filter to said first print image file to generate a second print image file simulating a printed copy of said original print, d) computing a copy detection performance for said digital file indicative of a measure of the differences between the printed copy and the original print, e) creating a modified digital file by modifying said test digital file by changing one or more pixels from dark to light or light to dark, f) applying steps b) to d) on the modified digital file, g) comparing the copy detection performance of the test digital file with the copy detection performance of the modified digital file, h) if the copy detection performance of the modified digital file is indicative of a greater difference between the printed copy and the original print than the copy detection performance of the test digital file, replacing the test digital file with the modified digital file, which now becomes the test digital file, i) if the copy detection performance of the modified digital file is indicative of a smaller difference between the printed copy and the original print than the copy detection performance of the digital file, repeating steps e) to h), j) repeating steps b) to i) until a termination condition is met, k) after step j), forming said CDP or said portion of CDP, from the test digital file.

17. Method according to the preceding claim wherein said second print simulation filter is identical to said first print simulation filter.

18. Method according to claim 16 wherein said second print simulation filter is different from said first print simulation filter.

19. Method according to the preceding claim wherein said second print simulation filter includes a Gaussian unsharp filter.

20. Method according to any preceding claim 16-19 wherein said first print simulation filter is a gaussian filter.

21 . Method according to claim 16 wherein the first and said second print simulation filters are neural networks.

22. Method according to any preceding claim 16-21 , wherein computing a copy detection performance includes computing a score for each of the first and second print image files representative of a difference between the digital file and each of said first and second print image files.

23. Method according to the preceding claim 16-22 wherein computing the score comprises measuring a Pearson correlation coefficient.

24. Method according to any preceding claim 16-23 wherein said first print simulation filter is a gaussian filter.

25. Method according to any preceding claim 16-24 wherein modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP of not more than 10%, preferably not more than 5%.

26. Method according to any preceding claim 16-25 wherein modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP between 0,2% and 2 %.

27. Method according to either of the two directly preceding claims wherein said percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP is decreased for later iterations.

28. Method according to any preceding claim wherein the dark pixels are black pixels and/or the light pixels are white pixels.

29. Method according to any preceding claim wherein said at least partially random two dimensional (2D) distribution of dark and light pixels are generated using a cryptographic key, and optionally the cryptographic key is randomly generated.

30. Method of generating a copy detection pattern (CDP), comprising generating a plurality of portions of a CDP, each using the method according to any preceding claim, and assembling said plurality of portions to form said CDP.

Description:
COPY DETECTION PATTERN

Field of the invention

The present invention relates to a copy detection pattern (CDP) for printing on an article or substrate, and a method of generating a CDP.

Background

A copy detection pattern (CDP), secure graphic (SG) or graphical code is a small random or pseudo-random digital image which is printed on documents, labels or products for counterfeit detection. Authentication is made by scanning the printed CDP using an image scanner or mobile phone camera.

In 1 , it is explained that the detection of counterfeits using a CDP relies on an "information loss principle", which states that every time a digital image is printed or scanned, some information is lost about the original digital image. A CDP is defined as a “maximum entropy image” that attempts to take advantage of this information loss, with counterfeits containing less information than original prints on the digital image of the CDP. Indeed, compared to producing an original print of a CDP, producing a counterfeit CDP requires one additional scanning and printing process and will have thus have less accurate information than an original CDP. By measuring the information in the scanned CDP, the detector can determine whether the CDP is an original print or a copy.

CDPs aim to address limitations of optical security features such as security holograms for detecting counterfeits. They are motivated by the need for security features that can be originated, managed and transferred digitally, and that are machine readable. Contrary to many traditional security printing techniques which are not secure anymore once the knowledge and equipment to produce them is disseminated, the algorithm for generating CDPs can be public as long as the key or image used to generate it or the digital CDP is not revealed. If a key is compromised, CDPs generated with other keys by the same algorithm maintain their security.

The CDP can be viewed as an image which is designed with the purpose to be as sensitive as possible to printing and copying, such that it is as difficult as possible to copy successfully by a

1 https://www.spiedigitallibrary.org/conference-proceedings-of -spie/5310/1 ZDigital-authentication-with- copy-detection-patterns/10.1117/12.528055.short?SSO=1 counterfeiter. However when CDPs were first introduced in 2002 2 , no method was provided to generate them in a way that would ensure their sensitivity to copy attempts and ability to detect counterfeits would be maximized. The patent only proposed to generate a random image, without consideration for the printer resolution, or whether the image should be grayscale or binary (for example offset printers only print binary images, and use halftoning for grayscale images). Therefore it was not revealed how to generate CDPs that would in effect be optimized for discerning originals from counterfeits.

It was observed later on 3 that as most printers actually print binary images at a single, native resolution, it is preferable to generate CDPs as random or pseudo-random binary images at the native printer resolution. As the maximum entropy of a digital binary image with N pixels is N bits, and is reached if the probability of a black versus a white pixel is 50%. However, the entropy of (or amount of information contained in) the digital CDP is typically significantly reduced by printing the CDP at the native printer resolution, due to the naturally occurring information loss caused by dot gain or ink smearing. It is even possible in some situations, for example If dot gain is too high, that the printed CDP becomes flat grey. In such cases, the entropy can become null after printing, as there is no remaining information from the digital CDP. Of course, such a CDP that has lost all its information after the first print, would lose its capacity to discern original prints from copies. In conclusion, simply maximizing the entropy of the digital image of the CDP is in general not the best way to optimize their capacity to discern original prints from counterfeits.

Attempts have been made to develop theoretical models from which the parameter values that optimize the CDP’s ability to differentiate originals from copies can be inferred. The problem can be modeled in a decision-theoretic way, with assumptions on the input signal, the attack channel (how the counterfeit is made), and the print channel. For example in 4 , the signal input (CDP) is binary and the print channel output is also binary, with a probability p that a pixel flips during printing and during copying. This probability describes the noise of the print channel. Based on this model, It was found that the detection performance is maximized for a probability of approx. 19.1%. In 5 and patent application W02009004172A2, the signal input is also binary and equiprobable, except the print channel is modeled as additive white Gaussian noise described https://www.researchgate.net/publication/341030602_Copy_Dete ctable_lmages_From_Theory_to_Practi ce

5 https://www.researchgate.net/publication/341030510_On_the_Se curity_of_Copy_Detectable_lmages with one parameter representative of the noise level of the print channel. The counterfeiter must take a binary decision on the binary value of each signal element. Assuming that the counterfeiter makes an optimal decision in the sense of choosing the signal values which minimize errors in the copy, It is found that the detection performance is optimal if the SNR is approximately 0.562.

These models provide insight into what makes a CDP sensitive to copy (e.g. a certain amount of noise or information loss in the channel is required) but they are misleading as well. One issue is that to apply the model requires to have a print channel with a level of noise which is ideally close to the optimum. In practice, the producer of documents or packaging does not get to select a printing channel: this channel is in most cases a given, based on the specifications of the printer, ink and paper. But even if such specifications could be controlled, there is no reason that the counterfeiter would have used the same parameters. In particular, if the specifications lead to printing with lower quality than what could be achieved with standard printing operations, then the counterfeiter would simply use a better print quality than was used for the original, and increase his chances to have a successful copy.

The models do not indicate how to model the source signal (generate the digital CDP) to approach this signal to noise ratio or error rate. In fact, it assumes binary equiprobable values (i.e. of black and white pixels). It is proposed in W02009004172A2 to approach the target SNR by adjusting the cell size (2x2 black pixels, 3x3 black pixels, etc), and use two sizes such that the printing noise will make lead to errors if the counterfeiter tries to determine what dot structure was used.

These models of printing lead to mistaken conclusions on how to design the CDP. First, they assume that printing has the same effect to black (printed) pixels and white (not printed) pixels. In reality, printing is fairly asymmetric, as a black pixel will rarely become white after printing, while due to dot gain and ink smearing a white pixel will much more frequently become black.

Secondly, they assume that the printing channel can be correctly modeled by a noisy channel. In reality, the noise part in printing a CDP is relatively very small, compared to the mostly deterministic effects of dot gain and ink smearing, which can be better modeled by a combination of low-pass filters and eventually, thresholding.

Thirdly, the application of the model to derive more effective generation of CDPs is suboptimal. In US8593696B2, to target an optimal channel noise factor, it is proposed to use cell sizes of printed black pixels, for example 2x2 or 3x3 pixels. Due to the printing variations, there is uncertainty in determining whether a 2x2 or 3x3 cell was printed. It is noted that for lower resolution printing, e.g. 1200 pixels and under, such a method does not work as using cell sizes of more than 1 pixels will lead to negligible information loss, and therefore near perfect copying will be possible.

In most of the prior art, and in much of the research on the field, binary equiprobable signals are used, meaning on average there are 50% black pixels, or 50% of black cells. In US9594993B2, it is proposed to use a density which differs significantly from 50%, for example 40% of black pixels and 60% of white pixels. The underlying objective is to maximize the information density in original prints and thereby to increase the degradation rate when reproduced, i.e. to maximize the information loss during reproduction of original 2D barcode. While this addresses some of the shortcomings mentioned previously, nothing in the method proposed gives proof, or at least confidence, that a CDP that is maximally effective at detecting copying is generated.

Finally, none of the prior art takes into account the impact that the scanning device may have. Indeed, it is possible that a certain CDP with very fine structure may be very effective if an optimal scanning device (for example a microscope) is used. However, the same CDP could be completely ineffective if scanned with a poor quality scanning device, such as a lower cost mobile camera, as the fine structure might be completely erased in the scan (which probably look like flat gray). For such a device, a more coarse structure, which might be able to discern photocopies, could be better suited. However, the prior art does not propose any method for determining such a structure.

Photocopies and even copies made from a high-resolution scan and using the same printer are generally easy to detect with CDPs. The problem we seek to address here is the detection of superior copies, that are made by applying high-pass filters which allow to pre-compensate the dot gain which occurs during printing of the copy. Copied CDPs can get dangerously close to originals, and while there may remain some statistically significant differences between copied CDPs and original CDPs, after taking into account natural printing and scanning variations, it may become difficult to discern originals from copies with very high reliability.

Besides the use of standard image processing (e.g. high-pass filters, binarization), there is an increasing amount of research on the use of machine learning techniques, and in particular deep learning, to make more accurate copies. For example, in R. Yadav, I. Tkachenko, A Tremeau, and T. Fournel, “Copy Sensitive Graphical Code Estimation: Physical vs Numerical Resolution," in IEEE Workshop on Information Forensics and Security, Delft, Netherlands, December 2019, an efficient estimation attack is developed using an auto-encoder, which is a special type of neural network. While CDPs do not get copied perfectly by such methods, the difference with originals is relatively small, such that it can become problematic to detect such copies in applications where the quality of prints and detection scans varies significantly. The prior art does not provide methods to improve the detection of such sophisticated copies.

Summary of the Invention

An object of the invention is provide a method of generating a CDP for printing on an article, and a CDP generated thereby, that allows very reliable detection of copy attempts.

It is advantageous to provide, a method of generating a CDP for printing on an article, and a CDP generated thereby, that allows very reliable detection of copy attempts for various printing

It is advantageous to provide, a method of generating a CDP for printing on an article, and a CDP generated thereby, that allows very reliable detection of copy attempts using various scanning techniques and various scanners. techniques and various printers.

It is advantageous to provide, a method of generating a CDP for printing on an article, and a CDP generated thereby, that is economical to implement and easy to use.

Objects of the invention have been achieved by providing a method according to claim 1. Dependent claims set forth some of the features of advantageous embodiments.

Disclosed herein is a method of generating a copy detection pattern (CDP), or a portion of a CDP, for printing on a substrate, including:

- generating a plurality of digital files, each of an image comprising an at least partially random two dimensional (2D) distribution of dark and light pixels, and

- applying an optimization process configured to increase a copy detection performance of a resultant digital file output by the method.

The optimization process comprises: a) comparing a copy detection performance of a first of said plurality of digital files, which constitutes a test digital file, with a copy detection performance of another of said plurality of digital files modified with respect to said test digital file, which constitutes a modified digital file, b) replacing the test digital file with said modified digital file, if the copy detection performance of said modified digital file is greater than the copy detection performance of the test digital file, in which case the modified digital file becomes the test digital file, and repeating steps a) and b) until a termination condition is met. The digital file with the highest copy detection performance may be retained as resultant digital file output by the method. This may in particular be the last test or modified digital file on termination of process, if an iterative process is employed.

The step of comparing a copy detection performance includes, for each of said test and modified digital files:

- applying a first print simulation filter configured to generate a first print image file simulating an original print of said image on a substrate,

- applying a second print simulation filter to said first print image file to generate a second print image file simulating a printed copy of said original print, and

- computing a copy detection performance for the digital file indicative of a measure of the differences between the printed copy and the original print.

In an advantageous embodiment, said computing a copy detection performance includes computing a score for each of the first and second print image files representative of a difference between the digital file and each of said first and second print image files.

In an advantageous embodiment, computing the score comprises measuring a Pearson correlation coefficient.

In an advantageous embodiment, the first print simulation filter is a gaussian filter.

In an embodiment, said second print simulation filter is identical to said first print simulation filter. In another embodiment, said second print simulation filter is different from said first print simulation filter.

In an advantageous embodiment, the second print simulation filter includes a Gaussian unsharp filter.

In an advantageous embodiment, the step of generating a plurality of digital files, includes modifying a current one of said test digital files by changing one or more pixels from dark to light or light to dark to form a current one of said modified digital files.

In an advantageous embodiment, the method comprises modifying said digital file by changing one or more pixels from dark to light or light to dark is performed randomly. In an advantageous embodiment, the modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP of not more than 10%, preferably not more than 5%, preferably between 0,5 and 2 %.

In an advantageous embodiment, the generation of a plurality of digital files is iterative.

In an advantageous embodiment, the dark pixels are black pixels.

In an advantageous embodiment, the light pixels are white pixels.

In an advantageous embodiment, the at least partially random two dimensional (2D) distribution of dark and light pixels are generated using a cryptographic key.

In an advantageous embodiment, the cryptographic key may be randomly generated.

In an advantageous embodiment, the method comprises generating a plurality of portions of a CDP, each using the method described above, and assembling said plurality of portions to form said CDP.

A termination condition may be when the score difference is higher than a preset value, or a certain number of iterations, or a rate of change of the detection performance is below a certain threshold, or after a set computation time.

Also disclosed herein, according to a preferred aspect of the invention, is a method of generating a copy detection pattern (CDP), or a portion of a CDP, for printing on a substrate, comprises: a) generating a test digital file of an image comprising an at least partially random two dimensional (2D) distribution of dark and light pixels, b) applying a first print simulation filter to said digital file configured to generate a first print image file simulating an original print of said image on a substrate, c) applying a second print simulation filter to said first print image file to generate a second print image file simulating a printed copy of said original print, d) computing a copy detection performance for said digital file indicative of a measure of the differences between the printed copy and the original print, e) creating a modified digital file by modifying said test digital file by changing one or more pixels from dark to light or light to dark, f) applying steps b) to d) on the modified digital file, g) comparing the copy detection performance of the test digital file with the copy detection performance of the modified digital file, h) if the copy detection performance of the modified digital file is indicative of a greater difference between the printed copy and the original print than the copy detection performance of the test digital file, replacing the test digital file with the modified digital file, which now becomes the test digital file, i) if the copy detection performance of the modified digital file is indicative of a smaller difference between the printed copy and the original print than the copy detection performance of the digital file, repeating steps e) to h), j) repeating steps b) to i) until a termination condition is met, k) after step j), forming said CDP or said portion of CDP, from the test digital file.

The above preferred aspect may further comprise additional features of the advantageous embodiments mentioned above.

In an advantageous embodiment, said second print simulation filter is identical to said first print simulation filter.

In an advantageous embodiment, said second print simulation filter is different from said first print simulation filter.

In an advantageous embodiment, said second print simulation filter includes a Gaussian unsharp filter.

In an advantageous embodiment, the step of computing a copy detection performance includes computing a score for each of the first and second print image files representative of a difference between the digital file and each of said first and second print image files.

In an advantageous embodiment, the step of computing the score comprises measuring a Pearson correlation coefficient.

In an advantageous embodiment, said first print simulation filter is a gaussian filter.

In an advantageous embodiment, the step of modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP of not more than 10%, preferably not more than 5%.

In an advantageous embodiment, the step of modifying said digital file by changing one or more pixels from dark to light or light to dark is performed on a percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP between 0,2% and 2 %.

In an advantageous embodiment, said percentage of pixels relative to a total amount of pixels forming said CDP or said portion of CDP is decreased for later iterations.

In an advantageous embodiment, the dark pixels are black pixels and/or the light pixels are white pixels.

In an advantageous embodiment, said at least partially random two dimensional (2D) distribution of dark and light pixels are generated using a cryptographic key, and optionally the cryptographic key is randomly generated.

Also disclosed herein is a method of generating a copy detection pattern (CDP), comprising generating a plurality of portions of a CDP, each using the method according to any preceding aspect or embodiment, and assembling said plurality of portions to form said CDP.

Embodiments of the invention also include a CDP generated according to any of the above methods and a product bearing a printed CDP generated according to any of the above methods.

In an advantageous embodiment, the first and said second print simulation filters may be neural networks.

Conventional CDPs are randomly generated, often pseudo-randomly from a cryptographic key which is itself randomly generated. The inventor has realized however that not are all secure graphics generated randomly have an essentially equivalent capability in separating originals from copies, but that some CDPs may be more effective than others.

The inventor has in particular observed that when printing a binary CDP, there are regions that tend to become mostly black and regions that tend to become mostly white. Figure 1 contains a 50x50 pixels digital CDP and the scan of a 600ppi print of that CDP (scanned at high resolution). Area 1 shows one of many areas that become mostly black after printing, while zone 2 shows an area which becomes mostly white. As can be observed in the digital CDP, the density of black pixels is a bit higher than average in zone 1 , while the density of white pixels is a bit higher than usual in zone 2. Of course, the local density of black versus white pixels varies naturally for CDPs with pixels that are determined pseudo-randomly with a certain distribution.

Regions which become mostly black or mostly white have actually lost their capacity to discriminate between originals and copies, as they will most likely no longer be modified when producing the copy: a white region in the original print will remain white, and a black region will remain black. On the other hand, transition regions from black to white (or vice versa), and regions where there is a balance of black and white pixels in the original prints, typically have more chances to get altered during the copying process. Therefore it is preferable to generate CDPs such that original prints of these CDPs have as many transition regions as possible.

Another observation from the inventor is that the loss of information through printing is only partially due to random effects, and the more significant effect appears to be a deterministic transformation of the CDP. This transformation does not vary significantly for two prints of the same CDP, as can be seen on Figure 2 which shows two scans of two different printed CDPs of the same digital CDP (which was printed at 812ppi on a HP Indigo printer). It can be observed that the two images are nearly similar. For illustrative purposes, one of the small differences is indicated, but it basically highlights that the two prints are essentially the same. The key takeaway is that making the assumption that CDPs are secure against copy on the premise that printing is a noisy process is not exactly true, as the print transformation is mostly deterministic. Indeed, printing an original CDP, or printing a copy from a scan of an original CDP, is more akin to a deterministic transformation than to the insertion of noise. Consequently, using a printing model where printing is described as a noise process, as has been done in the prior art, will lead to misleading conclusions on how CDPs should be designed to be optimally robust against copy.

One feature of this invention is to use mathematical models of the printing process and of the copying process, to search for optimized CDP structures in the CDP generation process. The process of finding an optimized CDP is initialized with a non-optimized CDP, which will be then optimized using methods described below. Based on the observations and discussion above, examples of the modelling of the detection performance, generation, printing, and copying according to embodiments of the invention are described below.

Brief Description of the drawings

Figure 1 illustrates a 50x50 pixels digital CDP and a scan of a 600ppi print of that CDP (scanned at high resolution);

Figure 2 shows two scans of two different printed CDPs of the same digital CDP; Figure 3a shows an example of a digital CDF which is a 50x50 pixels large;

Figure 3b shows the result of applying a Gaussian filter (with o = 1) to the digital CDF of Figure 3a;

Figure 3c shows a simulated optimized copy of the printed CDP of figure 3a;

Figure3d shows a simulated normal copy of the printed CDP of figure 3a;

Figure 4 illustrates different example algorithms, in python language, to simulate printing, as well as to simulate the copying process;

Figure 4a illustrates an example algorithm to generate an optimal CDP;

Figure 4b illustrates an example algorithm to simulate generation of a large number of original CDPs, copied CDPs, and measure the score difference that can be obtained;

Figure 4c illustrates an example algorithm to find optimal parameters of the unsharp filter used by the counterfeiter to prepare the copy;

Figure 5 shows a plot of the evolution of score difference when applying the algorithms of figure 4 to a previously generated CDP;

Figure 6 shows an initial CDP, and two optimized CDPs obtained by a process according to an embodiment of the invention;

Figure 7 shows a tiled optimized digital CDP, and the original prints and copies;

Figure 8a shows a plot of the Pearson correlation coefficient versus the op (sigma print) of the Gaussian blur filter used to simulate printing of the original CDP;

Figure 8b shows a plot of the op (sigma print) value corresponding to a given correlation coefficient, as calculated by interpolation;

Figure 9 illustrates another example algorithm to optimize a CDP;

Figure 10 shows an example of an initial digital image of a CDP (top), and of an optimized modified digital image of a CDP (bottom) obtained after 20’000 iterations using a method according to an embodiment of the invention;

Figure 11a illustrates a simplified flowchart of a method of generating a copy detection pattern according to an embodiment of the invention;

11

RECTIFIED SHEET (RULE 91) ISA/EP Figure 11 b illustrates a simplified flowchart of how detection performance is computed according to an embodiment of the invention;

Figure 12 illustrates an example of lookup table which allows to determine, for a real printer, the corresponding op (sigma) value of the Gaussian filter which allows to simulate the printing of an original CDP; this sigma value corresponds to a given average score, or Pearson correlation coefficient, on a set of scans of printed original CDP with this printer.

Detailed Description of embodiments

Detection performance

Given two images of identical size represented by n pixel values with x, and being the individual pixel values points indexed with / for each respective image, r xy is the Pearson correlation coefficient which has a value between -1 and +1 :

To assess detection performance, one can measure the Pearson correlation coefficient between the original and digital CDP, and between the copied CDP and digital CDP, to thus obtain an original score and a copy score. The score difference between the original score and the copy score may be used as a detection performance indicator: the larger it is, the better the CDP will potentially be at discerning originals from copies.

Figure 11 b illustrates how detection performance is calculated with a flowchart.

There are several methods of determining the score according to embodiments of the invention. One comprises measuring the bit error rate (BER) between the test CDP and the digital CDP. To determine the BER, the test CDP, which is typical a grayscale image with integer values between 0 and 255, may be binarized using a threshold. There are multiple thresholding methods. Otsu’s method is frequently used; another method comprises trying all integer threshold values and calculating the BER for each, then the smallest value found for the BER is returned.

It is also noted that multiple types of copies can be generated, using different algorithms, and the detection performance is calculated based on a combination of the individual scores for the different copies: for example the average, the maximum, or the median score of these different copies. When the images are actual scans, a step of image registration should be done first, in which the scan is spatially transformed to align with the target image, which is the digital CDP. The transformation is required to achieve correspondence between the pixels values of the (transformed) scanned image and the digital CDP. However, this step is generally not needed when working with simulated originals and copies.

CDP generation

One can generate pseudo-randomly an initial binary SG with 50% black pixel probability. Figure 3 (a) shows one example digital CDP which is 50x50 pixels large.

Other percentage values may be used, as well as different types of structures: for example 2x2, 3x3, 4x4 or larger structures can be used if the print resolution is higher than 600ppi. Asymmetric structures (e.g. 2x3 pixels) may also be used.

It is also possible to start with a non-random digital CDP, such as a grid with evenly spaced black pixel structures.

CDP printing model

One may simulate the printing process in a deterministic way by using a Gaussian blur filter. Optionally, the filtering can be followed by thresholding.

Below is an example Gaussian filter kernel (with the standard deviation parameter o = 0.84089642), taken from the Wikipedia page on https://en.wikipedia.org/wiki/Gaussian blur . The center element (at [4, 4]) has the largest value, decreasing symmetrically as distance from the center increases.

[ 0.00000067 0.00002292 0.00019117 0.00038771 0.00019117 0.00002292 0.00000067 0.00002292 0.00078633 0.00655965 0.01330373 0.00655965 0.00078633 0.00002292

0.00019117 0.00655965 0.05472157 0.11098164 0.05472157 0.00655965 0.00019117

0.00038771 0.01330373 0.11098164 0.22508352 0.11098164 0.01330373 0.00038771

0.00019117 0.00655965 0.05472157 0.11098164 0.05472157 0.00655965 0.00019117

0.00002292 0.00078633 0.00655965 0.01330373 0.00655965 0.00078633 0.00002292

0.00000067 0.00002292 0.00019117 0.00038771 0.00019117 0.00002292 0.00000067 ]

Figure 3b shows the result of applying such a Gaussian filter (with o = 1) to the digital CDP of Figure 3a. Other more complex functions, for example a neural network trained with various scans, may be used to simulate the scanning model.

CDP scanning model

Optionally, it may be advantageous to simulate imperfect scanning conditions for detection. In this case, the simulated original CDPs and simulated copies may be transformed by a model of the scanning process, which can be described as well with a Gaussian blur filter.

Other more complex functions, for example a neural network trained with various scans, may be used to simulate the scanning model.

Copying process

Photocopies, and even copies made from a high-resolution scan and using the same printer are generally easy to detect with CDPs, as long as the counterfeiter applies only basic image processing before printing the scan to obtain the copies. One of the problems addressed by the present invention is the detection of superior copies where the counterfeiter processes the scan in a specific, optimized way before printing the copy. Such copies can for example be made by applying high-pass filters with well-tuned parameters, in order to pre-compensate the dot gain which occurs during printing of the copy. It may be done as well using deep learning-based approaches to achieve a similar purpose. Even if there are still some statistically significant differences between original CDPs and these optimized copies of CDPs, it may become difficult to discern originals from these optimized copies with 100% accuracy after taking into account the natural printing and scanning variations which occur in real-world applications.

In the simulations, we may assume the following steps for simulating a normal copy, respectively an optimized copy. The optimized copy differs from the normal copy only in step 1 below :

1. Perfect scan of original CDP: there is no information loss or transformation of the original CDP

2. For each type of copy: a. Normal copy: No image processing b. Optimized copy: Application of a Gaussian unsharp mask filter with near optimal values for optimized copies (more on this below)

3. Image binarization using the image mean as threshold

4. Print copy in the same condition as the originals, which means using the same value of o in the Gaussian blur filter; alternatively, we may use a smaller or a larger o for the copy, to simulate conditions where the copy is printed in better, respectively worst conditions than the originals.

It is known that a Gaussian unsharp mask is an image sharpening technique, which uses a blurred, or "unsharp", negative image to create a mask of the original image. The unsharp mask is then combined with the original positive image, creating an image that is less blurry than the original. The typical blending formula for unsharp masking is

Sharpened image = original + (original image - blurred image) x Amount.

When using a Gaussian filter to determine the blurred image, the value of ct needs to be set, in addition to the value of the parameter “Amount “ in the formula above. This value is different from the o value used to simulate printing. In the following we will denote the parameter ct p as the ct parameter used to simulate the printer, and the parameter ct c as the value used in the unsharp filter by the counterfeiter.

For a given ct p , the counterfeiter can determine by a grid search the values of ct c and Amount which seek to minimize the detection performance (the difference between original and copy score).

It is noted that several other copying models can be used. In particular, copies based on neural networks, also called deep learning may be used. Two such examples of copying models can be found in previously mentioned paper R. Yadav, I. Tkachenko, A. Tremeau, and T. Fournel, “Copy Sensitive Graphical Code Estimation: Physical vs Numerical Resolution, ” in IEEE Workshop on Information Forensics and Security, Delft, Netherlands, December 2019 and: O. Taran, S. Bonev, T. Holotyak, and S. Voloshynovskiy, "Adversarial detection of counterfeited printable graphical codes: towards "adversarial games" in physical world, " in Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020.

In the following example, the print is simulated by a Gaussian filter with o p =1. Figure 3 shows the digital CDP (in (a) ), simulated printed CDP (in (b)), and simulated copies (optimized copy in (c) and normal copy (d)). The original score is 0.5634, while the normal copy score is 0.3820. There is a score difference of 0.1814, which in practice gives a comfortable margin to discern the original from the copy. For the optimized copy, the counterfeiter applies an unsharp filter. The optimal parameters after a grid search are o c =0.543 and amount=43, for a copy score is 0.5092. The score difference is 0.0542 which is considerably smaller. It can be appreciated from Figure 3c that this optimized copy of CDP is visually much more similar to the original CDP, than is the normal copy of Figure 3d. An example algorithm is given in Figure 4c to determine values optimal parameter values.

If one repeats this simulation one thousand times with differently randomly generated CDPs, one will statistically get a similar result: the average score is 0.572 for the original CDP, 0.385 for the normal copy, and 0.507 for the optimized copy. The average score difference is 0.065, and the minimum and maximum score difference over the one thousand simulations are respectively 0.047 and 0.080. While there are some variations in the score difference, the best digital CDP (in the sense of leading to the largest score difference) generated randomly seems only slightly better than the worst CDP.

In practice, considering the natural variability due to print quality variations as well as scan quality variations, this gap may be too small to discern original from copies with a sufficiently high reliability.

It may be noted that similar results are found for different variations of the models that simulate printing (for example by thresholding after applying the Gaussian low pass filter, in the printing model) as well as for the copying process.

Example of an algorithm to determine an optimized CDP according to an embodiment

The aim is to determine whether there are certain CDPs which would be naturally more robust to copying than others. For such “optimized CDPs”, the score difference with good copies should be larger than the largest score difference that we found, which was 0.080. As we observed from the above example simulation repeated 1000 times, such CDPs, if they exist, are unlikely to be found by chance. Example code to make such simulation is shown in Figure 4b.

The following iterative algorithm is an example of an embodiment, which at each iteration tests small modifications of a CDP, and keeps the modifications only if they lead to an “improved CDP” with a higher score difference. At each iteration, a random modification of a few pixels is made on the current digital CDP (flipping white pixels to back or vice-versa), to obtain a test digital CDP. Using the simulations described above to obtain an original CDP and a good copy CDP, the score difference may be computed for both the current CDP and the test CDP. If the score difference is higher for the test CDP, the current CDP becomes the test CDP for the next iteration, otherwise the current CDP remains the same. The algorithm terminates either after a certain number of iterations, when the score difference is higher than a preset value, or when it is unlikely that the test CDP will yield a higher score difference. Other termination conditions may be set (e.g. time, thresholds, etc). The algorithm receives as input the printing model parameter o p , and the copying model parameters o c and Amount. Note that the copying model parameters may not be necessary if some fixed values are used to generate the copy.

Here is the generic algorithm, described as a flowchart in Figure 11 :

• Receive printing and copying parameters.

• Initialize the current digital CDP by generating an initial digital CDP

• Iterate until termination conditions are met: o Randomly flip a given percentage of pixels of current digital CDP, to obtain a test digital CDP. o Calculate detection performance of current (i.e. modified) digital CDP as in Figure 11 b o Calculate detection performance of test digital CDP as in Figure 11 b o If detection performance of test digital CDP is superior to detection performance of current digital CDP, replace current digital CDP with test digital CDP

One implementation of this iterative algorithm in python code is provided as an example in Figure 4. In this example algorithm, a 50x50 pixels CDP initially generated from a binary equiprobable distribution, is iteratively modified randomly. The algorithm runs over 10’000 iterations (termination condition), using a 0.2% probability that each pixel of the digital CDP gets flipped at each iteration (on average, 5 pixels are flipped per iteration). The printing model uses a Gaussian blur with a a of 1 . The copying process model uses values previously mentioned of o c =0.543 and Amount=43.

Figure 5 shows the evolution of score difference when applying this algorithm to the previously generated CDP. The score difference starts at 0.0542, and increases four times to 0.223 at the 10’000th iteration. For the optimized CDP, the original score is 0.588 and the good copy score is 0.365. According to the simulation, the score difference between an optimized CDP and its optimized copy can be even larger than the score difference between a regular CDP and its optimized copy (which was of 0.1814 in the example above).

It is interesting to note that running this process multiple times from the same initial CDP, we typically obtain very different optimized CDPs: on average 43% of the pixel values differ. This suggests there is a potentially very large number of optimized CDPs. However optimized CDPs are extremely unlikely to be obtained by chance using a conventional CDP generation algorithm of the prior art. Figure 6 shows the initial CDP, and two optimized CDP obtained by the process described above. What is visually striking in these two optimized CDPs is that, contrary to the initial CDP, there are no regions where black pixels or large pixels predominate. Such regions are much more frequent in the initial CDP.

A probability of bit flipping of 0.2% may seem too small. However, if a probability of 1 % was used instead, the score difference after 10’000 iterations would only be of 0.155, as with this percentage it quickly becomes very unlikely to obtain random modifications that improve the score difference. For converging to an optimized CDP in a more efficient manner, the bit flipping probability may be set to a relatively high value initially, e.g. 1%, and then slowly decrease with the number of iterations.

Other termination conditions can be used, for example a certain detection performance, or a certain total number of modifications to the initial CDP.

Testing method with real data

While this algorithm appears very promising when applied to synthetically generated originals and copies, it is useful to verify its effectiveness when confronted to real printing conditions. For this purpose, we have printed both the initial and optimized CDP 100 times, by tiling the same CDP 100 times. Both the 100 initial and optimized CDPs were copied in the same condition (using a 2400dpi scan, followed by application of the unsharp filter (same parameters as before), thresholding and printing on the same printer). Figure 7 shows the tiled optimized digital CDP, and the original print and copy. The scans of the original and copy are made at 1200dpi. All scans are made using a Canon Canoscan 9000F, and the prints are made using a Canon IR-ADV C5535i laser printer.

For the initial CDP, the score of original CDPs is on average 0.4679. The score of copied CDPs is on average 0.3609, and the score difference is 0.1068.

To calculate an optimized CDP using the previously described algorithm, we need to determine the value of o p for the Gaussian low pass filter, which gives a score (correlation coefficient) close to what we obtained for the original CDPs, i.e. 0.4679. To be able to determine the o p value representative of different printers, we have generated a CDP, applied the Gaussian low pass filter with multiple values for o p between 0 and 3. Figure 8a shows the correlation coefficient versus o p . By interpolation, we are then able to determine the op sigma value corresponding to a given correlation coefficient, as shown in Figure 8b. A cubic spline interpolation using Python package SciPy can be used for instance, among many other possibilities, to do the interpolation. In this particular case, we find this value to be approximately 1.24. We note that in practice, well known methods such as a lookup table or various interpolation functions can be used. A lookup table is provided in Figure 12 for correlation coefficient of 0.25 to 0.7. This range of correlation coefficient is representative of printers that are used in practical applications. It is noted that it is not necessary to have a very precise value, as the algorithm will lead to a CDP with improved performance for a range of values, but it is preferable to be within the range that is representative of the printer.

For an optimized CDP in this example, the score of original CDPs is on average 0.4717, the score of copied CDPs is on average 0.2604, and the score difference is 0.1905. While the score difference is not quite as large as with synthetically generated prints and copies, the improvement in differentiation between original and copies remains very significant. This demonstrates that the optimization process of CDPs significantly increases their ability to detect copies. While in this experiment in a benchmarked environment, copies are fully separated from originals for the initial CDP as well, a larger score difference is very important to ensure higher robustness in real world scenarios, where there can be larger score variations due to evolving print conditions, as well as a larger variety of scan quality.

Algorithm for determining printing parameters

The previous algorithm requires a printing parameter o p as input. Here is an example of an algorithm to evaluate this value according to an embodiment:

1. Generate one or a set of digital CDPs (e.g. binary CDP with 50% black pixel probability)

2. Print CDPs on the printer of interest, to obtain set of printed CDPs

3. Scan printed CDPs to obtain a set of scanned CDPs

4. Measure score or each scanned CDP (e.g. correlation coefficient after image alignment with the source CDP) and determine average score.

5. Determine sigma parameter for average score, using the lookup table in Figure 12. The closest value to the average score in column “score” is used to determine the o p printing parameter value in column “sigma”

Alternative method to determine optimized CDP

The process above is relatively computationally intensive, as it does require a number of iterations to converge to an optimized CDP. However its speed can be improved by orders of magnitude in different ways. One very effective way comprises creating, in advance, several small optimized CDPs, using the process described above. For example, ten thousand optimized CDPs of size 10x10 pixels can be created in advance. Then, when generating a new CDP, for example of 50x50 pixels, for each of the twenty-five 10x10 blocks of the CDP, one of the ten thousand 10x10 structures will be pseudo-randomly selected. Once assembled, this CDP can be further optimized through a smaller number of iterations (e.g. 500), as this will help to eliminate possible suboptimal regions at the transition between blocks.

Other algorithms within the scope of the invention can be used to achieve a similar result at a lower computational cost, while preserving the spirit of the method. One such method comprises reducing the number of regions where there is a high proportion of black pixels or a high proportion of white pixels. This can be done by randomly picking regions of, say, 3x3 pixels, and setting the central pixel to white, respectively black, if more than half the pixels around it are black, respectively white. One example algorithm is given in Figure 9. This algorithm can be initialized with imgO being a random binary matrix with approximately 50% black pixel and the number of iterations nb_iterations=20000. Figure 10 shows an example with an initial image (top), and an image obtained after twenty thousand iterations (bottom).

• Until termination condition is met: o Select a random pixel in the CDP o Calculate sum of black pixels in a region around this pixel o If the pixel is white and the sum of black pixels is under a given threshold, set this pixel to white o If the pixel is black and the sum of black pixels is over a certain threshold, set this pixel to white

Additional remarks

It may be noted that the method can work with any printing model. More sophisticated and accurate printing models, for example derived from using neural networks, can be used as well.

In practice, the optimized CDP may be then sent to the printer that will be used to print the original items. It may be printed once on a single item, or an arbitrary number of times on different items. This CDP, information allowing to reconstruct that CDP, or any transformation of that CDP may be stored on the authentication server. Reference scans of the printed CDP may be stored, instead or in addition, on the authentication server.