Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
PRIVACY-ENHANCING TECHNOLOGIES FOR MEDICAL TESTS USING GENOMIC DATA
Document Type and Number:
WIPO Patent Application WO/2014/040964
Kind Code:
A1
Abstract:
In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by a patient's public keys at a Storage and Processing Unit (SPU). A part of the corresponding private key is also stored on the SPU. At the time of the test by a Medical Unit (MU), the patient provides the second part of the private key to the MU. A test with its associated markers is determined by the MU and sent to the SPU. The test is carried out on the encrypted values thanks to homomorphic operation and returned back to the MU. The latter uses the second part of the private key to access the result.

Inventors:
AYDAY ERMAN (CH)
HUBAUX JEAN-PIERRE (CH)
RAISARO JEAN-LOUIS (CH)
TELENTI AMALIO (CH)
FELLEY JACQUES (CH)
MC LAREN J PAUL (CH)
ROUGEMONT JACQUES (CH)
HUMBERT MATHIAS (CH)
Application Number:
PCT/EP2013/068658
Publication Date:
March 20, 2014
Filing Date:
September 10, 2013
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
ECOLE POLYTECH (CH)
International Classes:
G06F19/00; G16B50/40
Other References:
ERMAN AYDAY ET AL: "Privacy-enhancing technologies for medical tests using genomic data", INTERNET CITATION, 28 December 2012 (2012-12-28), pages 16pp, XP009168811, Retrieved from the Internet [retrieved on 20130912]
KANTARCIOGLU M ET AL: "A Cryptographic Approach to Securely Share and Query Genomic Sequences", IEEE TRANSACTIONS ON INFORMATION TECHNOLOGY IN BIOMEDICINE, IEEE SERVICE CENTER, LOS ALAMITOS, CA, US, vol. 12, no. 5, 1 September 2008 (2008-09-01), pages 606 - 617, XP011224108, ISSN: 1089-7771, DOI: 10.1109/TITB.2007.908465
THOMAS NEUBAUER ET AL: "A methodology for the pseudonymization of medical data", INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, ELSEVIER SCIENTIFIC PUBLISHERS, SHANNON, IR, vol. 80, no. 3, 19 October 2010 (2010-10-19), pages 190 - 204, XP028363237, ISSN: 1386-5056, [retrieved on 20101027], DOI: 10.1016/J.IJMEDINF.2010.10.016
DEFTEREOS S ET AL: "Technical Guidelines for Enhancing Privacy and Data Protection in Modern Electronic Medical Environments", IEEE TRANSACTIONS ON INFORMATION TECHNOLOGY IN BIOMEDICINE, IEEE SERVICE CENTER, LOS ALAMITOS, CA, US, vol. 9, no. 3, 1 September 2005 (2005-09-01), pages 413 - 423, XP011138592, ISSN: 1089-7771, DOI: 10.1109/TITB.2005.847498
A. CAVOUKIAN, PRIVACY BY DESIGN, 2009, Retrieved from the Internet
S. F. GURSES: "Multilateral privacy requirements analysis in online social network services", PHD THESIS, 2010
M. LANGHEINRICH: "Principles of privacy-aware ubiquitous systems", PROCEEDINGS OF UBIQUITOUS COMPUTING (UBICOMP, 2001
G. VAN BLARKOM; J. BORKING; J. OLK: "Handbook of privacy and privacy-enhancing technologies (the case of intelligent software agents", COLLEGE BESCHERMING PERSOONSGEGEVENS, 2003
J. R. TRONCOSO-PASTORIZA; S. KATZENBEISSER; M. CELIK: "Privacy preserving error resilient DNA searching through oblivious automata", CCS '07: PROCEEDINGS OF THE 14TH ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2007, pages 519 - 528, XP002495762, DOI: doi:10.1145/1315245.1315309
M. BLANTON; M. ALIASGARI: "Secure outsourcing of DNA searching via finite automata", DBSEC'10: PROCEEDINGS OF THE 24TH ANNUAL IFIP WG 11.3 WORKING CONFERENCE ON DATA AND APPLICATIONS SECURITY AND PRIVACY, 2010, pages 49 - 64, XP047307802, DOI: doi:10.1007/978-3-642-13739-6_4
S. JHA; L. KRUGER; V. SHMATIKOV: "Towards practical privacy for genomic computation", PROCEEDINGS OF THE 2008 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, 2008, pages 216 - 230, XP031259107
F. BRUEKERS; S. KATZENBEISSER; K. KURSAWE; P. TUYLS: "Privacy-preserving matching of DNA profiles", TECH. REP., 2008
M. KANTARCIOGLU; W. JIANG; Y. LIU; B. MALIN: "A cryptographic approach to securely share and query genomic sequences", IEEE TRANSACTIONS ON INFORMATION TECHNOLOGY IN BIOMEDICINE, vol. 12, no. 5, 2008, pages 606 - 617, XP011345491, DOI: doi:10.1109/TITB.2007.908465
P. BALDI; R. BARONIO; E. DE CRISTOFARO; P. GASTI; G. TSUDIK: "Countering GATTACA: efficient and secure testing of fully-sequenced human genomes", CCS '11: PROCEEDINGS OF THE 18TH ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2011, pages 691 - 702
M. CANIM; M. KANTARCIOGLU; B. MALIN: "Secure management of biomedical data with cryptographic hardware", IEEE TRANSACTIONS ON INFORMATION TECHNOLOGY IN BIOMEDICINE, vol. 16, no. 1, 2012, XP011490859, DOI: doi:10.1109/TITB.2011.2171701
D. EPPSTEIN; M. T. GOODRICH; P. BALDI: "Privacy-enhanced methods for comparing compressed DNA sequences", CORR
D. EPPSTEIN; M. T. GOODRICH: "Straggler identification in round-trip data streams via Newton's identities and invertible Bloom filters", IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, vol. 23, no. 2, 2011, pages 297 - 306, XP011390617, DOI: doi:10.1109/TKDE.2010.132
R. WANG; Y. F. LI; X. WANG; H. TANG; X. ZHOU: "Learning your identity and disease from research papers: information leaks in genome wide association study", CCS '09: PROCEEDINGS OF THE 16TH ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2009, pages 534 - 544
B. MALIN; L. SWEENEY: "How (not) to protect genomic data privacy in a distributed network: using trail re-identification to evaluate and design anonymity protection systems", JOURNAL OF BIOMEDICAL INFORMATICS, vol. 37, June 2004 (2004-06-01), pages 179 - 192
N. HOMER; S. SZELINGER; M. REDMAN; D. DUGGAN; W. TEMBE: "Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays", PLOS GENETICS, 4 August 2008 (2008-08-04)
J. GITSCHIER: "Inferential genotyping of Y chromosomes in Latter-Day Saints founders and comparison to Utah samples in the HapMap project", AM. J. HUM. GENET., vol. 84, 2009, pages 251 - 258
X. ZHOU; B. PENG; Y. F. LI; Y. CHEN; H. TANG; X. WANG: "To release or not to release: evaluating information leaks in aggregate human-genome data", ESORICS'11: PROCEEDINGS OF THE 16TH EUROPEAN CONFERENCE ON RESEARCH IN COMPUTER SECURITY, 2011, pages 607 - 627, XP047434861, DOI: doi:10.1007/978-3-642-23822-2_33
S. E. FIENBERG; A. SLAVKOVIC; C. UHLER: "Privacy preserving GWAS data sharing", PROCEEDINGS OF THE IEEE 11TH INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW, December 2011 (2011-12-01)
Y. CHEN; B. PENG; X. WANG; H. TANG: "Large-scale privacy-preserving mapping of human genomic sequences on hybrid clouds", NDSS'12: PROCEEDING OF THE 19TH NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM, 2012
R. WANG; X. WANG; Z. LI; H. TANG; M. K. REITER; Z. DONG: "Privacy-preserving genomic computation through program specialization", PROCEEDINGS OF THE 16TH ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2009, pages 338 - 347, XP058271042, DOI: doi:10.1145/1653662.1653703
R. AGRAWAL; A. EVFIMIEVSKI; R. SRIKANT: "Information sharing across private databases", PROCEEDINGS OF SIGMOD CONFERENCE, 2003
D. DACHMAN-SOLED; T. MALKIN; M. RAYKOVA; M. YUNG: "Efficient robust private set intersection", PROCEEDINGS OF THE 7TH INTERNATIONAL CONFERENCE ON APPLIED CRYPTOGRAPHY AND NETWORK SECURITY, 2009, pages 125 - 142, XP047430691, DOI: doi:10.1007/978-3-642-01957-9_8
S. KATHIRESAN; O. MELANDER; D. ANEVSKI; C. GUIDUCCI; N. BURTT: "Polymorphisms associated with cholesterol and risk of cardiovascular events", THE NEW ENGLAND JOURNAL OF MEDICINE, vol. 358, 2008, pages 1240 - 1249
E. ASHLEY; A. BUTTE; M. WHEELER; R.CHEN; T. KLEIN: "Clinical assessment incorporating a personal genome", THE LANCET, vol. 375, no. 9725, 2010, pages 1525 - 1535, XP027035657, DOI: doi:10.1016/S0140-6736(10)60452-7
S. SESHADRI; A. FITZPATRICK; M.A. LKRAM; A. DESTEFANO; V. GUDNASON; M. BOADA; J. BIS; A. SMITH; M. CARASSQUILLO; J. LAMBERT: "Genome-wide analysis of genetic loci associated with Alzheimer disease", JAMA, vol. 303, 2010, pages 1832 - 1840
D. GREENBAUM; A. SBONER; X. MU; M. GERSTEIN: "Genomics and privacy: Implications of the new reality of closed data for the field", PLOS COMPUTATIONAL BIOLOGY, vol. 7, no. 12, 2011
M. RAYKOVA; H. ZHAO; S. M. BELLOVIN: "Privacy enhanced access control for outsourced data sharing", FINANCIAL CRYPTOGRAPHY AND DATA SECURITY, 2012
M. T. GOODRICH; M. MITZENMACHER: "Privacy-preserving access of outsourced data via oblivious RAM simulation", PROCEEDINGS OF THE 38TH INTERNATIONAL CONFERENCE ON AUTOMATA, LANGUAGES AND PROGRAMMING, 2011, pages 576 - 587, XP047366878, DOI: doi:10.1007/978-3-642-22012-8_46
E. STEFANOV; E. SHI; D. SONG: "Towards practical oblivious RAM", NDSS'12: PROCEEDING OF THE 19TH NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM, 2012
E. BRESSON; D. CATALANO; D. POINTCHEVAL: "A simple public-key cryptosystem with a double trapdoor decryption mechanism and its applications", PROCEEDINGS OF ASIACRYPT 03, LNCS, vol. 2894, 2003, pages 37 - 54
M. PIRRETTI; P. TRAYNOR; P. MCDANIEL; B. WATERS: "ecure attribute-based systems", PROCEEDINGS OF THE 13TH ACM CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2006, pages 99 - 112
G. ATENIESE; K. FU; M. GREEN; S. HOHENBERGER: "Improved proxy re-encryption schemes with applications to secure distributed storage", ACM TRANSACTIONS ON INFORMATION AND SYSTEM SECURITY, vol. 9, February 2006 (2006-02-01), pages 1 - 30, XP001242926, DOI: doi:10.1145/1127345.1127346
D. S. FALCONER; T. F. MACKAY: "Introduction to Quantitative Genetics", 1996, ADDISON WESLEY LONGMAN
C. DIAZ; S. SEYS; J. CLAESSENS; B. PRENEEL: "Towards measuring anonymity", PROCEEDINGS OF PRIVACY ENHANCING TECHNOLOGIES SYMPOSIUM (PETS, 2002
A. SERJANTOV; G. DANEZIS: "Towards an information theoretic metric for anonymity", PRIOCEEDINGS OF PRIVACY ENHANCING TECHNOLOGIES SYMPOSIUM (PETS, 2002
B. H. BLOOM: "Space/time trade-offs in hash coding with allowable errors", ACM COMMUNICATIONS, vol. 13, no. 7, 1970, pages 422 - 426, XP058297944, DOI: doi:10.1145/362686.362692
F. HAO; M. KODIALAM; T. V. LAKSHMAN: "Building high accuracy Bloom filters using partitioned hashing", PROCEEDINGS OF ACM INTERNATIONAL CONFERENCE ON MEASUREMENT AND MODELING OF COMPUTER SYSTEMS, 2007, pages 277 - 288
P. S. ALMEIDA; C. BAQUERO; N. PREGUI,CA; D. HUTCHISON: "Scalable Bloom filters", INFORMATION PRO- CESSING LETTERS, vol. 101, no. 6, 2007, pages 255 - 261, XP005831231, DOI: doi:10.1016/j.ipl.2006.10.007
"A map of human genome variation from population- scale sequencing", NATURE, vol. 467, 2010, pages 1061 - 1073
Attorney, Agent or Firm:
LEMAN CONSULTING S.A. 284 (Nyon, CH)
Download PDF:
Claims:
Claims

1. Method to process genomic data comprising the steps of :

at an initialization stage :

by a Certified Institution :

- associating a patient identification ID for a given patient P,

- generating a pair of asymmetric keys related to said patient P comprising a private and a public key,

- analyzing an output of the DNA sequencer and preparing an aligned genomic data for said patient P comprising approved variants (SNP), said approved variants being approved by medical authorities, each approved variant representing a position in the genome and a value representing a nucleotide that varies between individuals,

- extracting real and potential variants from said approved variants, said real and potential variants having each a position, said real variants being a subset of the approved variants and are different for each human being, said potential variants being the remaining part of the approved variants, - encrypting the value of each real variant with the public key of the patient,

- storing the encrypted values with their respective positions into a Storage and Processing Unit (SPU), as well as the patient identification ID,

- dividing the private key into at least a first and a second part,

- storing the second part of the private key in the Certified Institution or in a patient device;

- transmitting the first part of the private key to the SPU,

at a test stage :

- providing by the Certified Institution or the patient device the second part of the private key to a medical unit MU,

- selecting by the medical unit MU a personalized clinical test genetic test to be carried out and the related genetic markers, each marker having a position and a contribution,

- determining by the medical unit MU the contribution of each marker according to the personalized clinical test selected,

- transmitting by the MU the genetic markers with their respective contribution to the SPU as well as an identification ID of the patient P,

- retrieving by the SPU the encrypted values for said patient P matching the position of the genetic markers, and for said patient,

- executing by the SPU a genetic test by using the retrieved values, and the contribution of those markers thanks to homomorphic operations,

- partially decrypting by the SPU the result of the genetic test using the first part of the private key, - sending by the SPU the decrypted result to the MU,

- using by the MU the second part of the private key to obtain the result of the performed personalized clinical test.

2. Method of claim 1 , wherein it comprises the steps of :

- selecting all or part of the potential variants,

- adding these selected potential variants in the encrypting and storing steps.

3. Method of claim 2, wherein it comprises the steps of :

- analyzing the correlation between the variants and a privacy sensitivity of the real variants,

- selecting a number of other potential variants, said number being determined according to the previous analysis and a level of privacy required,

- adding these selected potential variants in the encrypting and storing steps.

4. Method to process genomic data according to claim 1 , further comprising the steps of :

during the initialization stage, by the Certified Institution :

- generating a dummy variant comprising a dummy position and a dummy value, said dummy position being outside of the overall variant positions of a sequence,

- encrypting the positions of the real variants with the symmetric key of the patient,

- encrypting the dummy value with the public key of the patient,

- encrypting the position of the dummy variant with the symmetric key of the patient,

- storing together with the encrypted variants, the dummy variant as well as the encrypted positions and the encrypted dummy position into a Storage and Processing Unit (SPU),

- storing the position of the dummy variant into the patient device,

at a test stage :

- determining by the CI a set of positions which are common between the marker's position and the real variant's positions,

- encrypting by the MU the set of positions with the symmetric key of said patient, and for the marker's positions not present in the variant's position, encrypting the dummy position and adding the dummy position to the set of positions,

- sending by the CI to the SPU the encrypted marker's positions as well as an identification ID of the patient P,

- retrieving by the SPU the encrypted values for said patient P at these encrypted locations and sending them to the MU.

Description:
Privacy-Enhancing Technologies for Medical Tests Using Genomic Data Introduction

In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. First, we highlight the potential privacy threats on genomic data and the challenges of providing privacy-preserving algorithms. Then, focusing specifically on a typical disease-susceptibility test, we develop a new architecture (between the patient and the medical unit) and propose privacy-preserving algorithms.

1 Background Art

Privacy control can be defined as the ability of individuals to determine when, how, and to what extent information about themselves is revealed to others. In this way, the usage of private data will remain in context and it will be used exclusively for the purpose the data owner has in mind. Privacy is usually protected by both legal and technological means. By using legal actions, such as data protection directives and fair information practices, privacy regulations can enforce privacy protection on a large scale. Yet, this approach is mostly reactive, as it defines regulations after technologies are put in place. To avoid this issue, Privacy-Enhancing Technologies (PETs) [1-3] can be incorporated into the design of new systems in order to protect individuals' data. PETs protect privacy by eliminating or obfuscating personal data, thereby preventing misuse or involuntary loss of data, without affecting the functionality of the information system [4].

Their objective is to make it difficult for a malicious entity to link information to specific users. In order to obfuscate personal data, PETs often rely on cryptographic primitives, such as anonymous authentication and encryption.

Genomics is becoming the next significant challenge for privacy. The price of a complete genome profile has plummeted below $200 for genome-wide genotyping (i.e., the characterization of about one million common genetic variants), which is offered by a number of companies (located mostly in the US). Whole genome sequencing is also offered through the same direct-to-consumer model (but at a higher price). This low cost of DNA sequencing will break the physician/patient connection, because private citizens (from anywhere in the world) can have their genome sequenced without involving their family doctor. This can open the door to all kinds of abuse, not yet fully understood.

As a result of the rapid evolution in genomic research, substantial progress is expected in terms of improved diagnosis and better preventive medicine. However, the impact on privacy is unprecedented, because (i) genetic diseases can be unveiled, (ii) the propensity to develop specific diseases (such as Alzheimer's) can be revealed, (iii) a volunteer accepting de facto to have his genomic code made public (as it already happened) can leak substantial information about his ethnic heritage and genomic data of his relatives (possibly against their will), and (iv) complex privacy issues can arise if DNA analysis is used for criminal investigations and insurance purposes. Such issues could lead to genetic discrimination (e.g., ancestry discrimination or discrimination due to geographic mapping of people). Even though the Genetic Information Non-discrimination Act (GINA), which prohibits the use of genomic information in health insurance and employment, attempted to solve some of these problems in the US, these types of laws are very difficult to enforce. An even more severe case, currently not widely considered, is where a malicious party initiates a cross-layer attack by utilizing privacy-sensitive information belonging to a person retrieved from different sources (e.g., genomic data, location, online social network, etc.), thus creating the opportunity for a large variety of fraudulent uses of such data. For example, as stated in the Personal Genome Project (PGP) consent form [5], a malicious party could make synthetic DNA of a person and plant it at a crime scene to falsely accuse him.

In this hypothetical situation, the attacker can make his accusation stronger if he has the location patterns of the person to be blamed, and hence, knows that the person was close to the crime scene at the time of the crime. Similarly, an attacker can easily obtain information on close relatives of a target from online social network data, thus effectively increasing the potential access to target's genomic data if his relatives' DNA has been sequenced. In other words, even if the person has perfect privacy on his own genome, if the attacker has access to the DNA sequence of the relatives, he can obtain significant information about the person's DNA sequence.

Even though, at this stage, the field of genomics is generally free from serious attacks, it is likely that the above threats will become more serious as the number of sequenced individuals becomes larger. Such was the case of the Internet that was initially run and used by well-intentioned researchers. However, once it became more widely used, it became plagued by uncountable attacks such as spyware, viruses, spam, botnets, Denial-of-Service attacks, etc. Therefore, the need to adapt PETs to personal genomic data will only grow with time, as they are key tools for preventing an adversary from linking particular genomic data to a specific person or from inferring privacy-sensitive genomic data about a person.

It is obvious that users need to reveal personal and even privacy-sensitive information for genomic tests (e.g., paternity tests, disease-susceptibility tests, etc.). Nevertheless, they want to control how this information is used by the service providers (e.g., medical units such as healthcare centers or pharmaceutical companies, depending on the type of the test). Currently, the companies and hospitals that perform DNA sequencing store the genomic data of their customers and patients. Of course, tight legislation regulates their activities, but it is extremely difficult for them to protect themselves against the misdeeds of a hacker or a disgruntled employee. In a non-adversarial scenario, however, making use of this data requires legitimate professionals (e.g., physicians and pharmacists) to access the data in some way. Therefore, new architectures and protocols are needed to store and process this privacy-sensitive genomic data, while still enabling its utilization by the service providers (e.g., medical units).

In this work, our goal is to protect the privacy of users' genomic data while enabling medical units to access the genomic data in order to conduct medical tests or develop personalized medicine methods. In a medical test, a medical unit checks for different health risks (e.g., disease susceptibilities) of a user by using specific parts of his genome. Similarly, to provide personalized medicine, a

pharmaceutical company tests the compatibility of a user on a particular medicine, or a pharmacist checks the compatibility of a given medicine (e.g., over-the-counter drug) to a given user. In both scenarios, in order to preserve his privacy, the user does not want to reveal his complete genome to the medical unit or to the pharmaceutical company. Moreover, in some scenarios, it is the pharmaceutical companies who do not want to reveal the genetic properties of their drugs. To achieve these goals, we propose to store the genomic data at a Storage and Processing Unit (SPU) and conduct the computations on genomic data utilizing homomorphic encryption and proxy encryption to preserve the privacy of the genomic data.

The rest of the paper is organized as follows. In the rest of this section, we discuss the challenges in genomic privacy and summarize the related work on genomic privacy. In Section 2, we describe our proposed schemes for privacy-preserving medical tests and personalized medicine. Furthermore, we analyze the level of privacy provided by the proposed schemes for different design and genomic criteria. Then, in Section 3, we discuss the implementation of the proposed schemes and present their complexity and security evaluations.

Finally, in Section 4, we conclude the paper and discuss new research directions on genomic privacy.

1.1 Challenges of Genomic Privacy

Obviously, there are certain obstacles for achieving our goals on genomic privacy. These are mostly due to (i) the balance between privacy and reliability of the genomic data, (ii) the structure of the human genome, and (iii) the evolution of the genomic research.

PETs generally protect users' privacy by either breaking the link between individuals' identities and the data they provide (e.g., removing user's identities from the published genomic data), or by decreasing the information provided (e.g., by using cryptographic tools or obfuscation techniques). Both techniques might reduce the reliability and the accuracy of the genomic data. Thus, a major issue to be addressed when designing PETs is limiting private information leakage while keeping an acceptable level of reliability and accuracy of the genomic data for the researchers and medical units.

Moreover, developing PETs for genomic data has many unique challenges, due to the architecture of the human genome. The human genome is encoded in double stranded DNA molecules consisting of two complementary polymer chains. Each chain consists of simple units called nucleotides (A,C,G,T). The human genome consists of approximately three billion letters. Existing privacy-preserving methods do not scale to these large genomic data sizes; hence current algorithms are inadequate for privacy protection on the genomic level.

Finally, the rapid evolution in the field of genomics produces many new discoveries every year, which cause significant changes in the known facts. For example, the sensitivity of certain genomic information will change over time; hence it is crucial to develop dynamic algorithms that can smoothly adapt to this rapid evolution.

1.2 Related Work

Due to the sensitivity of genomic data, research on the privacy of genomic data has considerably accelerated over the past few years. We can put the research on genomic privacy in three main categories: (i) private string searching and comparison, (ii) private release of aggregate data, and (iii) private clinical genomics.

In [6], Troncoso-Pastoriza ef al. propose a protocol for string searching, which is then improved by Blanton and Aliasgari [7]. In this approach, one party with his own DNA snippet can verify the existence of a short template within his snippet by using a Finite State Machine in an oblivious manner. To compute the similarity of DNA sequences, in [8], Jha ef al. propose techniques for privately computing the edit distance of two strings by using garbled circuits. In [9], Bruekers ef al. propose privacy-enhanced comparison of DNA profiles for identity, paternity and ancestry tests using homomorphic encryption. Similar to our work, in [10], Kantarcioglu ef al. propose using homomorphic encryption to perform scientific investigations on integrated genomic data. In their scheme, all genomic data is encrypted by the same public key of the data storage site, and there is a single key holder site which can decrypt everything. Thus, a curious party at the key holder site can obtain the genomic information of all users in case of a possible data leakage from the data storage site. Moreover, in [10], only the encrypted variants (i.e., positions in the genome holding a nucleotide that varies between individuals) of the users are stored at the data storage site along with their plaintext locations (on the DNA), which can leak substantial information to the data storage site about the genomic sequences of the users, as we discuss in Section 2.4. As opposed to [10], we focus on personal use of genomic data (e.g., in medical tests and personalized medicine methods), propose methods in which each user's genomic data is encrypted via his own cryptographic key, and prevent the leakage of genomic data due to statistical dependence between the variants. In one of the recent works [1 1], Baldi ef al. make use of both medical and cryptographic tools for privacy-preserving paternity tests, personalized medicine, and genetic compatibility tests. Instead of utilizing public key encryption protocols, in [12], Canim et al. propose securing the biomedical data using cryptographic hardware. Finally, in [13], Eppstein ef al. propose a privacy-enhanced method for comparing two compressed DNA sequences by using Invertible Bloom Filter [14].

When releasing databases consisting of aggregate genomic data (e.g., for research purposes), it is shown that known privacy-preserving approaches (e.g., de-identification) are ineffective on (unencrypted) genomic data [15, 16]. Homer ef al. [17] prove that the presence of an individual in a case group can be determined using aggregate allele frequencies and his DNA profile. In another recent study [18], Gitschier shows that a combination of information, from genealogical registries and a haplotype analysis of the Y chromosome collected for the HapMap project, allows for the prediction of the surnames of a number of individuals held in the HapMap database. Thus, releasing genomic data (even in aggregate form) is currently banned by many institutions due to this privacy risk. In [19], Zhou ef al. study the privacy risks of releasing the aggregate genomic data. They propose a risk-scale system to classify aggregate data and a guide for the release of such data. Recently, using differential privacy was proposed by Fienberg ef al. [20]; they aim to ensure that two aggregated databases, differing from each other by only one individual's data (e.g., DNA sequence), have indistinguishable statistical features.

Recently, in [21], utilizing a public cloud, Chen ef al. propose a secure and efficient algorithm to align short DNA sequences to a reference (human) DNA sequence (i.e., read mapping). Finally, in [22], Wang ef al. propose a privacy-protection framework for important classes of genomic computations (e.g., search for homologous genes), in which they partition a genomic computation, distributing sensitive data to the data provider and the public data to the data user.

In this work, we focus on medical tests (e.g., disease-susceptibility test) and personalized medicine methods by using users' genomic data while protecting user's genomic privacy. As a result of our extensive collaboration with geneticists, clinicians, and biologists, we conclude that DNA string comparison is insufficient in many medical tests (that use genomic data) and would not be enough to pave the way to personalized medicine. As it will become clearer in the next sections, specific variants must be considered individually for each genetic test. Thus, as opposed to the above private string search and comparison techniques, which focus on privately comparing the distance between the genomic sequences, we use the individual variants of the users to conduct genetic disease susceptibility tests and develop personalized medicine methods. We consider the individual contribution of each variant to a particular disease, for which a string comparison algorithm (such as Private Set Intersection [23, 24]) would not work. Further, in our proposed algorithms, we consider the statistical relationship between the variants for the genomic privacy of the users. In addition, we make use of a Storage and Processing Unit (SPU) between the user (patient) and the medical unit to store the genomic data in encrypted form and make computations on it using homomorphic encryption and proxy encryption.

1.3 Brief description of the invention

The aim of the present invention is to propose a privacy-enhancing method for medical tests and personalized medicine methods, which utilize patients' genomic data. It is proposed a method to process genomic data comprising the steps of :

at an initialization stage :

- associating a patient identification ID for a given patient P,

- generating a pair of asymmetric keys related to said patient P comprising a private and a public key,

- preparing a DNA sequence for said patient P comprising approved variants(SNP), said approved variants being approved by medical authorities, each approved variant representing a position in the genome and a value representing a nucleotide that varies between individuals,

- extracting real and potential variants from said approved variants, said real and potential variants having each a position, said real variants being a subset of the approved variants and are different for each human being, said potential variants being the remaining part of the approved variants,

- encrypting the value of each real variant with the public key of the patient,

- storing the encrypted values with their respective positions into a Storage and Processing Unit (SPU), as well as the patient identification ID,

- dividing the private key into at least a first and a second part,

- transmitting the first part of the private key to the SPU,

at a test stage :

- providing the second part of the private key to a medical unit MU,

- selecting by the medical unit MU a genetic test to be carried out and the related genetic markers, each marker having a position and a contribution,

- determining the contribution of each marker according to the genetic test selected,

- transmitting by the MU the genetic markers with their respective contribution to the SPU as well as an identification ID of the patient P, - retrieving by the SPU the encrypted values for said patient P matching the position of the genetic markers, and for said patient,

- executing by the SPU a genetic test by using the retrieved values, and the contribution of those markers thanks to homomorphic operations,

- decrypting the result of the genetic test using the first part of the private key,

- sending the decrypted result to the MU,

- using the second part of the private key to obtain the final result.

The method of the invention is split into a first phase in which the DNA sequence is processed and stored in the SPU and a second phase during which a test is carried out.

During the first phase, the DNA sequence, produced by an authorized laboratory, is processed and encrypted as explained above. During this second phase, the medical test selected by the medical unit is carried out without having the possibility to retrieve all information of the patient.

The method proposed by the invention is based on the use of homomorphic encryption and proxy encryption. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by their public keys at a Storage and Processing Unit (SPU). The proposed algorithm lets the SPU (or the medical unit) process the encrypted genomic data for medical tests and personalized medicine methods while preserving the privacy of patients' genomic data. We extensively analyze the relationship between the storage cost (of the genomic data), the level of genomic privacy (of the patient), and the characteristics of the genomic data. Furthermore, we implement and show via a complexity analysis the practicality of the proposed schemes. Finally, we evaluate the security of the proposed schemes and propose new research directions on genomic privacy.

1.4 Brief description of the figures

The invention will be better understood thanks to the attached figures in which :

- The Figure 1 illustrates the General architecture between the patient, SPU, and the medical unit.

- The Figure 2 illustrates the Privacy-preserving protocol for disease-susceptibility test using Method 1 or Method 2.

- The Figure 3 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the number of LD pairs per SNP (i.e., [i( )) and storage redundancy.

- The Figure 4 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the LD strength between two SNPs (i.e., μ(Ι)) and storage redundancy.

- The Figure 5 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying standard deviation and mean values of the number of LD pairs per SNP (i.e., σ (k) and [i(k)) and storage redundancy. - The Figure 6 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying standard deviation and mean values of the LD strength between two SNPs (i.e., o(l) and μ(Ι)) and storage redundancy.

- The Figure 7 illustrates the Increase in genomic privacy of different types of patients with 100% increments in the storage redundancy. For example, increasing the storage redundancy from 400% to 500% would increase the privacy of Patient A (who carries mostly low severity real SNPs) by 5%, whereas the same scenario increases the privacy of Patient B (who carries mostly high severity SNPs) by 13%.

- The Figure 8 illustrates the Level of genomic privacy, as defined by (8), for different types of patients with varying storage redundancy.

- The Figure 9 illustrates the Privacy-preserving protocol for disease-susceptibility test using Method 3.

- The Figure 10 illustrates the Privacy, practicality, and storage overhead comparison of the proposed methods.

2 PETs for Medical Tests and Personalized Medicine Methods

In the present case, we study the privacy issues of medical tests and personalized medicine methods. Most medical tests and personalized medicine methods (that use genomic data) involve a patient and a medical unit. The patient is identified by a patient identification (ID), which could be a user name or a pseudonym (e.g., hash value of his social security number). In general, the medical unit is the family doctor, a physician, a pharmacist, or a medical council. In this study, we consider a malicious medical unit as the potential attacker. That is, a medical unit can be a malicious institution trying to obtain private information about a patient. Even if the medical unit is non-malicious, it is extremely difficult for medical units to protect themselves against the misdeeds of a hacker or a disgruntled employee. Similarly, the genomic data is too sensitive to be stored on users' personal devices (mostly due to security, availability, and storage issues), hence it is risky to leave the users' genomic data in their own hands. In addition, extreme precaution is needed between the patient and the medical unit due to the sensitivity of genomic data. Thus, we believe that a Storage and Processing Unit (SPU) should be used to store and process the genomic data. We note that a private company (e.g., cloud storage service), the government, or a non-profit organization could play the role of the SPU. We also assume that the SPU is an honest organization, but it might be curious (e.g., existence of a curious party at the SPU), hence genomic data should be stored at the SPU in encrypted form (i.e., the SPU should not be able to access the content of patients' genomic data). This general architecture is illustrated in Fig. 1.

For the simplicity of presentation, in the rest of this section, we will focus on a particular medical test (namely, computing genetic disease susceptibility). We note that similar techniques would apply for other medical tests and personalized medicine methods. In a typical disease-susceptibility test, a medical unit (MU) wants to check the susceptibility of a patient (P) to a particular disease X (i.e., probability that the patient P will develop disease X). It is shown that a genetic disease-susceptibility test can be realized by analyzing particular Single Nucleotide Polymorphisms (SNPs) of the patient via some operations, such as weighted averaging [25] or Likelihood Ratio (LR) test [26]. A SNP is a position in the genome holding a nucleotide (A, T, C or G), which varies between individuals. For example, it is reported that there are three particular genes bearing a total of ten particular SNPs necessary to analyze a patient's susceptibility to Alzheimer's disease [27]. Each SNP contributes to the susceptibility in a different amount and the contribution amount of each SNP is determined by previous studies on case and control groups (these studies are published in several papers).

Furthermore, some of the SNPs contribute to the development of a disease, whereas some are protective.

In general, there are two alleles observed at a given SNP position: (i) The major allele is the most frequently observed nucleotide, and (ii) the minor allele is the rare nucleotide. Everyone inherits one allele of every SNP location from each of his parents. If an individual receives the same allele from both parents, he is said to have a homozygous variant for that SNP location. If, however, he inherits a different allele from each parent (one minor and one major), he has a heterozygous variant. There are approximately 40 million approved variants (SNPs) in the human population as of now (according to the NCBI dbSNP [28]) and each patient carries on average 4 million SNPs (i.e., real variants) out of this 40 million. Moreover, this set of 4 million SNPs is different for each patient. From now on, to avoid confusion, for each patient, we refer to these 4 million variants as the real SNPs and the remaining non-variants (approved SNPs that do not exist for the considered patient) as the potential SNPs of the patient; when we only say "SNPs", we mean both the real and potential SNPs.

At this point, it can be argued that these 4 million real SNPs (nucleotides) could be easily stored on the patient's computer or mobile device, instead of the SPU. However, we assert that this should be avoided due to the following issues. On one hand, the number of approved SNPs in human population continues to increase with new discoveries. Further, types of variations in human population are not limited to SNPs, and there are other types of variations such as Copy-Number Variations (CNVs), rearrangements, or translocations (our proposed privacy-preserving mechanisms can be smoothly adapted for these alternative variations), consequently the required storage per patient is likely to be considerably more than only 4 million nucleotides. This higher storage cost might still be affordable to an average patient (via desktop computers or USB drives), however, genomic data of the patient should be available any time (e.g., for emergencies), thus it should be stored at a reliable source such as the SPU. On the other hand, as we discussed before, leaving the patient's genomic data in his own hands and letting him store it on his computer or mobile device is risky, because his mobile device can be stolen or his computer can be hacked.

A potential attacker can learn about the susceptibilities of the patient to privacy-sensitive diseases if he obtains some specific real SNPs of the patient. Moreover, the knowledge of 75 real SNPs (out of approximately 4 million), if not fewer, will enable the attacker to identify a person [29]. These situations could lead to genetic discrimination such as denying a person's access to health (or life) insurance or obstructing his employment opportunities. As we discussed before, in our setting, both the MU and SPU pose a threat to the patient's privacy. On one hand, the MU can either be a malicious institution trying to obtain private information about the patient or it can be hacked by another malicious entity. On the other hand, the SPU is considered as an honest but curious entity. Thus, our goal is to build mechanisms in which the patient can preserve the privacy of his genomic sequence (his real SNPs) while enabling the MU to access his genomic data and conduct genetic tests. We assume that the whole genome sequencing is done by a Certified Institution (CI) with the consent of the patient. Moreover, the genomic data of the patient is encrypted by the same CI (using the patient's public key) and uploaded to the SPU so that only the patient can decrypt the stored (potential or real) SNPs, and the SPU cannot access the SNPs of the patient. We are aware that the number of discovered SNPs increases with time. Thus, the patient's complete DNA sequence is also encrypted as a single vector file (via symmetric encryption using the patient's key) and stored at the SPU, thus when new SNPs are discovered, these can be included in the pool of the previously stored SNPs of the patient. We also assume the SPU does not have access to the real identities of the patients and data is stored at the SPU by using pseudonyms; this way, the SPU cannot associate the conducted genomic tests to the real identities of the patients. As an alternative, the privacy of the genomic data at the SPU can be further increased using privacy enhanced access control [30] or Oblivious RAM (O- RAM) storage [31] techniques, in which the data access patterns are completely hidden from the server (SPU). Note that even the most efficient implementation of ORAM introduces high storage overhead to the client (patient), and it introduces 20 ~ 25 times more overhead with respect to non-

oblivious storage. Thus once it becomes more efficient, O-RAM storage could be considered as a future add-on to the proposed privacy-preserving mechanisms.

Depending on the access rights of the MU, the SPU can either (i) compute Pr(X), the probability that the patient will develop the disease X by checking the patient's encrypted SNPs via homomorphic encryption techniques [33] (In one of our proposed schemes, see Method 3 in Section 2.4, Pr(X) is computed at the MC via homomorphic operations), or (ii) provide the relevant SNPs to the MU (e.g., for complex diseases that cannot be interpreted using homomorphic operations). These access rights are defined either jointly by the MU and the patient or by the medical authorities. Further, access rights can be enforced by using a secure attributebased system as in [34]. We note that homomorphic encryption lets the SPU (or MU) compute Pr(X) using encrypted SNPs of the patient P. In other words, the SPU (or MU) does not access P's SNPs to compute his predicted disease susceptibility. We use a modification of the Paillier cryptosystem (described in Section 2.1 ) to support the homomorphic operations at the SPU (or MU).

We propose four different techniques for the storage and process of the SNPs at the SPU and the preservation of the patient's privacy: (i) Method 0 in Section 2.2, (ii) Method 1 in Section 2.3, (iii) Method 2 in Section 2.4, and (vi) Method 3 in Section 2.5. We describe these proposed techniques in detail in the following subsections. We also discuss the computation of genetic disease susceptibility by using homomorphic operations in Section 2.6.

In the rest of this work, for simplicity of the presentation, we do not consider the type of the variant at a real SNP location (i.e., whether the variation is homozygous or heterozygous for that real SNP); we only consider whether the patient has a real SNP or not at a particular location. However, the proposed approaches and the analysis (in Section 2.4) can easily be extended to cover the types of the variants. In order to facilitate future references, frequently used notations are listed in Table I for the different stages of the proposed schemes.

2.1 Paillier Cryptosystem

In this section, we briefly review the modified Paillier cryptosystem (described in detail in [33, 35]), which we use in this work, and its homomorphic properties. We note that the usual notation in Paillier cryptosystem is to use a pair of keys named public and secret key. However, for the present description, we will use the denote the keys as public and private.

The public key of the patient P is represented as (n, g, h = g x ), where the strong private key is the factorization of n = pq (p, q are safe primes), the weak private key is x e [1 , n 2 /2], and g of order (p-1 )(q-1 )/2. Such a g can be easily found by selecting a random a e Z * 2 and computing g = -a 2 ".

Encryption of a message: To encrypt a message m e Z * 2 , we first select a random r e [1 , n/4] and generate the ciphertext pair (T 1: T 2 ) as below:

T x = g r mod n' and T 2 = h r (l + mn) mod n 2 (1 )

Re-encryption of a message: An encrypted message (T ? , T 2 ) can be re-encrypted under the same public key, using a new random number r-ι e [1 , n/4] as below:

7 = g ri 7 mod « 2 and T = h ri T modn (2)

Decryption of a message: The message m can be recovered as follows:

m = A(T 2 / T l x ) , (3)

(u - l)mod « 2 , f _ 2 i , . )

where A (u ) ■i '- , for u e \ u < n « = lmod« Homomorphic properties: Assume two messages m-ι and m 2 are encrypted using two different random numbers r-ι and r 2 , under the same public key, (n, g, h = g x ), such that E(m-i , r-ι , g x ) = (T^ , T 2 ) and E(m 2 , r 2 , g x ) = (7 , T 2 ). Assume also that c is a constant number. Then the below-mentioned homomorphic properties are supported by Paillier cryptosystem:

· The product of two ciphertexts will decrypt to the sum of their corresponding plaintexts.

D(E(m 1 ,r 1 ,g x ) - E(m 2 , r 2 , g x )) = D^ 1 - T , T · Γ 2 2 mod« 2 ) = m l + m 2 modn. (4)

• An encrypted plaintext raised to a constant c will decrypt to the product of the plaintext and the constant.

D (E (m 1 , r x , g x J J = D J , (τ^ J mod n 2 J = cm, mod n. (5) These homomorphic operations are conducted at the SPU (or MU depending on which approach is used) to compute the predicted susceptibility of the patient P to disease X, as will be discussed in Section 2.6.

Proxy encryption: The patient's weak private key x is divided (preferably randomly or by any other rule) into two shares: x < ) and x (2) (such that x = x (1) + x (2) ). x (1) is given to the SPU and x (2) is given to the MU. Using the above Paillier cryptosystem, an encrypted message ( T^ , T 2 ) (under the patient's public key) can be partially decrypted by the SPU (using x (1) ) to generate the ciphertext pair { %) as below: °= Τ 2 and † > = T 2 1 " mod n . (6)

Now, ( f v f 2 ) can be decrypted at the MU using x (2) to recover the original message. x (2) can be provided to the MU once the patient is registered to the medical unit or through the patient's digital ID card. Further details about the distribution of shares are out of the scope of this paper. We note that this approach is not proxy re-encryption; it is based on secret-sharing.

Overall, this modified Paillier cryptosystem is not key optimal, because the size of the MU's and SPU's secret storages do not remain constant. That is, both the MU and SPU need to store a secret for every patient.

However, this storage cost can be considered negligible when compared to the storage of the genomic data. Further, the shares (e.g., x (1) and x (2) ) can be stored by the patient and sent to the MU and SPU only when it is needed in order to resolve this storage issue at the expense of extra communication overhead. Furthermore, the above modified Paillier cryptosystem is not proxy invisible, because all participants of the systems (i.e., P, MU and SPU) should be aware of the existence of the proxy. We discuss the security evaluation of this cryptosystem in Section 3.2.

2.2 Method 0: Only Store the Real SNPs at the SPU

In this approach, the real SNPs of the patient are stored encrypted (via the patient's public key) and the locations of the corresponding real SNPs are stored in plaintext at the SPU. We assume that SNP, at the patient P is represented as SNP and SNPf = 1 , if P has a real SNP

(i.e., variant) at this location, and SNP = 0, if P does not have a variant at this location. We let Y be the set of real SNPs of the patient P (at which SNP = 1 ). We also let P represent the set of potential

SNPs (at which SNPf = 0).

Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique.

• Step 0: The asymmetric keys (public and private keys) of each patient are generated and distributed to the patients during the initialization period. Then, symmetric keys are established between the parties, using which the communication between the parties is protected from an eavesdropper. We note that the distribution, update and revocation of cryptographic keys are handled by a trusted entity (similar to e-banking platforms).

• Step 1 : The patient (P) provides his sample (e.g., his saliva) to the Certified Institution (CI) for sequencing.

• Step 2: The CI sequences P, and encrypts the contents of his real SNP locations (in Y ) by using P's public key.

• Step 3: The CI sends the encrypted real SNPs of P to the SPU (so that the SPU cannot access to P's SNPs).

• Step 4: We divide the private key into a first and a second part, the patient provides the first part of his private key (x (1) ) to the SPU.

· Step 5: The MU wants to conduct a susceptibility test on P to a particular disease X, and P provides the second part of his private key (x (2) ) to the MU as well as his identification ID.

• Step 6: The MU provides genetic variant markers, along with their individual contributions (to the disease susceptibility), to the SPU.

• Step 7: If the disease susceptibility can be interpreted by homomorphic operations, the SPU computes P's total susceptibility to disease Xfrom the individual effects of SNPs by using the homomorphic properties of the Paillier cryptosystem as described in Section 2.6. Otherwise, the SPU provides the relevant real SNPs to the MU based on MU's access rights.

• Step 7: The SPU partially decrypts the end-result (or the relevant SNPs) using the first part of P's private key for example by following a proxy encryption protocol (Section 2.1 ).

· Step 8: The SPU sends the partially decrypted end-result (or the relevant real SNPs) to the MU.

• Step 9: The MU decrypts the message received from the SPU using the second part of P's private key and recovers the end-result (or the relevant real SNPs).

2.3 Method 1 : Plaintext Locations at the SPU

Method 0 in Section 2.2 might leak private information to the curious party at the SPU. As the locations of the SNPs are stored in plaintext, if the SPU only stores the real SNPs in Y , a curious party at the SPU can learn all real SNP locations of the patient, and hence, much about his genomic sequence. The nucleotides corresponding to variants at particular locations of the DNA sequence are public knowledge. Thus, even though the contents of patient's real SNPs are encrypted, a curious party at the SPU can infer the nucleotides corresponding to these SNPs from their plaintext locations. Therefore, in this method, the SPU stores the contents of both real and potential SNP locations (in |Y UQ p } ) in order to preserve the privacy of the patient. The locations of the corresponding SNPs are again stored in plaintext at the SPU. This is because, when a particular SNP (or set of SNPs) are queried by the MU, the SPU should know which SNPs to process (or send to the MU).

As before, we assume that SNP, at the patient P is represented as SNP and SNP = 1 , if P has a real SNP (i.e., variant) at this location, and SNP = 0, if P does not have a variant at this location. We let Y be the set of real SNPs of the patient P (at which SNP f = 1 ). We also let P represent the set of potential SNPs (at which SNP = 0). Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique. This approach is illustrated in Fig. 2.

• Step 0: The asymmetric keys (public and private keys) of each patient are generated and distributed to the patients during the initialization period. Then, symmetric keys are established between the parties, using which the communication between the parties is protected from an eavesdropper. We note that the distribution, update and revocation of cryptographic keys are handled by a trusted entity (similar to e-banking platforms).

• Step 1 : The patient (P) provides his sample (e.g., his saliva) to the Certified Institution (CI) for sequencing.

• Step 2: The CI sequences P, and encrypts the contents of his real and potential SNP locations (in {Y P UQ p } ) by using P's public key.

• Step 3: The CI sends the encrypted SNPs of P to the SPU (so that the SPU cannot access to P's SNPs).

· Step 4: We divide the private key into a first and a second part, the patient provides the first part of his private key (x (1) ) to the SPU.

• Step 5: The MU wants to conduct a susceptibility test on P to a particular disease X, and P provides the second part of his private key (x (2) ) to the MU as well as his identification ID.

• Step 6: The MU provides genetic variant markers, along with their individual contributions (to the disease susceptibility), to the SPU.

• Step 7: If the disease susceptibility can be interpreted by homomorphic operations, the SPU computes P's total susceptibility to disease Xfrom the individual effects of SNPs by using the homomorphic properties of the Paillier cryptosystem as described in Section 2.6. Otherwise, the SPU provides the relevant SNPs to the MU based on MU's access rights.

· Step 7: The SPU partially decrypts the end-result (or the relevant SNPs) using the first part of P's private key for example by following a proxy encryption protocol (Section 2.1 ). • Step 8: The SPU sends the partially decrypted end-result (or the relevant SNPs) to the MU.

• Step 9: The MU decrypts the message received from the SPU using the second part of P's private key and recovers the end-result (or the relevant SNPs).

The above technique provides a high level of privacy and practicality for the patient, because (i) from the view point of a curious party at the SPU, inferring the locations of the patient's real SNPs with the stored information is equivalent to inferring them with no information about the patient, and (ii) the patient is not involved in the protocol after the sequencing (except for the consent between the patient and the MU for a particular test). However, this level of privacy and practicality comes at the cost of extra storage overhead at the SPU (due to the storage of both real and potential SNPs as discussed in Section 3.1 ).

2.4 Method 2: Redundant Storage at the SPU

Due to the significant storage overhead mentioned in Section 2.3, here we propose another technique that reduces the storage overhead at the SPU at the expense of decrease in privacy. In a nutshell, we leave everything the same as in Section 2.3, but, instead of storing the contents of all potential and real SNP locations, we store the real SNPs of the patient along with a certain level of redundancy (i.e., contents of some potential SNP locations). In other words, to mislead a curious party at the SPU, among the 40 million discovered SNPs, we store the approximately 4 million real SNPs (for which p P

SNP i = 1 , / e ¾ ) along with some redundant content from ¾ (with SNPJ = 0), for each patient.

Again, we assume that the location of the encrypted (real or potential) SNPs are stored in plaintext at the SPU and there exists a potential curious party at the SPU trying to infer the real SNPs of the patient (in ). An important issue to consider in this approach is the Linkage Disequilibrium (LD) between SNPs [36].

LD occurs when SNPs at two loci (SNP positions) are not independent of each other. For simplicity, we represent the LD relationship between two SNPs i and j as Pr(SNPi|SNP j ), where SNP, (or SNP j ) takes values from the set {0, 1}. In compliance with genetic observations, we assume that the LD between two SNPs are not symmetric, i.e., Pr(SNPi|SNP j ) ≠ Pr(SNP |SNP|). We note that LD relationships are defined among all 40 million discovered SNPs, regardless of their type (i.e., real or potential) at a particular patient.

As in Section 2.3, the SPU provides the end-result of a disease-susceptibility test or the relevant SNPs to the MU. However, in this case, if a particular potential SNP (requested by the MU or needed in the

p

susceptibility test) is not stored at the SPU (i.e., SNPJ = 0), one of the following two scenarios occurs: (i) If the SPU provides the relevant SNPs to the MU, MU infers the missing potential SNPs from the reference genome (since it is known that the missing potential SNPs are not a variant for P), p or (ii) if the SPU provides the end-result of the susceptibility test, the SPU uses the fact that SNPJ = 0 for each missing potential SNP j . As expected, the amount of storage redundancy (due to the storage of the content from ¾ ), along with the LD between the SNPs and their characteristics, determine the level of a patient's genomic privacy.

Therefore, in the rest of this section, we analyze the relationship between the amount of redundancy, LD values, characteristics of the SNPs, and the level of privacy. To do so, first, we observe the average probability of correctly inferring the locations of P's real SNPs (in ^p ) considering varying amounts of redundancy and the LD values between the SNPs. That is, how much information from a patient's un-stored potential SNPs is revealed to the curious party at the SPU about the locations of his real SNPs ? This problem can also be formulated similarly if the goal of the attacker is to determine

P

the type of the variant at a real SNP location (e.g., homozygous or heterozygous). In this case, SNP ϊ can take three different values from the set {0, 1 , 2}, 0 representing a potential SNP (i.e., non-variant) 1 representing a real homozygous SNP, and 2 representing a real heterozygous SNP for P. It is worth noting that for this study, we create realistic models for the LD values and the characteristics of the SNPs. Further, for the created models, we try a wide range of parameters and observe a wide range of results to address most potential scenarios. However, as the field of genomics becomes more mature, our models can be replaced by the values obtained from the medical research.

We let and ¾ denote the set of P's potential SNPs that are stored (for redundancy) and not stored at the SPU, respectively (¾ U = % ). Further, K, is the set of SNPs with which a particular SNP i has LD, and |K,| = k (for each SNP, these k SNPs are chosen among approximately 40 million SNPs). We assume that k > 0 and it is a truncated Gaussian random variable with only discrete values and obtained from a distribution with mean o(k) and standard deviation (k).

F y ¾

Initially, we compute Pr(SNP j ) for all (real and potential) SNPs in { } by using the LD relationships between these SNPs and those in l % . As all SNPs in { } are encrypted and stored at the SPU, only the LD relationships between these SNPs and the un-stored SNPs in "p are helpful for the curious party.

P P

Therefore, for each real SNP / fc Y ? , we observe Pr(SNP : = 1 | SNP« = 0) for all

P

« e Pip ) flie £½½ l i }, get the average of these values, and compute Pr(SNP i = 1 ).

P p

Similarly, for each potential SNP j - f l , we observe Pr(SNP = 0| SNPm = 0) for all

= 0). We let / be the indicator of

P F

the LD strength between two SNPs. Thus, we represent Pr(SNP t = 1 | SNPIB = 0) = / ( e Y p , ) and Pr(SNP; = 0| SNPm = 0) = / (I · Q ¼ »t 6 ) i f½% as truncated Gaussian random variables with range [0.5, 1], obtained from a distribution with mean μ(Ι) and standard deviation o(l).

Finally, if |K,| = k = 0 or | , | = 0 for a SNP / in { }, we update Pr(SNP . = 1 ) considerin the fact that the expected value of all stored SNPs is known by the curious party as below: . (7)

In the following, we illustrate our numerical results that represent the relationship between storage, inference power of the curious party at the SPU, and LD values. We assume | Y | = 4 million and |

Y ΙΙΩ ρ I = 40 million. We define the percentage of storage redundancy at the SPU as P» l| χ 100

P

and compute the average value of Pr(SNP t = 1 ) for a SNP in Y for varying values of \i(k), (k), μ(Ι),

P

and σ(Ι). Higher values of Pr(SNP j = 1 ) indicate a higher inference power for the curious party at the SPU. We repeat each simulation 100 times to obtain an average. Note that Method 1 (in Section 2.3) is a special case of Method 2 (when the storage redundancy at the SPU is 900%), hence its privacy is the same as 900% redundancy in the following results.

P

In Fig. 3, we illustrate the variance in the average value of Pr(SNP i = 1 ) for different values of \i(k), when μ(Ι) = 0.8, σ(Ι) = 0.15, and a(k) = 0.75. We note that "no LD" curve in the figure represents the case in which the LD values between the SNPs are ignored. We observe that as the available side information (i.e., number of un-stored potential SNPs in ¾ having LD with the stored ones) increases, the inference power of the curious party increases, especially for low values of storage redundancy. For example, to have the same inference power for the curious party, 200% storage redundancy is required when \i(k) = 0, whereas it is 700% when [i( ) = 4. Furthermore, even at the maximum (i.e, 900%) storage redundancy, the curious party still has a slight probability of inferring the variants of the patient, because it knows that 4 out of 40 million of the stored content are variants. Next, in Fig. 4, we illustrate the variance in the same probability, this time for different values of μ(Ι), when [i(k) = 2, a(k) = 0.75, and σ(Ι) = 0.25. For higher values of o(l), the gap between the different μ(Ι) curves becomes negligible, because the distribution becomes almost uniform, rather than truncated Gaussian. As expected, the inference power of the curious party increases when the strength of LD between the SNPs increases (i.e., when μ(Ι) increases).

We observe that the strength of LD, however, does not affect the inference power as strong as k. Then, figure 3 illustrates the average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the number of LD pairs per SNP (i.e., [i(k)) and storage redundancy. The figure 4 illustrates the average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the LD strength between two SNPs (i.e., μ(Ι)) and storage redundancy.

P

In the Figs. 5 and 6, we show the Average{Pr(SNP i = 1 )} for varying standard deviations of k and I, and with 500% storage redundancy as follows: (i) in Fig. 5, we vary a(k) and [i( ), when μ(Ι) = 0.8 and σ(Ι) = 0.15, and (ii) in Fig. 6, we vary σ (I) and μ(Ι), when [i(k) = 2 and a(k) = 0.75. We observe that the inference power of the curious party varies (either increases or decreases) with an increasing value of a(k) (σ(Ι)) depending on [i(k) (μ(Ι)), and, as expected, all curves converge to a single value for higher values of a(k) (σ(Ι)).

Next, considering the individual characteristics of the real SNPs (i.e., their severity levels), we analyze the level of genomic privacy of a patient against a curious party at the SPU. By the level of genomic privacy, we understand the level of information that a third party can infer about the real variants of a

J* patient. The severity of a SNP i can be defined as the privacy-sensitivity of the SNP when SNP i = 1 (i.e. , when it exists as a variant at the patient P). For example, a real SNP revealing the predisposition of a patient for Alzheimer's disease can be considered more severe than another real SNP revealing his predisposition to a more benign disease. Severity values of the SNPs are determined as a result of medical studies (depending on their contributions to various diseases) and tables of disease severities provided by insurance companies (e.g ., percentage of invalidity). We denote the severity of a real SNP i as Vi, and 0≤ Vi≤ 1 (1 denotes the highest severity). Thus, we define the genomic privacy of the patient P as below : p = -∑log 2 (PrOSNif = l)) x ^ . (8)

We do not use the traditional entropy metric [37, 38] to quantify privacy, as only one state of

SNF poses privacy risks (i.e. SNFf = 1 ), as discussed before.

First, we study the relationship between the storage redundancy and the severity of the real SNPs by focusing on three types of patients: (i) patient A, carrying mostly low severity real SNPs (in Y A ), (ii) patient B, carrying mostly high severity real SNPs (in Y B ), and (iii) patient C, carrying mixed severity real SNPs (in Y c ). For each patient, the highest level of privacy is achieved when the storage redundancy is maximum (as in Method 1 in Section 2.3). Thus, we recognize this level as 100% genomic privacy for the patient. For the evaluation, we take the highest privacy level of patient C as the base and normalize everything with respect to this value. We use the following parameters for the simulation. The severities of patient A's and patient B's real SNPs are represented as truncated Gaussian random variables with (μΑ, oA) = (0.25, 0.15) and (μΒ, oB) = (0.75, 0.15), respectively. Furthermore, the severity of patient C's real SNPs are represented as a uniform distribution between 0 and 1 . We also set μ(Ι) = 0.8, σ(Ι) = 0.25, i(W) = 2, and o(k) = 0.75. In Fig. 7, we illustrate the increase in privacy with increments in the storage redundancy for these three types of patients (A, B, and C). We observe that by increasing the storage redundancy, a patient with high severity real SNPs gains more privacy than a patient with lower severity real SNPs, hence the storage redundancy can be customized for each patient differently based on the types of his real SNPs. It can be argued that the amount of storage redundancy for a patient can leak information (to the curious party the SPU) about the severities of his real SNPs. However, the severity of the SNPs is not the only criteria to determine the storage redundancy for a desired level of genomic privacy as we discuss next.

Finally, we study the relationship between the severity of the real SNPs, the number of LD pairs per SNP (number of SNPs with which a particular SNP has LD, i.e., k), and the storage redundancy. We assign the Vi values of the real SNPs (in Y ) following a uniform distribution between 0 and 1. We set the LD parameters as μ(Ι) = 0.8, σ(Ι) = 0.25, [i(k) = 2, and o(k) = 1.5. Then, we observe and compare the following potential scenarios in different types of patients: (i) The real low severity SNPs of the patient (i.e., his real SNPs with low Vi values) have a higher number of LD pairs (i.e., higher k values) with respect to his high severity real SNPs (we note that, in all cases, k values are obtained from the same truncated Gaussian distribution with \i(k) = 2, and a(k) = 1.5); (ii) /( values are assigned randomly to the SNPs; and (iii) the real high severity SNPs of the patient (i.e., his real SNPs with high values) have a higher number of LD pairs (i.e., higher /( values) with respect to his low severity real SNPs. Again, we set a patient's genomic privacy to 100% when the storage redundancy is maximum at the SPU (as in Method 1 in Section 2.3). We illustrate our results in Fig. 8, and show different storage redundancy requirements for different types of patients (to provide the same level of privacy). For example, to achieve 40% genomic privacy, the SPU requires 400% storage redundancy for a patient whose less severe real SNPs have more LD pairs, whereas it requires 600% storage redundancy for another patient whose more severe real SNPs have more LD pairs (which means more storage per patient, as discussed in Section 3.1 ). This result also supports our belief to customize the storage redundancy for each patient.

We obtained similar patterns for further variations of the variables but we do not present these results due to the space limitation. In summary, depending on the actual [i(k), o(k), μ(Ι), σ(Ι), and values (which will be determined as a result of the medical research), the storage redundancy can be determined (and customized for each patient based on the types of his variations) for this approach to keep the genomic privacy of the patient at a desired level. Note that the curious party at the SPU cannot infer the real SNPs of the patient (or the severities of the patient's real SNPs) from the amount of customized storage redundancy, because the storage redundancy (for a desired level of genomic privacy) depends on various factors. For example, a patient with low storage redundancy (for a desired level of genomic privacy) could mean that (i) he carries mostly low severity real SNP (as in Fig. 7), (ii) he carries mixed severity real SNPs, but his less severe real SNPs have more LD pairs (as in Fig. 8), (iii) his real SNPs (regardless of their severities) have low number of LD pairs (as in Fig. 3), or (iv) his real SNPs (regardless of their severities) have low LD strengths (as in Fig. 4).

2.5 Method 3: Encrypted Locations at the SPU

Let if = {L i : i e Y p } represent the set of locations (on the DNA sequence) of the patient P's real SNPs (in Y ). As opposed to the previous two approaches, here, we propose to encrypt the locations of the SNPs along with their contents. By doing so, we save storage costs by storing only the real SNPs in Y at the SPU (around 4 million) while providing the highest level of privacy (as in Section

2.3). These benefits, however, come with a cost in the practicality of the algorithm, introducing extra steps for the patient (P) during the protocol. Although we can assume that these extra steps can easily be handled via the patient's device such as smart card or mobile device, this approach still requires more message exchanges (as will be described next) between the parties, compared to the previous two approaches.

In some environments, dividing the weak private of the patient, and distributing two shares of the weak private key to the SPU and MU might not be acceptable (e.g., when it is likely that the SPU and MU will collaborate to retrieve patient's weak private). Therefore, for the sake of completeness, in the following, we present Method 3 with and without proxy encryption (i.e., without distributing the patient's private key to other parties). The Method 1 and Method 2 can also be modified similarly to avoid proxy encryption.

2.5.1 With Proxy Encryption

The initial steps of the protocol are the same as in Section 2.3, except for Steps 2 and 3 in which the locations of the SNPs are encrypted and a Bloom filter [39] is generated. Below, we summarize the different steps of this approach (the unchanged steps are not repeated). These steps are illustrated in Fig. 9.

• Step 2: The Certified Institution (CI) first determines the locations of P's real SNPs (in Y p ) and constructs L P . Then, the CI constructs a Bloom filter using the elements of i as inputs.

A Bloom filter is a simple space-efficient randomized data structure for representing a set in order to support membership queries [39]. A Bloom filter for representing a set L P is described by an array of κ bits, initially all set to 0. It employs independent hash functions H l 5 ..., H with range {1 , . . . , κ}.

For every element L i e L p , the bits H^L,.), ..., H in the array are set to 1. A location can be set to 1 multiple times, but only the first change has an effect.

After constructing the Bloom filter, the CI encrypts each element in L P by using a symmetric key shared between the CI and P (established during Step 0 as in Section 2.3) and generates

L P E = {E{L t ) : i G Y p } . The CI also encrypts a dummy variant (representing the potential SNPs in

Ω ) along with the real SNPs of the patient (using P's public key). Furthermore, the CI associates a dummy position ^for this dummy variant and encrypts using the symmetric key between the CI and P to obtain the encrypted dummy position E^).

• Step 3: The CI sends the constructed Bloom filter and the encrypted dummy position E^) to the patient for storage into the patient device, and encrypted SNPs and locations to the SPU.

• Step 6: The MU tells the patient the locations of the SNPs that are required for the susceptibility test or requested directly as the relevant SNPs. • Step 7: The patient inputs each requested location L. to the Bloom filter to determine if the corresponding location is stored at his Bloom filter (i.e., to determine if he has a real SNP at the corresponding location).

To check if L. belongs to L r , the patient checks whether all H ^L■), ... , Ή. (L .) are set to 1. If not, L j definitely does not belong to L P . Otherwise, the patient assumes L j e , although this may be wrong with some probability. That is, a Bloom filter could yield a false positive, where it suggests that L. is in L p even though it is not. This probability can be decreased at the expense of increasing

Bloom filter length (i.e., κ ). Further, the false positive probability can be significantly reduced by using some proposed techniques such as [40, 41 ]. As a result of this process

(a) If the location is in his Bloom filter (i.e., if he has a real SNP at the corresponding location), P encrypts the location with the symmetric key between the CI and P.

(b) If the location is not in his Bloom filter (i.e., if he does not have a real SNP at the corresponding location), P uses E^) as the encrypted location.

We note that the above operations can be easily done via the patient's device (e.g., by reading the patient's device at the MU as a consent to the test) or mobile device (e.g., by consenting via a smart phone application) by using the stored Bloom filter output, E^), and symmetric key between the CI and P.

• Step 8: The patient sends the SPU the encrypted locations of the SNPs which will be provided to the MU.

· Step 9: The encrypted SNPs are sent to the MU in the same order as they are requested in Step 6.

(a) If only the end-result is requested, the corresponding SNPs are re-encrypted at the SPU under the patient's public key (re-encryption under the same public key is discusses in Section 2.1 ). As there is only one value stored at the SPU representing the contents of the potential SNPs at which P does not have a variant (at position E^)), this value is re-encrypted for each different request of a non-variant, so that the MU cannot infer the locations of the non-variants of the patient.

(b) If relevant SNPs are requested, the SPU partially decrypts the relevant SNPs by using a part of P's private key following a proxy encryption protocol (Section 2.1 ).

• Step 10: Re-encrypted (or partially decrypted) SNPs are sent to the MU by the SPU.

• Step 11 : One of the following two scenarios occur at the MU: (a) If only the end-result is requested, the MU computes P's total susceptibility to disease X by using the homomorphic properties of the

Paillier cryptosystem (similar to the discussion in Section 2.6) under the patient's public key. Although the discussion in Section 2.6 is held considering Method 1 (or Method 2), a similar technique is used for this approach at the MU, hence we do not discuss it again.

(b) If relevant SNPs are requested, the MU decrypts the message received from the SPU by using the other part of P's private key and recovers the relevant SNPs.

• Step 12: The MU sends the encrypted end-result to the SPU. • Step 13: The SPU partially decrypts the end-result using a part of P's private key by following a proxy encryption protocol (Section 2.1 ) and sends it back to the MU.

• Step 14: The MU decrypts the message received from the SPU by using the other part of P's private key and recovers the end-result.

2.5.2 Without Proxy Encryption

In this approach, the SPU stores only the encrypted SNPs and encrypted locations. Genomic data encrypted by P's public key is only decrypted at P, and the weak private key of P remains only at P (i.e., shares of the weak private key are not distributed to the SPU or MU). Most of this approach is the same as Method 3 with proxy encryption. Indeed, the first 8 steps of the algorithm are the same, except for the distribution of parts of P's private key. The only difference is the transfer of the end- result or the relevant SNPs to the MU as follows:

• If the relevant SNPs are requested by the MU, the SPU sends the encrypted SNPs (by P's public key) to P. P decrypts these SNPs (using his weak private key) and sends them to the MU.

• If the end-result of the susceptibility test is requested by the MU, the disease-susceptibility test is done (via homomorphic operations) at the MU and the encrypted end-result is sent to P. Then, P decrypts the end-result and sends it back to the MU.

We note that the security of the communication between P and the MU is provided by symmetric keys as discussed before. The above operations put some more burdens on the patient during the protocol. However, we emphasize that these operations can be smoothly done on the patient's device without requiring a substantial effort from the patient himself.

In summary, as the locations of the real SNPs are encrypted, a curious party at the SPU cannot infer the contents of the SNPs from their locations (as in Section 2.3), hence it is enough to store only the real SNPs in Y . Furthermore, the privacy provided by this approach (with or without proxy encryption) is the same as 900% redundancy in Method 2 (i.e., similar to Method 1 ), hence we do not discuss it again. Another advantage of this approach (i.e., Method 3 in general) is that individual contributions of the genetic variant markers remain secret at the MU, because the homomorphic operations are conducted at the MU. This advantage might become more significant when this approach is used for personalized medicine methods in which the pharmaceutical company

(embodied in this case as the medical unit) does not want to reveal the genetic properties of its drugs. Thus, if introducing the described extra steps for the patient and few additional message exchanges between the parties are tolerated, this approach operates with relatively modest storage and yet provides very good privacy.

2.6 Computing Disease Susceptibility via Homomorphic Operations

We now present the disease-susceptibility test via homomorphic operations at the SPU for Method 1 (Section 2.3) and Method 2 (Section 2.4). Similar techniques can be used for Method 3 at the MU, as discussed in Section 2.5.

The SPU uses a proper function to compute P's predicted disease susceptibility via homomorphic encryption. There are different functions for computing the predicted susceptibility. In [25], focusing on one example of many diseases that require a susceptibility test involving multiple SNPs, Kathiresan ei al. propose to count the number of unfavorable alleles carried by the patient for each SNP related to a particular disease. Similarly, in [26], Ashley ei al. propose to multiply the Likelihood Ratios (LRs) of the most important SNPs for a particular disease in order to compute a patient's predicted susceptibility. LR values are determined as a result of medical studies. Furthermore, a weighted a veraging function can also be used, which computes the predicted susceptibility by weighting the contributions of SNPs by their contributions (e.g., LR values of the SNPs). Note that our proposed privacy-preserving mechanisms are not limited by the types of the functions (used to test the disease susceptibility). It is expected that these functions will evolve over time; hence the proposed algorithms can be developed to keep up with this evolution.

In the following, we discuss how to compute the predicted disease susceptibility at the SPU by using a toy example to show how the homomorphic encryption is used at the SPU. Initially, we assume that the function at the SPU is weighted averaging (which is an advanced version of the function proposed in [25]) and show how the predicted susceptibility is computed using encrypted SNPs. Then, we show how the function proposed in [26] (i.e., multiplication of LR values) can be utilized at the SPU.

2.6.1 Weighted Averaging

Assume that (for simplicity) the susceptibility to disease X is determined by the set of SNPs Ω =

P P

{SNP m ,SNP„}, which occur at particular locations of the DNA sequence. SNP?« and SNPn are not necessarily among the real SNPs of the patient P (i.e., P does not need to have a variant at those locations). The contributions of different states of SNP for i e {m, ) to the susceptibility to disease X are computed via previous studies (on case and control populations) and they are already known by the MU. That is, pj, ( ) @Pr( | S N P P = 0) and pj ( ) @Pr( | S N P P = 1) (/ ' e {m, n}) are determined and known by the MU. Further, the contribution (e.g., LR value) of SNP, to the susceptibility to disease X is denoted by Cf . Note that these contributions are also computed by previous studies on case and control groups and they are known by the MU.

As we have discussed before, the SPU stores the set of SNPs of the patient P, encrypted by P's public key (n, g, h = g x ). Encryption is done using the modified Paillier cryptosystem as discussed in

Section 2.1. Thus, the SPU uses E(SNP P , g x ) and E(SNP P , g x ) for the computation of predicted susceptibility of P to disease X. From now on, we drop the r values in the above encrypted messages for the clarity of the presentation (r values are chosen randomly from the set [1 , n/4] for every encrypted message as discussed in Section 2.1 ). Similarly, the MU provides the following to the SPU in plaintext: (i) the markers for disease X (SNP m and SNP„), (ii) corresponding probabilities p j ' (X) , i e |m,«} and j e {θ, ΐ} , and (iii) the contributions of each SNP Cf .

Next, the SPU encrypts j(j e {0, l}) using P's public key to obtain E(0, g x ) and E(1 , g x ) for the homomorphic computations. This encryption can also be done at the MU and sent to the SPU.

Alternatively, we might assume that SNPs of a patient are stored at the SPU in pairs of {E(|SNP - 0\, g x ), EflSNPf - 1 |, g x )} for each SNP , instead of the actual values of the SNPs. In this case, the above encryption at the SPU would not be required.

The SPU computes the predicted susceptibility of the patient P to disease X by using weighted averaging.

This can be com uted in plaintext as below:

The computation in (9) can be realized using the encrypted SNPs of the patient (and utilizing the homomorphic properties of the Paillier cryptosystem) to compute the encrypted disease susceptibility,

E( S^ , g x ) as below:

E(si,g") : Π x E(SNF ,g * )E(0,g")

(10) where

Δ! (1 1a) Δ (1 1 b) Θ (1 1 c)

0 - 1 1 - 0 c x + c x

We note that the end-result in (10) is encrypted by P's public key.

Then, the SPU partially decrypts the end-result E( S X , g x ) using its share (x (1) ) of P's private key

< 2 >

(x) as discussed in Section 2.1 to obtain E( S p , ) and sends it to the MU. Finally, the MU

< 2 >

decrypts E( S x , ) using its share (x (2) ) of P's private key to recover the end-result S X

In some genetic tests, the types of the real SNPs (e.g., homozygous or heterozygous) become also important. In this case, SNP can take three different values from the set {0, 1 , 2} to represent a potential SNP (i.e., nonvariant), a real homozygous SNP, and a real heterozygous SNP, respectively. In such a scenario, to conduct the disease-susceptibility test via homomorphic operations, the SPU should store the squared values of the SNPs. That is, for each SNP of the patient P, the SPU should store E((SNP †, g x ). Depending on the types of genomic tests that would be supported by the SPU

(and the functions required for these tests), the format of storage of patient's SNPs can be determined beforehand, and SNPs can be stored accordingly just after the sequencing process.

2.6.2 Likelihood Ratio Test

We now assume that the predicted disease susceptibility is computed from the multiplication of Likelihood Ratios (LRs) of the corresponding SNPs as in [26] and show how such a computation would be handled at the SPU by using homomorphic operations.

In this approach, the predicted disease susceptibility is computed by multiplying the initial risk of the patient (e.g., for disease X) by the LR value of each SNP related to that disease (LR value of a SNP ; ' depends on the value of SNP at the patient P). The initial risk of the patient P for the disease X is represented as f x . We note that f x is determined by considering several factors (other than patient's genomic data) such as patient's age, gender, height, weight, and environment. Thus, this initial risk can be computed directly by the MU. We also note that if the LR value corresponding to a particular SNP is less than one, the risk for the disease decreases. Otherwise, if the LR value is greater than one, the risk increases for the corresponding disease.

Similar to before, we assume that the susceptibility to disease X is determined by the set of SNPs in Ω = {SNP m ,SNP„}. We denote the LR values due to SNP = 0 and SNP = 1 for disease as L' x (0) and li x (1 ), respectively.

The SPU stores the SNPs of the patient P, encrypted by P's public key. The MU sends the following to the SPU: (i) V x (J) values ( i e |m,«} and j e {0, 1} ) in plaintext, and (ii) the markers for disease X.

The MU also encrypts the log of initial risk value, ln(/f ), by P's public key and sends E(ln(/f ), g x ) to the SPU. Alternatively, the contribution of the initial risk to the disease susceptibility can be included to the end-result at the end, at the MU.

The Paillier cryptosystem does not support multiplicative homomorphism in ciphertext (it only supports the multiplication of a ciphertext with a constant as discussed in Section 2.1 ). Thus, instead of multiplying the LR values, we propose using addition in log-domain at the SPU. Thus, the SPU computes the predicted susceptibility of P to disease X as below:

E (ln(S£, g' ) = E (ln(/ , g * )xn E (S P ) - E ( * ) _1 " x E (SNP/ ) - E (O, ^ ) (12)

where

Ξ , = |n ( Χ (°) η ) (13a) Ξ , = |n ( Χ ( ! ) η ) (13b)

(0 - 1) ' (1 - 0)

We note that 12) corresponds to the below computation in plaintext:

/ r ln (L ! „(l))]

ln(s?) [SNP/' - Oj x ■ (14)

As before, the SPU partially decrypts E(ln(SX P ), g x ) using x(1 ) (its share of P's private key) to obtain E(ln( S p ), g ) and sends it to the MU. Finally, the MU decrypts E(ln( ), 8 ) using x (2) (its share

(in(Sp ))

of P's private key) to recover ln( S p ), and computes e ' to obtain S p . Similar to weighted averaging, if the types of the real SNPs are used for the test (in which there are three possible states for SNP ), squared values of the SNPs should be stored at the SPU for each patient.

3 Evaluation and Implementation of the Proposed Methods In Fig. 10, based on the discussion in the previous sections, we graphically compare the proposed methods considering the level of privacy they provide, their practicality (for the patient), and their storage requirements (at the SPU). In this section, we report our findings about the complexity and security of the proposed methods.

3.1 Implementation and Complexity Evaluation

To evaluate the practicality of the proposed privacy-preserving algorithms, we implemented them, and assessed their storage requirements and computational complexities on Intel Core i7-2620M CPU with 2.70 GHz processor under Windows 7 Enterprise 64-bit Operating System. We set the size of the security parameter (n in Paillier cryptosystem in Section 2.1 ) to 1024 bits. We computed the disease susceptibility using weighted averaging (at the SPU or MU, see Section 2.6.1 as well as LR test in Section 2.6.2 which also has similar complexity) and real SNP profiles from [42]. Our implementation relies on a MySQL 5.5 database managed by the open source tool MySQL Workbench. To provide a platform-independent implementation, we used the Java programming language along with the open- source Integrated Development Environment, NetBeans IDE 7.1 .1 ., for the implementation of the Java code. We note that our code for the implementation is not optimized, and better results can be expected with an optimized implementation.

In Table I I, we summarize the computational and storage complexities of the proposed methods at (i) Certified Institution (CI), (ii) SPU, (iii) MU, and (iv) P. We evaluate the proposed methods considering the following costs: (i) encryption of patient's variants, (ii) disease-susceptibility test at the SPU via homomorphic operations (using ten variants), (iii) decryption of the end-result (or relevant SNPs), (iv) proxy encryption, and (v) storage costs, in which Θ represent the percentage of storage redundancy at

the SPU. We did not explicitly implement the Bloom filter (for Method 3) and symmetric

encryption/decryption between the parties for the security of the communication. However, the computational costs due to these operations are negligible compared to Paillier encryption/decryption and homomorphic operations.

We emphasize that the encryption of the variants at the CI is a one-time operation and is significantly faster than the sequencing and analysis of the sequence (which takes days). Further, this encryption can be conducted much more efficiently by computing some parameters, such as ( g r , W ) pairs, offline for various r values, for each patient. Indeed, by computing (g r , W ) pairs offline, we observe that the encryption takes only 0.017 ms per variant at the CI. Method 1 and Method 2

Table 2: Computational and Storage Complexities of the Proposed Methods

It is also possible to conduct private statistical tests (by a medical researcher) on the data stored at the SPU in order to get statistics about the variants of multiple patients. Conducting such a statistical test for a variant (about its type) on 100K patients takes around 55 minutes at the SPU and scales linearly with the number of patients. Note that such a statistical test is only possible with Method 1 or Method 2; using Method 3 and querying the encrypted locations of SNPs from 100K patients is not practical for this application.

In summary, all these numbers show the practicality of our privacy-preserving algorithms.

3.2 Security Evaluation

The proposed schemes preserve the privacy of patients' genomic data relying on the security strength of modified Paillier cryptosystem (in Section 2.1 ). The extensive security evaluation of the modified Paillier cryptosystem can be found in [33]. Below we summarize two important security features of this cryptosystem.

· One-wayness: This property means that no efficient adversary has any significant chance of finding a preimage to the ciphertext when he sees only the ciphertext and the public key of the patient. It is shown in [33] that the one-wayness of the modified Paillier cryptosystem can be related to the Lift Diffie-Hellman problem which is shown to be as hard as the partial Discrete Logarithm problem.

• Semantic security: This property ensures that an adversary will be unable to distinguish pairs of ciphertexts based on the message they encrypt. It is shown in [33] that if Decisional Diffie-Hellman Assumption (a computational hardness assumption about a certain problem involving discrete logarithms in cyclic groups) in Z * 2 holds, then the modified Paillier cryptosystem is semantically secure.

Finally, if the weak private key of the patient, x, is randomly divided and distributed to the Storage and Processing Unit (SPU) and Medical unit (MU) as in Method 1 , this weak private key could be revealed if the MU colludes with the SPU, but the factors n, p, and q remain secret. We note that such a collusion is not considered in this study. However, for the sake of completeness, in Section 2.5.2, we present an alternative approach (Method 3 without proxy encryption) that avoids distributing the patient's weak private key to other parties, hence is robust against such a collusion.

References [I] A. Cavoukian, "Privacy by design," 2009, http://www.ontla.on.ca/library/repository/mon/23002/ 289982.pdf.

[2] S. F. Gurses, "Multilateral privacy requirements analysis in online social network services," 2010, PhD thesis, KU Leuven.

[3] M. Langheinrich, "Principles of privacy-aware ubiquitous systems," Proceedings of Ubiquitous Computing (UbiComp), 2001.

[4] G. van Blarkom, J. Borking, and J. Oik, "Handbook of privacy and privacy-enhancing technologies (the case of intelligent software agents)," College bescherming persoonsgegevens, 2003.

[5] http://www.personalgenomes.org/consent/PGP Consent Approved 02212012.pdf.

[6] J. R. Troncoso-Pastoriza, S. Katzenbeisser, and M. Celik, "Privacy preserving error resilient DNA searching through oblivious automata," CCS Ό7: Proceedings of the 14th ACM Conference on Computer and Communications Security, pp. 519-528, 2007.

[7] M. Blanton and M. Aliasgari, "Secure outsourcing of DNA searching via finite automata," DBSec'10: Proceedings of the 24th Annual IFIP WG 11.3 Working Conference on Data and Applications Security and Privacy, pp. 49-64, 2010.

[8] S. Jha, L. Kruger, and V. Shmatikov, "Towards practical privacy for genomic computation," Proceedings of the 2008 IEEE Symposium on Security and Privacy, pp. 216-230, 2008.

[9] F. Bruekers, S. Katzenbeisser, K. Kursawe, and P. Tuyls, "Privacy-preserving matching of DNA profiles," Tech. Rep., 2008.

[10] M. Kantarcioglu, W. Jiang, Y. Liu, and B. Malin, "A cryptographic approach to securely share and query genomic sequences," IEEE Transactions on Information Technology in Biomedicine, vol. 12, no. 5, pp.606-617, 2008.

[I I] P. Baldi, R. Baronio, E. De Cristofaro, P. Gasti, and G. Tsudik, "Countering GATTACA: efficient and secure testing of fully-sequenced human genomes," CCS '11 : Proceedings of the 18th ACM Conference on Computer and Communications Security, pp. 691-702, 2011.

[12] M. Canim, M. Kantarcioglu, and B. Malin, "Secure management of biomedical data with cryptographic hardware," IEEE Transactions on Information Technology in Biomedicine, vol. 16, no. 1 , 2012.

[13] D. Eppstein, M. T. Goodrich, and P. Baldi, "Privacy-enhanced methods for comparing

compressed DNA sequences," CoRR, vol. abs/1107.3593, 2011. [Online]. Available:

http://arxiv.Org/abs/1107.3593

[14] D. Eppstein and M. T. Goodrich, "Straggler identification in round-trip data streams via Newton's identities and invertible Bloom filters," IEEE Transactions on Knowledge and Data Engineering, vol. 23, no. 2, pp. 297-306, 2011.

[15] R. Wang, Y. F. Li, X. Wang, H. Tang, and X. Zhou, "Learning your identity and disease from research papers: information leaks in genome wide association study," CCS '09: Proceedings of the 16th ACM Conference on Computer and Communications Security, pp. 534-544, 2009. [16] B. Malin and L. Sweeney, "How (not) to protect genomic data privacy in a distributed network: using trail re-identification to evaluate and design anonymity protection systems," Journal of

Biomedical Informatics, vol. 37, pp. 179-192, Jun. 2004.

[17] N. Homer, S. Szelinger, M. Redman, D. Duggan, and W. Tembe, "Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays," PLoS Genetics, vol. 4, Aug. 2008.

[18] J. Gitschier, "Inferential genotyping of Y chromosomes in Latter-Day Saints founders and comparison to Utah samples in the HapMap project," Am. J. Hum. Genet., vol. 84, pp. 251-258, 2009.

[19] X. Zhou, B. Peng, Y. F. Li, Y. Chen, H. Tang, and X. Wang, "To release or not to release:

evaluating information leaks in aggregate human-genome data," ESORICS'1 1 : Proceedings of the 16th European Conference on Research in Computer Security, pp. 607-627, 201 1.

[20] S. E. Fienberg, A. Slavkovic, and C. Uhler, "Privacy preserving GWAS data sharing," Proceedings of the IEEE 1 1th International Conference on Data Mining Workshops (ICDMW), Dec. 201 1.

[21] Y. Chen, B. Peng, X. Wang, and H. Tang, "Large-scale privacy-preserving mapping of human genomic sequences on hybrid clouds," NDSS'12: Proceeding of the 19th Network and Distributed System Security Symposium, 2012.

[22] R. Wang, X. Wang, Z. Li, H. Tang, M. K. Reiter, and Z. Dong, "Privacy-preserving genomic computation through program specialization," Proceedings of the 16th ACM Conference on Computer and Communications Security, pp. 338-347, 2009.

[23] R. Agrawal, A. Evfimievski, and R. Srikant, "Information sharing across private databases," Proceedings of SIGMOD Conference, 2003.

[24] D. Dachman-Soled, T. Malkin, M. Raykova, and M. Yung, "Efficient robust private set intersection," Proceedings of the 7th International Conference on Applied Cryptography and Network Security, pp.125-142, 2009.

[25] S. Kathiresan, O. Melander, D. Anevski, C. Guiducci, and N. Burtt, "Polymorphisms associated with cholesterol and risk of cardiovascular events," The New England Journal of Medicine, vol. 358, pp. 1240- 1249, 2008.

[26] E. Ashley, A. Butte, M. Wheeler, R.Chen, and T. Klein, "Clinical assessment incorporating a personal genome," The Lancet, vol. 375, no. 9725, pp. 1525-1535, 2010.

[27] S. Seshadri, A. Fitzpatrick, M.A. Ikram, A. DeStefano, V. Gudnason, M. Boada, J. Bis, A. Smith, M. Carassquillo, J. Lambert, C. Consortium, G. Consortium, and E. Consortium, "Genome-wide analysis of genetic loci associated with Alzheimer disease," JAMA, vol. 303, pp. 1832-1840, 2010.

[28] http://www.ncbi.nlm.nih.gov/projects/SNP/.

[29] D. Greenbaum, A. Sboner, X. Mu, and M. Gerstein, "Genomics and privacy: Implications of the new reality of closed data for the field," PLoS Computational Biology, vol. 7, no. 12, 201 1.

[30] M. Raykova, H. Zhao, and S. M. Bellovin, "Privacy enhanced access control for outsourced data sharing," Financial Cryptography and Data Security, 2012. [31] M. T. Goodrich and M. Mitzenmacher, "Privacy-preserving access of outsourced data via oblivious RAM simulation," Proceedings of the 38th International Conference on Automata, Languages and Programming - Volume Part II, pp. 576-587, 201 1.

[32] E. Stefanov, E. Shi, and D. Song, "Towards practical oblivious RAM," NDSS'12: Proceeding of the 19th Network and Distributed System Security Symposium, 2012.

[33] E. Bresson, D. Catalano, and D. Pointcheval, "A simple public-key cryptosystem with a double trapdoor decryption mechanism and its applications," Proceedings of Asiacrypt 03, LNCS 2894, pp. 37-54, 2003.

[34] M. Pirretti, P. Traynor, P. McDaniel, and B. Waters, "Secure attribute-based systems,"

Proceedings of the 13th ACM Conference on Computer and Communications Security, pp. 99-1 12, 2006.

[35] G. Ateniese, K. Fu, M. Green, and S. Hohenberger, "Improved proxy re-encryption schemes with applications to secure distributed storage," ACM Transactions on Information and System Security, vol. 9, pp. 1-30, Feb. 2006.

[36] D. S. Falconer and T. F. Mackay, Introduction to Quantitative Genetics (4th Edition). Harlow, Essex, UK: Addison Wesley Longman, 1996.

[37] C. Diaz, S. Seys, J. Claessens, and B. Preneel, "Towards measuring anonymity," Proceedings of Privacy Enhancing Technologies Symposium (PETS), 2002.

[38] A. Serjantov and G. Danezis, "Towards an information theoretic metric for anonymity,"

Prioceedings of Privacy Enhancing Technologies Symposium (PETS), 2002.

[39] B. H. Bloom, "Space/time trade-offs in hash coding with allowable errors," ACM Communications, vol. 13, no. 7, pp. 422-426, 1970.

[40] F. Hao, M. Kodialam, and T. V. Lakshman, "Building high accuracy Bloom filters using partitioned hashing," Proceedings of ACM International Conference on Measurement and Modeling of Computer Systems, pp. 277-288, 2007.

[41] P. S. Almeida, C. Baquero, N. Pregui ^ ca, and D. Hutchison, "Scalable Bloom filters," Information Pro- cessing Letters, vol. 101 , no. 6, pp. 255-261 , 2007.

[42] The 1000 Genomes Project Consortium, "A map of human genome variation from population- scale sequencing," Nature, vol. 467, pp. 1061-1073, 2010.