Privacy-Preserving Machine Learning: A Primer
I have a bit of a cybersecurity background, so a year or two ago I started paying attention to how often data breaches happen - and I noticed something depressing: they happen literally 3+ times each day on average in the United States alone! (For a particularly sobering review check out this blog post.) All that information, your information, is out there, floating through the ether of the internet, along with much of my information: email addresses, old passwords, and phone numbers.
Governments, countries, and individuals - with good reason - are taking notice and making changes. The GDPR was groundbreaking in its comprehension, but hardly alone in its goal to shape policy surrounding individual data rights. More policies will pass as the notions of individual privacy and data rights are explored, defined, and refined.
Reconstruction Attacks
A reconstruction attack is when an adversary gains access to the feature vectors that were used to train the ML model and uses them to reconstruct the raw private data (i.e., your geographical location, age, sex, or email address). For this attack to work, the adversary must have white-box access to the ML model; that is, the internal state of the system can be easily understood or ascertained. Consider, for example, that some ML models (SVMs, KNNs) store the feature vectors as part of the model itself. If an attacker were to gain access to the model they would also have access to the feature vectors, and could very likely reconstruct the raw private data. Yikes!
Model Inversion Attacks
A model inversion attack is one step removed from the above. In this attack, the adversary only has access to the results produced from the trained ML model, not the feature vectors. However, many ML models report not only an “answer” (label, prediction, etc.) but a confidence score (probability, SVM decision value, etc.). For example, an image classifier might categorize an image as containing a panda with a confidence of .98. This additional piece of information - the confidence - opens the door for a model inversion attack. The adversary can now feed many examples into your trained model, record the output label and confidence, and use these to glean what feature vectors the original model was trained on in order to produce those results.
Membership Inference Attacks
This is a particularly pernicious attack. Even if you keep the original data and the feature vectors completely under lock and key (e.g., encrypted), and only report predictions without confidence scores, you’re still not 100% safe from a membership inference attack. In this attack, the adversary has a dataset and seeks to learn which of the samples might have been used to train the target ML model by comparing its output with the label associated with that sample. This type of attack is illustrated in the dashed box in the figure above. Successful attempts allow the attacker to learn which individual records were used to train the model, and thus identify an individual’s private data.
Anonymization is a privacy technique that sanitizes data by removing all personally identifying information before releasing the dataset for public use. Unfortunately, this technique also suffers from membership inference attacks as was demonstrated in the now-infamous Netflix case. The attack is the same as in the ML case: in that instance, researchers used the open-source internet movie database (IMBD) to infer the identities of the anonymized Netflix users.
Privacy-Preserving Solutions
So what are responsible machine learning enthusiasts and practitioners to do? Give up and go home? Obviously not. It’s time to present some solutions! While some of these ideas aren’t particularly new (for example, differential privacy has been around for at least 13 years), developing practical ML applications with these privacy techniques is still a nascent area of research. And while there is not yet a one-stop-shop for ML privacy on the market, this is definitely a space to follow as developments come quickly.
Secret Sharing and Secure Multi-Party Computation
Secret sharing is a method of distributing a secret without exposing it to anyone else. At its most basic, this scheme works by assigning a “share” of the secret to each party. Only by pooling their shares together can they reveal the secret. This idea has been developed into a framework of protocols that make secure multi-party computation (SMPC) possible. In this paradigm, the “secret” is an entity’s private data which is typically secured cryptographically. SMPC then allows multiple entities (banks, hospitals, individuals) to pool their encrypted data together for training an ML model without exposing their data to other members of the party. This type of protection strongly protects against reconstruction attacks.
Get Connected Here: =============
Website Link: https:// databasescientist.org/
Nomination Link: https:// databasescientist.org/award- nomination/?ecategory=Awards& rcategory=Awardee
Contact Us For Enquiry: contact@ databasescientist.org
Social media=======
Youtube: https://www.youtube. com/@databasescientist
Instagram: https://www. instagram.com/ databasescientist123/
Pinterest: https://in. pinterest.com/ databasescientist/
Blogger: https://www.blogger. com/blog/posts/ 1267729159104340550
#DatabaseScience #DataManagement #DatabaseExpert #DataProfessional #DatabaseDesign #DataArchitecture #DatabaseDevelopment #DataSpecialist #DatabaseAdministration #DataEngineer #DatabaseProfessional #DataAnalyst #DatabaseArchitect #DataScientist #DatabaseSecurity #DataStorage #DatabaseSolutions #DataManagementSolutions #DatabaseInnovation #DataExpertise
Website Link: https://
Nomination Link: https://
Contact Us For Enquiry: contact@
Social media=======
Youtube: https://www.youtube.
Instagram: https://www.
Pinterest: https://in.
Blogger: https://www.blogger.
#DatabaseScience #DataManagement #DatabaseExpert #DataProfessional #DatabaseDesign #DataArchitecture #DatabaseDevelopment #DataSpecialist #DatabaseAdministration #DataEngineer #DatabaseProfessional #DataAnalyst #DatabaseArchitect #DataScientist #DatabaseSecurity #DataStorage #DatabaseSolutions #DataManagementSolutions #DatabaseInnovation #DataExpertise
Comments
Post a Comment