What is differential privacy and how can it protect consumer data?

What is differential privacy and how can it protect consumer data?


Play all audios:

Loading...

It’s no secret that big tech companies like Facebook, Google, Apple and Amazon are increasingly infiltrating our personal and social interactions to collect vast amounts of data on us every


day. At the same time, privacy violations in cyberspace regularly make front page news. So how should privacy be protected in a world where data is gathered and shared with increasing speed


and ingenuity? Differential privacy is a new model of cyber security that proponents claim can protect personal data far better than traditional methods. The maths it is based on was


developed 10 years ago, and the method has been adopted by Apple and Google in recent years. ADVERTISEMENT WHAT IS DIFFERENTIAL PRIVACY? Differential privacy makes it possible for tech


companies to collect and share aggregate information about user habits, while maintaining the privacy of individual users. For example, say you wanted to show the most popular routes people


take walking through a park. You track the routes of 100 people who regularly walk through the park, and whether they walk on the path or through the grass. But instead of sharing the


specific people taking each route, you share the aggregate data collected over time. People viewing your results might know that 60 out of 100 people prefer to take a short-cut through the


grass, but not which 60 people. WHY DO WE NEED IT? Many of the world’s governments have strict policies about how tech companies collect and share user data. Companies who do not follow the


rules can face huge fines. A Belgian court recently ordered Facebook to stop collecting data on users’ browsing habits on external websites, or face fines of €250,000 a day. For many


companies, especially multinationals operating in different jurisdictions, this leaves them in a delicate position when it comes to the collection and use of customer data. On the one hand,


these companies need users’ data so they can provide high-quality services that benefit users, such as personalised recommendations. On the other hand, they may face charges if they collect


too much user data, or if they try to move data from one jurisdiction to another. Traditional privacy-preserving tools such as cryptography can’t resolve this dilemma since it prevents tech


companies from accessing the data at all. And anonymity reduces the value of data – an algorithm can’t serve you personalised recommendations if it doesn’t know what your habits are. HOW


DOES IT WORK? Let’s continue the example of walking routes through a park. If you know the identities of those included in the study, but you don’t know who took which route then you might


assume that privacy is protected. But that may not be the case. Say someone viewing your data wants to ascertain if Bob prefers to walk through the grass or on the path. They have obtained


background information about the other 99 people in the study, which tells them that 40 people prefer to walk on the path and 59 prefer to walk through the grass. Therefore, they can deduce


that Bob, who is the 100th person in the database, is the 60th person who prefers to walk through the grass. This type of attack is a called a differentiated attack, and it is quite hard to


defend against as you cannot control how much background knowledge someone can obtain. Differential privacy aims to defend against this type of attack. Someone deducing your walking route


might not sound too serious, but if you replace walking routes with HIV test results, then you can see there is potential for a serious invasion of privacy. The differential privacy model


guarantees that even if someone has complete information about 99 of 100 people in a data set, they still cannot deduce the information about the final person. The primary mechanism to


achieve that is to add random noise to the aggregate data. In the path example, you may say the number of people who prefer to cross the grass is 59 or 61, rather than exact number of 60.


The inaccurate number can preserve the privacy of Bob, but it will have very little impact on the pattern: around 60% people prefer to take a short-cut. Apple emojis The noise is carefully


designed. When Apple employed differential privacy in iOS 10, it added noise to individual user inputs. That means it can track, for example, the most frequently used emojis, but the emoji


usage of any individual user is masked. Cynthia Dwork, the inventor of the differential privacy, has proposed wonderful mathematical proofs on how much noise is enough to achieve the


requirement of differential privacy. WHAT ARE ITS PRACTICAL APPLICATIONS? Differential privacy can be applied to everything from recommendation systems to location-based services and social


networks. Apple uses differential privacy to gather anonymous usage insights from devices like iPhones, iPads and Macs. The method is user-friendly, and legally in the clear. Differential


privacy would also allow a company like Amazon access your personalised shopping preferences while hiding sensitive information about your historical purchase list. Facebook could use it to


collect behavioural data for targeted advertising, without violating a country’s privacy policies. HOW COULD IT BE USED IN THE FUTURE? Different countries have varying privacy policies, and


sensitive documents currently have to be manually checked before they move from one country to another. This is time-consuming and expensive. Recently, a team from Deakin University


developed differential privacy technology to automate privacy processes within cloud-sharing communities across countries. They propose using mathematical formulas to model the privacy laws


of each country that could be translated to “middleware” (software) to ensure the data conforms. Employing differential privacy in this way could protect the privacy of users and resolve a


data sharing headache for tech companies. _TIANQING ZHU IS A LECTURER IN CYBER SECURITY, FACULTY OF SCIENCE, ENGINEERING & BUILT ENVIRONMENT AT DEAKIN UNIVERSITY. THIS ARTICLE WAS


ORIGINALLY PUBLISHED ON THE CONVERSATION. READ THE ORIGINAL ARTICLE._