Powered by OpenAIRE graph
Found an issue? Give us feedback

Stewardship of Privacy-Enhancing Technologies for Scientific Research and Policymaking

Funder: UK Research and InnovationProject code: MR/Y015711/1
Funded under: UKRI FLF Funder Contribution: 1,579,590 GBP

Stewardship of Privacy-Enhancing Technologies for Scientific Research and Policymaking

Description

We generate vast amounts of data concerning our health, movements, and habits when interacting with technology and digital services. These digital traces are a vital key to solving society's biggest problems-for example, electronic health records can support cancer surveillance efforts, and mobile location data can support humanitarian action for disaster relief. While privacy researchers have proposed numerous techniques to safely collect, analyse, and share personal data, these systems are not without their limits. Indeed, a number of supposedly anonymous datasets have been re-identified, and a lack of public confidence derailed the NHS's care.data and GP data collection scheme that tried to share de-identified health data for research. To address the privacy threats involved in releasing sensitive human data, regulators have advocated for use of modern privacy-enhancing technologies (PETs) that have stronger privacy guarantees. However, some PET techniques-such as injecting noise into the data, or creating 'synthetic' datasets-can fundamentally distort data in unknown but potentially harmful ways, for example if rare diseases are suppressed from synthetic data, or vulnerable communities are further marginalised. A group of 50 US academics led by Prof. Gary King recently warned the US Census Bureau that the secrecy of anonymisation techniques can lead to "biases that have never been publicly quantified". This lack of understanding of how PETs will impact research and data analysis-and the policy interventions that rely on it-complicates recent calls to "unlock the power of data" for the public good. Over the course of this Fellowship, I will provide a pathway to guarantee both the privacy of data subjects *and* the utility and integrity of research data. My proposal pioneers a statistical learning and computational approach to guide the development of fair and usable PETs, allowing regulators and civil society-for the first time-to make evidence-based determinations for which privacy mechanisms to use when collecting and releasing sensitive datasets, and researchers to independently audit the validity and integrity of any anonymised data they receive. It will pioneer computationally-heavy replication studies to understand how PETs can cause harm (WP1); statistical methods to help PET developers 'extrapolate' guarantees from lab studies to the real world (WP2); technical standards and certification to quantify the impact of PETs (WP3); will produce online tools to allow researchers to audit deployed PETs (WP4); and undertake a broad programme of outreach and engagement to inform practice in policy, industry and academia (WP5). The Fellowship will thereby provide a framework to make research using digital traces safe and reliable; support data-driven policy interventions that rely on anonymised administrative data; and inform the regulation of underlying AI technologies-such as generative AI for synthetic data.

Data Management Plans
Powered by OpenAIRE graph
Found an issue? Give us feedback

Do the share buttons not appear? Please make sure, any blocking addon is disabled, and then reload the page.

All Research products
arrow_drop_down
<script type="text/javascript">
<!--
document.write('<div id="oa_widget"></div>');
document.write('<script type="text/javascript" src="https://beta.openaire.eu/index.php?option=com_openaire&view=widget&format=raw&projectId=ukri________::82a43f2e4acad6ac9055338f7e93a9a5&type=result"></script>');
-->
</script>
For further information contact us at helpdesk@openaire.eu

No option selected
arrow_drop_down