Differential Privacy for Software Analysis

Differential privacy has emerged as a leading theoretical framework for privacy-preserving data gathering and analysis. In addition to a rich body of theoretical results, differentially-private analyses have transitioned to industry in companies such as Google, Apple, and Microsoft, as well as to government agencies such as the U.S. Census Bureau. Differential privacy allows meaningful statistics to be collected for a population without revealing "too much" information about any individual member of the population. For the purposes of software analysis, this machinery allows run-time execution data from many users of a deployed software system to be collected and analyzed in a privacy-preserving manner. Such a solution is appealing to many stakeholders: software users, software developers, providers of software analysis infrastructures, and government agencies that enforce consumer protections. The main focus of our current work is on developing differentially-private analyses for gathering profiling data and coverage data from software executions. This work is funded by NSF grant 1907715. Specific contributions are as follows:
main page