Adversaries or allies? Privacy and deep learning in big data era
- Publication Type:
- Journal Article
- Citation:
- Concurrency Computation, 2019, 31 (19)
- Issue Date:
- 2019-10-10
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
Liu_et_al-2018-Concurrency_and_Computation__Practice_and_Experience.pdf | Published Version | 1.67 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
© 2018 John Wiley & Sons, Ltd. Deep learning methods have become the basis of new AI-based services on the Internet in big data era because of their unprecedented accuracy. Meanwhile, it raises obvious privacy issues. The deep learning–assisted privacy attack can extract sensitive personal information not only from the text but also from unstructured data such as images and videos. In this paper, we proposed a framework to protect image privacy against deep learning tools, along with two new metrics that measure image privacy. Moreover, we propose two different image privacy protection schemes based on the two metrics, utilizing the adversarial example idea. The performance of our solution is validated by simulations on two different datasets. Our research shows that we can protect the image privacy by adding a small amount of noise that has a humanly imperceptible impact on the image quality, especially for images of complex structures and textures.
Please use this identifier to cite or link to this item: