Applying differential privacy mechanism in artificial intelligence

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
Proceedings - International Conference on Distributed Computing Systems, 2019, 2019-July, pp. 1601-1609
Issue Date:
2019-07-01
Filename Description Size
08885331.pdfPublished version248.19 kB
Adobe PDF
Full metadata record
© 2019 IEEE. Artificial Intelligence (AI) has attracted a large amount of attention in recent years. However, several new problems, such as privacy violations, security issues, or effectiveness, have been emerging. Differential privacy has several attractive properties that make it quite valuable for AI, such as privacy preservation, security, randomization, composition, and stability. Therefore, this paper presents differential privacy mechanisms for multi-agent systems, reinforcement learning, and knowledge transfer based on those properties, which proves that current AI can benefit from differential privacy mechanisms. In addition, the previous usage of differential privacy mechanisms in private machine learning, distributed machine learning, and fairness in models is discussed, bringing several possible avenues to use differential privacy mechanisms in AI. The purpose of this paper is to deliver the initial idea of how to integrate AI with differential privacy mechanisms and to explore more possibilities to improve AIs performance.
Please use this identifier to cite or link to this item: