Empowerment-driven Policy Gradient Learning with Counterfactual Augmentation in Recommender Systems
- Publisher:
- Institute of Electrical and Electronics Engineers (IEEE)
- Publication Type:
- Conference Proceeding
- Citation:
- Proceedings - IEEE International Conference on Data Mining, ICDM, 2022, 2022-November, pp. 885-890
- Issue Date:
- 2022-01-01
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
Empowerment-driven_Policy_Gradient_Learning_with_Counterfactual_Augmentation_in_Recommender_Systems.pdf | Published version | 392.98 kB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
Deep reinforcement learning (DRL) has been proven its efficiency in capturing users' dynamic interests in recent literature. However, training a DRL agent is challenging, because of the sparse environment in recommender systems (RS), DRL agents could spend times either exploring informative user-item interaction trajectories or using existing trajectories for policy learning. It is also known as the exploration and exploitation trade-off which affects the recommendation performance significantly when the environment is sparse. It is more challenging to balance the exploration and exploitation in DRL RS where RS agent need to deeply explore the informative trajectories and exploit them efficiently in the context of recommender systems. As a step to address this issue, We design a novel empowerment-driven exploration method to increase the capability of exploring informative interaction trajectories in the sparse environment, which are further enriched via a counterfactual augmentation strategy for more efficient exploitation. The extensive experiments on four offline datasets and an online simulation platform demonstrate the superiority of our model to a set of existing state-of-the-art methods.
Please use this identifier to cite or link to this item: