Stein Points
- Publisher:
- International Machine Learning Society (IMLS)
- Publication Type:
- Conference Proceeding
- Citation:
- 35th International Conference on Machine Learning, 2018, 2 pp. 1320 - 1350
- Issue Date:
- 2018
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
chen18f.pdf | Published version | 1.14 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
An important task in computational statistics and machine learning is to approximate a posterior distribution p(x) with an empirical measure supported on a set of representative points {xi} n i=1 . This paper focuses on methods where the selection of points is essentially deterministic, with an emphasis on achieving accurate approximation when n is small. To this end, we present Stein Points. The idea is to exploit either a greedy or a conditional gradient method to iteratively minimise a kernel Stein discrepancy between the empirical measure and p(x). Our empirical results demonstrate that Stein Points enable accurate approximation of the posterior at modest computational cost. In addition, theoretical results are provided to establish convergence of the method.
Please use this identifier to cite or link to this item: