Personalised automated assessments
- Publication Type:
- Conference Proceeding
- Citation:
- Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS, 2016, pp. 1115 - 1123
- Issue Date:
- 2016-01-01
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
p1115-gutierrez.pdf | Published version | 1.01 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
Copyright © 2016, International Foundation for Autonomous Agents and Multiagent Systems (www.ifaamas.org). All rights reserved. Consider a person who needs to assess a large amount of information. For instance, think of a teacher of a massive open online course with thousands of enrolled students, or a senior program committee member in a large conference who needs to decide what are the final marks of reviewed papers, or a buyer in an e-commerce scenario who needs to build up her opinion about products. When assessing a large number of objects, sometimes it is simply unfeasible to evaluate them all and very often one needs to rely on the opinions of others. In this paper, we provide a model that uses peer assessments (assessments made by others) in an online community to approximate the assessments that a particular member of the community would generate given the occasion to do so (e.g. the tutor, the SPC member or the buyer-we refer to this person as the leader). Furthermore, we provide a measure of the uncertainty of the computed assessments and a ranking of the objects that should be assessed next. The model, although inspired by human societies is thought to be used in the organisation of agent communities.
Please use this identifier to cite or link to this item: