Refresh Strategies in Continuous Active Learning
MetadataShow full item record
High recall information retrieval is crucial to tasks such as electronic discovery and systematic review. Continuous Active Learning (CAL) is a technique where a human assessor works in loop with a machine learning model; the model presents a set of documents likely to be relevant and the assessor provides relevance feedback. Our focus in this thesis is on one particular aspect of CAL: refreshing, which is a crucial and recurring event in the CAL process. During a refresh, the machine learning model is trained with the relevance judgments and a new list of likely-to-be-relevant documents is produced for the assessor to judge. It is also computationally the most expensive step in CAL. In this thesis, we investigate the effects of the default and alternative refresh strategies on the effectiveness and efficiency of CAL. We find that more frequent refreshes can significantly reduce the human effort required to achieve certain recall. For moderately sized datasets, the high computation cost of frequent refreshes can be reduced through a careful implementation. For dealing with resource constraints and large datasets, we propose alternative refresh strategies which provide the benefits of frequent refreshes at a lower computation cost. In this thesis, we also discuss the design of a modern implementation of the CAL algorithm which is efficient and extensible. Our implementation can be used as a research tool as well as for practical applications.
Cite this version of the work
Nimesh Ghelani (2018). Refresh Strategies in Continuous Active Learning. UWSpace. http://hdl.handle.net/10012/13669