Advertisement

Evaluation and Refinement of Clustered Search Results with the Crowd

Evaluation and Refinement of Clustered Search Results with the Crowd Evaluation and Refinement of Clustered Search Results with the Crowd
Amy X. Zhang, Jilin Chen, Wei Chai, Jinjun Xu, Lichan Hong, Ed CHI

IUI '19: 24th International Conference on Intelligent User Interfaces
Session: Collaborative Interfaces II

Abstract
When searching on the web or in an app, results are often returned as lists of hundreds to thousands of items, making it difficult for users to understand or navigate the space of results. Research has demonstrated that using clustering to partition search results into coherent, topical clusters can aid in both exploration and discovery. Yet clusters generated by an algorithm for this purpose are often of poor quality and do not satisfy users. To achieve acceptable clustered search results, experts must manually evaluate and refine the clustered results for each search query, a process that does not scale to large numbers of search queries. In this article, we investigate using crowd-based human evaluation to inspect, evaluate, and improve clusters to create high-quality clustered search results at scale. We introduce a workflow that begins by using a collection of well-known clustering algorithms to produce a set of clustered search results for a given query. Then, we use crowd workers to holistically assess the quality of each clustered search result to find the best one. Finally, the workflow has the crowd spot and fix problems in the best result to produce a final output. We evaluate this workflow on 120 top search queries from the Google Play Store, some of whom have clustered search results as a result of evaluations and refinements by experts. Our evaluations demonstrate that the workflow is effective at reproducing the evaluation of expert judges and also improves clusters in a way that agrees with experts and crowds alike.

DOI::
WEB::

Recorded at the 24th International Conference on Intelligent User Interfaces, Los Angeles, USA, March 16 - 20 2019

SIGCHI,IUI 2019,cluster evaluation,Relevance assessment,Clustering and classification,search engines,mobile app stores,Evaluation of retrieval results,Presentation of retrieval results,clusters,crowdsourcing,Clustered search results,Empirical studies in collaborative and social computing,human computation,

Post a Comment

0 Comments