首页    期刊浏览 2024年12月15日 星期日
登录注册

文章基本信息

  • 标题:Quality Of Crowdsourced Relevance Judgments In Association With Logical Reasoning Ability
  • 本地全文:下载
  • 作者:Sri Devi Ravana ; Parnia Samimi ; Prabha Rajagopal
  • 期刊名称:Malaysian Journal of Computer Science
  • 印刷版ISSN:0127-9084
  • 出版年度:2018
  • 卷号:31
  • 期号:5
  • 出版社:University of Malaya * Faculty of Computer Science and Information Technology
  • 摘要:Crowdsourcing has become a lowcost and scalable alternative to gather relevance assessments through crowdsourcing platforms. However, when gathering subjective human judgments, it can be challenging to enforce data quality due to the lack of vision about how judges make a decision. It is important to determine the attributes that could affect the effectiveness of crowsourcedjudgments in an information retrieval systems evaluation. The purpose of the experiment that is discussed in this paper is to investigate if logical reasoning ability of the crowd workers is related to the quality of the relevant judgments produced through the crowdsource process. The study also evaluates the effect of cognitive characteristics on the quality of relevance judgment compared to the gold standard dataset. Through this experiment, a comparison study is done between the quality of the judgments obtained through the crowdsourcing process and the original baseline judgments generated by the hired experts by TREC. In the study, the systems performances were measured using both of these sets of relevance judgments to see its correlation. The experimentation reveals that quality of relevance judgments is highly correlated with the logical reasoning ability of individuals. The judgment difficulty level reported by the crowdsource workers and the confidence level claimed by the workers showed a significant correlation with the quality of the judgments. Unexpectedly though, selfreported knowledge about a given topic and demographics data have no correlation with the quality of judgments produced through crowdsourcing.
  • 关键词:information retrieval evaluation; human judgments; quality of relevance judgments; crowdsourcing; logical reasoning
国家哲学社会科学文献中心版权所有