Human question answering performance using an interactive document retrieval system

Mark D. Smucker, James Allan, Blagovest Dachev
2012 Proceedings of the 4th Information Interaction in Context Symposium on - IIIX '12  
Every day, people widely use information retrieval (IR) systems to find documents that answer their questions. Compared to these IR systems, question answering (QA) systems aim to speed the rate at which users find answers by retrieving answers rather than documents. To better understand how IR systems compare to QA systems, we measured the performance of humans using an interactive IR system to answer questions. We conducted our experiments within the framework of the TREC 2007 complex,
more » ... tive question answering (ciQA) track. We found that the average QA system was comparable to humans using an IR system. Our results also show that for some users IR systems can be powerful question answering systems. After only 5 minutes of usage per question, one user of the IR system obtained an average F (β = 3) score of 0.800, which outperformed the best QA system by 27% and the average QA system by 40%. After 10 minutes of usage, 5 of 8 users of the IR system obtained a higher performance than the average QA system. To achieve superior performance, future QA systems should combine the flexibility and precision of IR systems with the ease-of-use and recall advantages of QA systems.
doi:10.1145/2362724.2362735 dblp:conf/iiix/SmuckerAD12 fatcat:brijt767lfgwnhmgl2otpt4wji