TY - JOUR
T1 - Identifying the reasons contributing to question deletion in educational Q&A
AU - Rath, Manasa
AU - Shah, Chirag
AU - Floegel, Diana
N1 - Funding Information:
The authors thank Sameera Somisetty and Dhrashti Mehta for their inputs in this work. The work reported in this paper is supported by the US Institute of Museum and Library Services (IMLS) grant LG-81-16-0025-16. We are also thankful to Brainly for providing us with the data.
Publisher Copyright:
Copyright © 2017 by Association for Information Science and Technology
PY - 2017/1
Y1 - 2017/1
N2 - Community question-answering (CQA) services are widely used by information seekers looking to ask questions and obtain accurate, personalized answers. Though general CQA sites such as Yahoo! Answers attract a diverse pool of users from many walks of life, other sites cater to a specific user pool. While identifying bad CQA content is generally important in order to improve sites' overall health and community knowledge-sharing, examining educational CQAs is particularly urgent in order to help struggling students understand why their questions fail, re-frame their inquiries in a more accurate manner based on feedback, and ultimately receive correct answers that facilitate their learning process. Otherwise, students' questions would merely be deleted, meaning they lose multiple opportunities to enrich their knowledge base. In this work, we focus on questions posted to Brainly, the largest educational CQA site, in order to first identify “bad” questions and next understand what textual (content-based) features contribute to such questions' poor quality. Using a sample of 1,000 questions–500 of which were deemed “good” and 500 of which were deemed “bad” by site moderators– we attempt to automatically classify question quality in order to label which questions would be deleted and therefore go unanswered. We then use human assessment to expand upon a typology to classify poor quality questions based on 14 textual features in order to identify why they have been marked for deletion. Finally, we propose a method to automatically identify questions' problematic textual features in order to provide feedback to students posting “bad” questions and ensure that they are given the opportunity to revise and improve their inquiries to obtain accurate answers that resolve their information needs.
AB - Community question-answering (CQA) services are widely used by information seekers looking to ask questions and obtain accurate, personalized answers. Though general CQA sites such as Yahoo! Answers attract a diverse pool of users from many walks of life, other sites cater to a specific user pool. While identifying bad CQA content is generally important in order to improve sites' overall health and community knowledge-sharing, examining educational CQAs is particularly urgent in order to help struggling students understand why their questions fail, re-frame their inquiries in a more accurate manner based on feedback, and ultimately receive correct answers that facilitate their learning process. Otherwise, students' questions would merely be deleted, meaning they lose multiple opportunities to enrich their knowledge base. In this work, we focus on questions posted to Brainly, the largest educational CQA site, in order to first identify “bad” questions and next understand what textual (content-based) features contribute to such questions' poor quality. Using a sample of 1,000 questions–500 of which were deemed “good” and 500 of which were deemed “bad” by site moderators– we attempt to automatically classify question quality in order to label which questions would be deleted and therefore go unanswered. We then use human assessment to expand upon a typology to classify poor quality questions based on 14 textual features in order to identify why they have been marked for deletion. Finally, we propose a method to automatically identify questions' problematic textual features in order to provide feedback to students posting “bad” questions and ensure that they are given the opportunity to revise and improve their inquiries to obtain accurate answers that resolve their information needs.
KW - classification
KW - community question-answering
KW - quality factors
KW - question deletion
KW - textual features
UR - http://www.scopus.com/inward/record.url?scp=85040766922&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85040766922&partnerID=8YFLogxK
U2 - 10.1002/pra2.2017.14505401036
DO - 10.1002/pra2.2017.14505401036
M3 - Article
AN - SCOPUS:85040766922
SN - 2373-9231
VL - 54
SP - 327
EP - 336
JO - Proceedings of the Association for Information Science and Technology
JF - Proceedings of the Association for Information Science and Technology
IS - 1
ER -