Let's Agree to Disagree: Measuring Agreement between Annotators for Opinion Mining Task
نویسندگان
چکیده
There is a need to know up to what degree humans can agree when classifying a sentence as carrying some sentiment orientation. However, a little research has been done on assessing the agreement between annotators for the different opinion mining tasks. In this work we present an assessment of agreement between two human annotators. The task was to manually classify newspaper sentences into one of three classes. For assessing the level of agreement, Cohen’s kappa coefficient was computed. Results show that annotators agree more for negative classes than for positive or neutral. We observed that annotators might agree up to a level of substantial agreement of 0.65 for the best case or 0.30 for the worst.
منابع مشابه
O-17: Development A Questionnaire Measuring Attitude toward Oocyte Donation
Background Since there was not a valid and pervasive questionnaire regarding attitude toward oocyte donation, this study aimed to design and develop a tool to measure attitude toward donated oocyte. MaterialsAndMethods In this methodological research, qualitative research was done on 15 infertile. Also, literature review and search in various data bases were carried out. Validity of this questi...
متن کاملWord Sense Annotation of Polysemous Words by Multiple Annotators
We describe results of a word sense annotation task using WordNet, involving half a dozen well-trained annotators on ten polysemous words for three parts of speech. One hundred sentences for each word were annotated. Annotators had the same level of training and experience, but interannotator agreement (IA) varied across words. There was some effect of part of speech, with higher agreement on n...
متن کاملThe CASS Technique for Evaluating the Performance of Argument Mining
Argument mining integrates many distinct computational linguistics tasks, and as a result, reporting agreement between annotators or between automated output and gold standard is particularly challenging. More worrying for the field, agreement and performance are also reported in a wide variety of different ways, making comparison between approaches difficult. To solve this problem, we propose ...
متن کاملOpinion Mining of Spanish Customer Comments with Non-Expert Annotations on Mechanical Turk
One of the major bottlenecks in the development of data-driven AI Systems is the cost of reliable human annotations. The recent advent of several crowdsourcing platforms such as Amazon’s Mechanical Turk, allowing requesters the access to affordable and rapid results of a global workforce, greatly facilitates the creation of massive training data. Most of the available studies on the effectivene...
متن کاملMeaning Unit Segmentation in English and Chinese: a New Approach to Discourse Phenomena
We present a new approach to dialogue processing in terms of “meaning units”. In our annotation task, we asked speakers of English and Chinese to mark boundaries where they could construct the maximal concept using minimal words. We compared English data across genres (news, literature, and policy). We analyzed the agreement for annotators using a state-ofthe-art segmentation similarity algorit...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Research in Computing Science
دوره 110 شماره
صفحات -
تاریخ انتشار 2016