Identifying Expressions of Emotion in Text



Yüklə 326,16 Kb.
səhifə5/9
tarix17.05.2023
ölçüsü326,16 Kb.
#115815
1   2   3   4   5   6   7   8   9
Identifying Expressions of Emotion in Text

Category

a↔b

a↔c

a↔d

average

happiness

0.76

0.84

0.71

0.77

sadness

0.68

0.79

0.56

0.68

anger

0.62

0.76

0.59

0.66

disgust

0.64

0.62

0.74

0.67

surprise

0.61

0.72

0.48

0.60

fear

0.78

0.80

0.78

0.79

mixed emotion

0.24

0.61

0.44

0.43

Within the emotion sentences, there are seven possible categories of emotion to which a sentence can be assigned. Table 4 shows the value of kappa for each of these emotion categories for each annotator pair. The agreement was found to be highest for fear and happiness. From this, we can surmise that writers express these emotions in more explicit and unambiguous terms, which makes them easy to identify. The mixed emotion category showed least agreement which was expected, given the fact that this category was added to account for the sentences which had more than one emotions, or which would not fit into any of the six basic emotion categories.


Agreement on emotion intensities can also be measured using kappa, as there are distinct categories – high, medium, and low. Table 5 shows the values of inter- annotator agreement in terms of kappa for each emotion intensity. The judges agreed more when the emotion intensity was high; agreement declined with decrease in the intensity of emotion. It is a major factor in disagreement that where one judge perceives a low-intensity, another judge may find no emotion.
Table 5. Pair-wise agreement in emotion intensities



Intensity

a↔b

a↔c

a↔d

average

High

0.69

0.82

0.65

0.72

Medium

0.39

0.61

0.38

0.46

Low

0.31

0.50

0.29

0.37

Emotion indicators are words or strings of words selected by annotators as marking emotion in a sentence. Since there are no predefined categories in this case, we cannot use kappa to calculate the agreement between judges. Here we need to find agreement between the sets of text spans selected by the two judges for each sentence.


Several methods of measuring agreement between sets have been proposed. For our task, we chose the measure of agreement on set-valued items (MASI), previously used for measuring agreement on co-reference annotation [10] and in the evaluation of automatic summarization [11]. MASI is a distance between sets whose value is 1 for identical sets, and 0 for disjoint sets. For sets A and B it is defined as:
MASI = J * M, where the Jaccard metric is
J = |AB| / |AB|
and monotonicity is
1, if A B




2 / 3, if A B or B A
M
1 / 3, if A B   , A B   , and B A  
0, if A B  
If one set is monotonic with respect to another, one set's elements always match those of the other set – for instance, in annotation sets {crappy} and {crappy, best} for (6). However, in non-monotonic sets, as in {crappy, relationship} and {crappy, best}, there are elements not contained in one or the other set, indicating a greater degree of disagreement. The presence of monotonicity factor in MASI therefore ensures that the latter cases are penalized more heavily than the former.
While looking for emotion indicators in a sentence, often it is likely that the judges may identify the same expression but differ in marking text span boundaries. For example in sentence (6) the emotion indicator identified by two annotators are “crappy” and “crappy relationship”, which essentially refer to the same item, but disagree on the placement of the span boundary. This leads to strings of varying lengths. To simplify the agreement measurement, we split all strings into words to ensure that members of the set are all individual words. MASI was calculated for each pair of annotations for all sentences in the corpus (see Table 6).

  1. We've both had our share of crappy relationship, and are now trying to be the best we can for each other.

We adopted yet another method of measuring agreement between emotion indicators. It is a variant of the IOB encoding [13] used in text chunking and named entity
recognition tasks. We use IO encoding, in which each word in the sentence is labeled as being either In or Outside an emotion indicator text span, as shown in (7).

  1. Sorry/I for/O the/O ranting/I post/O, but/O I/O am/O just/O really/I annoyed/I.

Binary IO labeling of each word in essence reduces the task to that of word-level classification into non-emotion and emotion indicator categories. It follows that kappa can now be used for measuring agreement; pair-wise kappa values using this method are shown in Table 6. The average kappa value of 0.66 is lower than that observed at sentence level classification. This is in line with the common observation that agreement on lower levels of granularity is generally found to be lower.



Yüklə 326,16 Kb.

Dostları ilə paylaş:
1   2   3   4   5   6   7   8   9




Verilənlər bazası müəlliflik hüququ ilə müdafiə olunur ©azkurs.org 2024
rəhbərliyinə müraciət

gir | qeydiyyatdan keç
    Ana səhifə


yükləyin