Table 3. The average number of lines of text messages for the different categories distinguished. The agreement of the two coders is quantified by the kappa score. A kappa score above 0.8 suggests very good agreement, and a kappa score between 0.6 and 0.8 suggests good agreement.


Topic category Average number per round per group Kappa score
Discussion, past rounds (evaluative) 4.2 0.92
Discussion, past rounds (procedural) 0.8 0.74
Punishment (positive) 0.3 0.74
Punishment (negative) 2.3 0.78
Punishment (general threats) 0.4 0.70
General strategy (temporal) 1.0 0.75
General strategy (spatial) 1.2 0.66
General strategy (mode) 2.2 0.63
General strategy (general) 1.4 0.84
Specific strategy (time: proposed) 0.4 0.77
Specific strategy (time: discussion) 6.9 0.82
Specific strategy (space: proposed) 0.3 0.70
Specific strategy (space: discussion) 7.6 0.80
Affirmation 0.5 0.66
Experiment (intent) 0.7 0.81
Experiment (procedures) 1.8 0.75
Experiment (software) 1.2 0.78
Experiment (uncertainty) 0.1 0.81
General discussion 9.4 0.75
Off topic 7.4 0.85