|
1
|
|
|
|
|
2
|
|
|
|
|
3
|
1 |
|
import logging |
|
4
|
1 |
|
import math |
|
5
|
|
|
# import datetime |
|
6
|
|
|
# import itertools |
|
7
|
|
|
# import pdb |
|
8
|
|
|
|
|
9
|
1 |
|
from collections import Counter |
|
10
|
|
|
# from pprint import pprint |
|
11
|
|
|
# from datetime import datetime |
|
12
|
|
|
# from collections import defaultdict |
|
13
|
|
|
|
|
14
|
1 |
|
import numpy as np |
|
15
|
1 |
|
import pandas as pd |
|
16
|
|
|
|
|
17
|
|
|
# import crowdtruth.models.unit |
|
18
|
|
|
|
|
19
|
1 |
|
SMALL_NUMBER_CONST = 0.00000001 |
|
20
|
|
|
|
|
21
|
1 |
|
class Metrics(): |
|
22
|
|
|
|
|
23
|
|
|
|
|
24
|
|
|
# Sentence Quality Score |
|
25
|
1 |
|
@staticmethod |
|
26
|
|
|
def sentence_quality_score(sentence_id, sent_work_rel_dict, wqs, rqs): |
|
27
|
|
|
''' |
|
28
|
|
|
sentence_id |
|
29
|
|
|
work_sent_rel_dict |
|
30
|
|
|
rqs: dict of relation_id (string) -> relation quality (float) |
|
31
|
|
|
wqs: dict of worker_id (string) -> worker quality score |
|
32
|
|
|
''' |
|
33
|
|
|
|
|
34
|
1 |
|
sqs_numerator = 0.0 |
|
35
|
1 |
|
sqs_denominator = 0.0 |
|
36
|
1 |
|
worker_ids = list(sent_work_rel_dict[sentence_id].keys()) |
|
37
|
|
|
|
|
38
|
1 |
|
for worker_i in range(len(worker_ids) - 1): |
|
39
|
1 |
|
for worker_j in range(worker_i + 1, len(worker_ids)): |
|
40
|
|
|
# print worker_ids[i] + " - " + worker_ids[j] + "\n" |
|
41
|
1 |
|
numerator = 0.0 |
|
42
|
1 |
|
denominator_i = 0.0 |
|
43
|
1 |
|
denominator_j = 0.0 |
|
44
|
|
|
|
|
45
|
1 |
|
worker_i_vector = sent_work_rel_dict[sentence_id][worker_ids[worker_i]] |
|
46
|
1 |
|
worker_j_vector = sent_work_rel_dict[sentence_id][worker_ids[worker_j]] |
|
47
|
|
|
|
|
48
|
1 |
|
for relation in worker_i_vector: |
|
49
|
1 |
|
worker_i_vector_rel = worker_i_vector[relation] |
|
50
|
1 |
|
worker_j_vector_rel = worker_j_vector[relation] |
|
51
|
1 |
|
numerator += rqs[relation] * (worker_i_vector_rel * worker_j_vector_rel) |
|
52
|
1 |
|
denominator_i += rqs[relation] * (worker_i_vector_rel * worker_i_vector_rel) |
|
53
|
1 |
|
denominator_j += rqs[relation] * (worker_j_vector_rel * worker_j_vector_rel) |
|
54
|
|
|
|
|
55
|
1 |
|
weighted_cosine = numerator / math.sqrt(denominator_i * denominator_j) |
|
56
|
|
|
|
|
57
|
1 |
|
sqs_numerator += weighted_cosine * wqs[worker_ids[worker_i]] * wqs[worker_ids[worker_j]] |
|
58
|
1 |
|
sqs_denominator += wqs[worker_ids[worker_i]] * wqs[worker_ids[worker_j]] |
|
59
|
|
|
|
|
60
|
1 |
|
if sqs_denominator < SMALL_NUMBER_CONST: |
|
61
|
1 |
|
sqs_denominator = SMALL_NUMBER_CONST |
|
62
|
1 |
|
return sqs_numerator / sqs_denominator |
|
63
|
|
|
|
|
64
|
|
|
|
|
65
|
|
|
# Worker - Sentence Agreement |
|
66
|
1 |
|
@staticmethod |
|
67
|
|
|
def worker_sentence_agreement(worker_id, sent_rel_dict, work_sent_rel_dict, sqs, rqs, wqs): |
|
68
|
|
|
''' |
|
69
|
|
|
worker_id |
|
70
|
|
|
sent_rel_dict |
|
71
|
|
|
work_sent_rel_dict |
|
72
|
|
|
sentence_vectors: data frame of sentence vectors |
|
73
|
|
|
sqs (sentence quality score): dict sentence_id -> sentence quality (float) |
|
74
|
|
|
rqs: dict of relation_id (string) -> relation quality (float) |
|
75
|
|
|
wqs: quality score of the given worker |
|
76
|
|
|
''' |
|
77
|
1 |
|
wsa_numerator = 0.0 |
|
78
|
1 |
|
wsa_denominator = 0.0 |
|
79
|
1 |
|
work_sent_rel_dict_worker_id = work_sent_rel_dict[worker_id] |
|
80
|
|
|
|
|
81
|
1 |
|
for sentence_id in work_sent_rel_dict_worker_id: |
|
82
|
1 |
|
numerator = 0.0 |
|
83
|
1 |
|
denominator_w = 0.0 |
|
84
|
1 |
|
denominator_s = 0.0 |
|
85
|
|
|
|
|
86
|
1 |
|
worker_vector = work_sent_rel_dict[worker_id][sentence_id] |
|
87
|
1 |
|
sentence_vector = sent_rel_dict[sentence_id] |
|
88
|
|
|
|
|
89
|
1 |
|
for relation in worker_vector: |
|
90
|
1 |
|
worker_vector_relation = worker_vector[relation] * wqs |
|
91
|
1 |
|
sentence_vector_relation = sentence_vector[relation] |
|
92
|
|
|
|
|
93
|
1 |
|
numerator += rqs[relation] * worker_vector_relation * \ |
|
94
|
|
|
(sentence_vector_relation - worker_vector_relation) |
|
95
|
1 |
|
denominator_w += rqs[relation] * \ |
|
96
|
|
|
(worker_vector_relation * worker_vector_relation) |
|
97
|
1 |
|
denominator_s += rqs[relation] * ( \ |
|
98
|
|
|
(sentence_vector_relation - worker_vector_relation) * \ |
|
99
|
|
|
(sentence_vector_relation - worker_vector_relation)) |
|
100
|
1 |
|
weighted_cosine = None |
|
101
|
1 |
|
if math.sqrt(denominator_w * denominator_s) < SMALL_NUMBER_CONST: |
|
102
|
1 |
|
weighted_cosine = SMALL_NUMBER_CONST |
|
103
|
|
|
else: |
|
104
|
1 |
|
weighted_cosine = numerator / math.sqrt(denominator_w * denominator_s) |
|
105
|
1 |
|
wsa_numerator += weighted_cosine * sqs[sentence_id] |
|
106
|
1 |
|
wsa_denominator += sqs[sentence_id] |
|
107
|
1 |
|
if wsa_denominator < SMALL_NUMBER_CONST: |
|
108
|
1 |
|
wsa_denominator = SMALL_NUMBER_CONST |
|
109
|
1 |
|
return wsa_numerator / wsa_denominator |
|
110
|
|
|
|
|
111
|
|
|
|
|
112
|
|
|
# Worker - Worker Agreement |
|
113
|
1 |
|
@staticmethod |
|
114
|
|
|
def worker_worker_agreement(worker_id, work_sent_rel_dict, sent_work_rel_dict, wqs, sqs, rqs): |
|
115
|
|
|
''' |
|
116
|
|
|
worker_id |
|
117
|
|
|
work_sent_rel_dict |
|
118
|
|
|
sent_work_rel_dict |
|
119
|
|
|
worker_vectors: data frame of worker vectors |
|
120
|
|
|
sqs (sentence quality score): dict sentence_id -> sentence quality (float) |
|
121
|
|
|
rqs: dict of relation_id (string) -> relation quality (float) |
|
122
|
|
|
''' |
|
123
|
|
|
|
|
124
|
1 |
|
wwa_numerator = 0.0 |
|
125
|
1 |
|
wwa_denominator = 0.0 |
|
126
|
|
|
|
|
127
|
1 |
|
wv = work_sent_rel_dict[worker_id] |
|
128
|
1 |
|
sentence_ids = list(work_sent_rel_dict[worker_id].keys()) |
|
129
|
|
|
|
|
130
|
|
|
|
|
131
|
1 |
|
for sentence_id in sentence_ids: |
|
132
|
1 |
|
wv_sentence_id = wv[sentence_id] |
|
133
|
1 |
|
sent_work_rel_dict_sentence_id = sent_work_rel_dict[sentence_id] |
|
134
|
1 |
|
for other_worker_id in sent_work_rel_dict_sentence_id: |
|
135
|
1 |
|
if worker_id != other_worker_id: |
|
136
|
1 |
|
numerator = 0.0 |
|
137
|
1 |
|
denominator_w = 0.0 |
|
138
|
1 |
|
denominator_ow = 0.0 |
|
139
|
|
|
|
|
140
|
1 |
|
sent_work_rel_dict_sentence_id_other_worker_id = sent_work_rel_dict_sentence_id[other_worker_id] |
|
141
|
1 |
|
for relation in wv_sentence_id: |
|
142
|
1 |
|
sent_work_rel_dict_sentence_id_other_worker_id_relation = sent_work_rel_dict_sentence_id_other_worker_id[relation] |
|
143
|
1 |
|
wv_sentence_id_relation = wv_sentence_id[relation] |
|
144
|
|
|
|
|
145
|
1 |
|
numerator += rqs[relation] * (wv_sentence_id_relation * sent_work_rel_dict_sentence_id_other_worker_id_relation) |
|
146
|
|
|
|
|
147
|
1 |
|
denominator_w += rqs[relation] * (wv_sentence_id_relation * wv_sentence_id_relation) |
|
148
|
|
|
|
|
149
|
1 |
|
denominator_ow += rqs[relation] * (sent_work_rel_dict_sentence_id_other_worker_id_relation * |
|
150
|
|
|
sent_work_rel_dict_sentence_id_other_worker_id_relation) |
|
151
|
|
|
|
|
152
|
1 |
|
weighted_cosine = numerator / math.sqrt(denominator_w * denominator_ow) |
|
153
|
|
|
# pdb.set_trace() |
|
154
|
1 |
|
wwa_numerator += weighted_cosine * wqs[other_worker_id] * sqs[sentence_id] |
|
155
|
1 |
|
wwa_denominator += wqs[other_worker_id] * sqs[sentence_id] |
|
156
|
1 |
|
if wwa_denominator < SMALL_NUMBER_CONST: |
|
157
|
1 |
|
wwa_denominator = SMALL_NUMBER_CONST |
|
158
|
1 |
|
return wwa_numerator / wwa_denominator |
|
159
|
|
|
|
|
160
|
|
|
|
|
161
|
|
|
|
|
162
|
|
|
# Sentence - Relation Score |
|
163
|
1 |
|
@staticmethod |
|
164
|
|
|
def sentence_relation_score(sentence_id, relation, sent_work_rel_dict, wqs): |
|
165
|
1 |
|
srs_numerator = 0.0 |
|
166
|
1 |
|
srs_denominator = 0.0 |
|
167
|
|
|
|
|
168
|
1 |
|
worker_ids = sent_work_rel_dict[sentence_id] |
|
169
|
1 |
|
for worker_id in worker_ids: |
|
170
|
1 |
|
srs_numerator += worker_ids[worker_id][relation] * wqs[worker_id] |
|
171
|
1 |
|
srs_denominator += wqs[worker_id] |
|
172
|
1 |
|
if srs_denominator < SMALL_NUMBER_CONST: |
|
173
|
1 |
|
srs_denominator = SMALL_NUMBER_CONST |
|
174
|
1 |
|
return srs_numerator / srs_denominator |
|
175
|
|
|
|
|
176
|
|
|
|
|
177
|
|
|
# Relation Quality Score |
|
178
|
1 |
|
@staticmethod |
|
179
|
|
|
def relation_quality_score(relations, work_sent_rel_dict, sqs, wqs): |
|
180
|
1 |
|
rqs_numerator = dict() |
|
181
|
1 |
|
rqs_denominator = dict() |
|
182
|
|
|
|
|
183
|
1 |
|
for relation in relations: |
|
184
|
1 |
|
rqs_numerator[relation] = 0.0 |
|
185
|
1 |
|
rqs_denominator[relation] = 0.0 |
|
186
|
|
|
|
|
187
|
1 |
|
worker_ids = list(work_sent_rel_dict.keys()) |
|
188
|
1 |
|
for worker_i, work_sent_rel_dict_worker_i in work_sent_rel_dict.items(): |
|
189
|
|
|
#work_sent_rel_dict_worker_i = work_sent_rel_dict[worker_i] |
|
190
|
1 |
|
work_sent_rel_dict_i_keys = list(work_sent_rel_dict_worker_i.keys()) |
|
191
|
1 |
|
for worker_j, work_sent_rel_dict_worker_j in work_sent_rel_dict.items(): |
|
192
|
|
|
#work_sent_rel_dict_worker_j = work_sent_rel_dict[worker_j] |
|
193
|
1 |
|
work_sent_rel_dict_j_keys = list(work_sent_rel_dict_worker_j.keys()) |
|
194
|
|
|
|
|
195
|
|
|
#print worker_i, worker_j,np.intersect1d(np.array(work_sent_rel_dict[worker_i].keys()),np.array(work_sent_rel_dict[worker_j].keys())) |
|
196
|
1 |
|
if worker_i != worker_j and len(np.intersect1d(np.array(work_sent_rel_dict_i_keys),np.array(work_sent_rel_dict_j_keys))) > 0: |
|
197
|
|
|
|
|
198
|
1 |
|
for relation in relations: |
|
199
|
1 |
|
numerator = 0.0 |
|
200
|
1 |
|
denominator = 0.0 |
|
201
|
|
|
|
|
202
|
1 |
|
for sentence_id, work_sent_rel_dict_worker_i_sent in work_sent_rel_dict_worker_i.items(): |
|
203
|
1 |
|
if sentence_id in work_sent_rel_dict_worker_j: |
|
204
|
|
|
#work_sent_rel_dict_worker_i_sent = work_sent_rel_dict_worker_i[sentence_id] |
|
205
|
1 |
|
work_sent_rel_dict_worker_j_sent = work_sent_rel_dict_worker_j[sentence_id] |
|
206
|
|
|
|
|
207
|
1 |
|
work_sent_rel_dict_worker_j_sent_rel = work_sent_rel_dict_worker_j_sent[relation] |
|
208
|
|
|
#print worker_i,worker_j,sentence_id,relation |
|
209
|
1 |
|
numerator += sqs[sentence_id] * (work_sent_rel_dict_worker_i_sent[relation] * |
|
210
|
|
|
work_sent_rel_dict_worker_j_sent_rel) |
|
211
|
1 |
|
denominator += sqs[sentence_id] * work_sent_rel_dict_worker_j_sent_rel |
|
212
|
|
|
|
|
213
|
1 |
|
if denominator > 0: |
|
214
|
1 |
|
rqs_numerator[relation] += wqs[worker_i] * wqs[worker_j] * numerator / denominator |
|
215
|
1 |
|
rqs_denominator[relation] += wqs[worker_i] * wqs[worker_j] |
|
216
|
|
|
|
|
217
|
|
|
|
|
218
|
1 |
|
rqs = dict() |
|
219
|
1 |
|
for relation in relations: |
|
220
|
1 |
|
if rqs_denominator[relation] > SMALL_NUMBER_CONST: |
|
221
|
1 |
|
rqs[relation] = rqs_numerator[relation] / rqs_denominator[relation] |
|
222
|
|
|
|
|
223
|
|
|
# prevent division by zero by storing very small value instead |
|
224
|
1 |
|
if rqs[relation] < SMALL_NUMBER_CONST: |
|
225
|
1 |
|
rqs[relation] = SMALL_NUMBER_CONST |
|
226
|
|
|
else: |
|
227
|
1 |
|
rqs[relation] = SMALL_NUMBER_CONST |
|
228
|
1 |
|
return rqs |
|
229
|
|
|
|
|
230
|
1 |
|
@staticmethod |
|
231
|
1 |
|
def run(results, config, max_delta = 0.001): |
|
232
|
|
|
|
|
233
|
1 |
|
judgments = results['judgments'].copy() |
|
234
|
1 |
|
units = results['units'].copy() |
|
235
|
|
|
|
|
236
|
|
|
#sent_work_rel_dict, work_sent_rel_dict, sent_rel_dict |
|
237
|
|
|
# TODO: change to use all vectors in one unit |
|
238
|
1 |
|
col = list(config.output.values())[0] |
|
239
|
1 |
|
sent_rel_dict = dict(units.copy()[col]) |
|
240
|
|
|
|
|
241
|
1 |
|
def expandedVector(worker, unit): |
|
242
|
|
|
#print worker, unit |
|
243
|
1 |
|
vector = Counter() |
|
244
|
1 |
|
for rel in unit: |
|
245
|
1 |
|
if rel in worker: |
|
246
|
1 |
|
vector[rel] = worker[rel] |
|
247
|
|
|
else: |
|
248
|
1 |
|
vector[rel] = 0 |
|
249
|
1 |
|
return vector |
|
250
|
|
|
|
|
251
|
|
|
# fill judgment vectors with unit keys |
|
252
|
1 |
|
for index,row in judgments.iterrows(): |
|
253
|
|
|
# judgments.set_value(index, col, expandedVector(row[col], units.at[row['unit'], col])) |
|
254
|
1 |
|
judgments.at[index, col] = expandedVector(row[col], units.at[row['unit'], col]) |
|
255
|
|
|
|
|
256
|
|
|
#print judgments.head() |
|
257
|
|
|
|
|
258
|
1 |
|
sent_work_rel_dict = judgments[['unit','worker',col]].copy().groupby('unit') |
|
259
|
1 |
|
sent_work_rel_dict = {name : group.set_index('worker')[col].to_dict() for name, group in sent_work_rel_dict} |
|
260
|
|
|
|
|
261
|
|
|
#print sent_work_rel_dict |
|
262
|
|
|
|
|
263
|
1 |
|
work_sent_rel_dict = judgments[['worker','unit',col]].copy().groupby('worker') |
|
264
|
1 |
|
work_sent_rel_dict = {name : group.set_index('unit')[col].to_dict() for name, group in work_sent_rel_dict} |
|
265
|
|
|
# print [i for i in list(sent_work_rel_dict)] |
|
266
|
|
|
# sent_work_rel_dict = {k : dict(sent_work_rel_dict[k]) for k in sent_work_rel_dict} |
|
267
|
|
|
|
|
268
|
|
|
#pprint(work_sent_rel_dict) |
|
269
|
|
|
|
|
270
|
|
|
#initialize data structures |
|
271
|
1 |
|
sqs_list = list() |
|
272
|
1 |
|
wqs_list = list() |
|
273
|
1 |
|
wwa_list = list() |
|
274
|
1 |
|
wsa_list = list() |
|
275
|
1 |
|
rqs_list = list() |
|
276
|
|
|
|
|
277
|
1 |
|
sqs = dict((sentence_id, 1.0) for sentence_id in sent_work_rel_dict) |
|
278
|
1 |
|
wqs = dict((worker_id, 1.0) for worker_id in work_sent_rel_dict) |
|
279
|
1 |
|
wwa = dict((worker_id, 1.0) for worker_id in work_sent_rel_dict) |
|
280
|
1 |
|
wsa = dict((worker_id, 1.0) for worker_id in work_sent_rel_dict) |
|
281
|
|
|
|
|
282
|
1 |
|
sqs_list.append(sqs.copy()) |
|
283
|
1 |
|
wqs_list.append(wqs.copy()) |
|
284
|
1 |
|
wwa_list.append(wwa.copy()) |
|
285
|
1 |
|
wsa_list.append(wsa.copy()) |
|
286
|
|
|
|
|
287
|
|
|
# initialize RQS depending on whether or not it is an open ended task |
|
288
|
1 |
|
rqs = dict() |
|
289
|
1 |
|
if not config.open_ended_task: |
|
290
|
1 |
|
rqs_keys = list(sent_rel_dict[list(sent_rel_dict.keys())[0]].keys()) |
|
291
|
1 |
|
for relation in rqs_keys: |
|
292
|
1 |
|
rqs[relation] = 1.0 |
|
293
|
|
|
else: |
|
294
|
1 |
|
for sentence_id in sent_rel_dict: |
|
295
|
1 |
|
for relation in sent_rel_dict[sentence_id]: |
|
296
|
1 |
|
rqs[relation] = 1.0 |
|
297
|
1 |
|
rqs_list.append(rqs.copy()) |
|
298
|
|
|
|
|
299
|
1 |
|
sqs_len = len(list(sqs.keys())) * 1.0 |
|
300
|
1 |
|
wqs_len = len(list(wqs.keys())) * 1.0 |
|
301
|
1 |
|
rqs_len = len(list(rqs.keys())) * 1.0 |
|
302
|
|
|
|
|
303
|
|
|
# compute metrics until stable values |
|
304
|
1 |
|
iterations = 0 |
|
305
|
1 |
|
while max_delta >= 0.001: |
|
306
|
1 |
|
sqs_new = dict() |
|
307
|
1 |
|
wqs_new = dict() |
|
308
|
1 |
|
wwa_new = dict() |
|
309
|
1 |
|
wsa_new = dict() |
|
310
|
|
|
|
|
311
|
1 |
|
avg_sqs_delta = 0.0 |
|
312
|
1 |
|
avg_wqs_delta = 0.0 |
|
313
|
1 |
|
avg_rqs_delta = 0.0 |
|
314
|
1 |
|
max_delta = 0.0 |
|
315
|
|
|
|
|
316
|
|
|
# pdb.set_trace() |
|
317
|
|
|
|
|
318
|
1 |
|
if not config.open_ended_task: |
|
319
|
|
|
# compute relation quality score (RQS) |
|
320
|
1 |
|
rqs_new = Metrics.relation_quality_score(list(rqs.keys()), work_sent_rel_dict, |
|
321
|
|
|
sqs_list[len(sqs_list) - 1], |
|
322
|
|
|
wqs_list[len(wqs_list) - 1]) |
|
323
|
1 |
|
for relation, _ in rqs_new.items(): |
|
324
|
1 |
|
max_delta = max(max_delta, abs(rqs_new[relation] - rqs_list[len(rqs_list) - 1][relation])) |
|
325
|
1 |
|
avg_rqs_delta += abs(rqs_new[relation] - rqs_list[len(rqs_list) - 1][relation]) |
|
326
|
1 |
|
avg_rqs_delta /= rqs_len |
|
327
|
|
|
|
|
328
|
|
|
# compute sentence quality score (SQS) |
|
329
|
1 |
|
for sentence_id, _ in sent_work_rel_dict.items(): |
|
330
|
1 |
|
sqs_new[sentence_id] = Metrics.sentence_quality_score(sentence_id, sent_work_rel_dict, |
|
331
|
|
|
wqs_list[len(wqs_list) - 1], |
|
332
|
|
|
rqs_list[len(rqs_list) - 1]) |
|
333
|
1 |
|
max_delta = max(max_delta, abs(sqs_new[sentence_id] - sqs_list[len(sqs_list) - 1][sentence_id])) |
|
334
|
1 |
|
avg_sqs_delta += abs(sqs_new[sentence_id] - sqs_list[len(sqs_list) - 1][sentence_id]) |
|
335
|
1 |
|
avg_sqs_delta /= sqs_len |
|
336
|
|
|
|
|
337
|
|
|
# compute worker quality score (WQS) |
|
338
|
1 |
|
for worker_id, _ in work_sent_rel_dict.items(): |
|
339
|
1 |
|
wwa_new[worker_id] = Metrics.worker_worker_agreement( |
|
340
|
|
|
worker_id, work_sent_rel_dict, |
|
341
|
|
|
sent_work_rel_dict, |
|
342
|
|
|
wqs_list[len(wqs_list) - 1], |
|
343
|
|
|
sqs_list[len(sqs_list) - 1], |
|
344
|
|
|
rqs_list[len(rqs_list) - 1]) |
|
345
|
1 |
|
wsa_new[worker_id] = Metrics.worker_sentence_agreement( |
|
346
|
|
|
worker_id, |
|
347
|
|
|
sent_rel_dict, |
|
348
|
|
|
work_sent_rel_dict, |
|
349
|
|
|
sqs_list[len(sqs_list) - 1], |
|
350
|
|
|
rqs_list[len(rqs_list) - 1], |
|
351
|
|
|
wqs_list[len(rqs_list) - 1][worker_id]) |
|
352
|
1 |
|
wqs_new[worker_id] = wwa_new[worker_id] * wsa_new[worker_id] |
|
353
|
1 |
|
max_delta = max( |
|
354
|
|
|
max_delta, |
|
355
|
|
|
abs(wqs_new[worker_id] - wqs_list[len(wqs_list) - 1][worker_id])) |
|
356
|
1 |
|
avg_wqs_delta += abs(wqs_new[worker_id] - wqs_list[len(wqs_list) - 1][worker_id]) |
|
357
|
1 |
|
avg_wqs_delta /= wqs_len |
|
358
|
|
|
|
|
359
|
|
|
# save results for current iteration |
|
360
|
1 |
|
sqs_list.append(sqs_new.copy()) |
|
361
|
1 |
|
wqs_list.append(wqs_new.copy()) |
|
362
|
1 |
|
wwa_list.append(wwa_new.copy()) |
|
363
|
1 |
|
wsa_list.append(wsa_new.copy()) |
|
364
|
1 |
|
if not config.open_ended_task: |
|
365
|
1 |
|
rqs_list.append(rqs_new.copy()) |
|
|
|
|
|
|
366
|
1 |
|
iterations += 1 |
|
367
|
|
|
|
|
368
|
|
|
# reconstruct sent_rel_dict with worker scores |
|
369
|
1 |
|
new_sent_rel_dict = dict() |
|
370
|
1 |
|
for sent_id, rel_dict in sent_rel_dict.items(): |
|
371
|
1 |
|
new_sent_rel_dict[sent_id] = dict() |
|
372
|
1 |
|
for relation, _ in rel_dict.items(): |
|
373
|
1 |
|
new_sent_rel_dict[sent_id][relation] = 0.0 |
|
374
|
1 |
|
for work_id, srd in work_sent_rel_dict.items(): |
|
375
|
1 |
|
wqs_work_id = wqs_new[work_id] |
|
376
|
1 |
|
for sent_id, rel_dict in srd.items(): |
|
377
|
1 |
|
for relation, score in rel_dict.items(): |
|
378
|
1 |
|
new_sent_rel_dict[sent_id][relation] += score * wqs_work_id |
|
379
|
|
|
# pdb.set_trace() |
|
380
|
1 |
|
sent_rel_dict = new_sent_rel_dict |
|
381
|
|
|
|
|
382
|
1 |
|
logging.info(str(iterations) + " iterations; max d= " + str(max_delta) + " ; wqs d= " + str(avg_wqs_delta) + "; sqs d= " + str(avg_sqs_delta) + "; rqs d= " + str(avg_rqs_delta)) |
|
383
|
|
|
|
|
384
|
|
|
#if iterations == 1: |
|
385
|
|
|
# break |
|
386
|
|
|
#pprint(sqs_list) |
|
387
|
|
|
#pprint(wqs_list) |
|
388
|
|
|
#pprint(rqs_list) |
|
389
|
|
|
|
|
390
|
1 |
|
srs = Counter() |
|
391
|
1 |
|
for sentence_id in sent_rel_dict: |
|
392
|
1 |
|
srs[sentence_id] = Counter() |
|
393
|
1 |
|
for relation in sent_rel_dict[sentence_id]: |
|
394
|
1 |
|
srs[sentence_id][relation] = Metrics.sentence_relation_score(sentence_id, relation, sent_work_rel_dict, wqs_list[len(wqs_list) - 1]) |
|
395
|
|
|
|
|
396
|
1 |
|
srs_initial = Counter() |
|
397
|
1 |
|
for sentence_id in sent_rel_dict: |
|
398
|
1 |
|
srs_initial[sentence_id] = Counter() |
|
399
|
1 |
|
for relation in sent_rel_dict[sentence_id]: |
|
400
|
1 |
|
srs_initial[sentence_id][relation] = Metrics.sentence_relation_score(sentence_id, relation, sent_work_rel_dict, wqs_list[0]) |
|
401
|
|
|
|
|
402
|
1 |
|
results['units']['uqs'] = pd.Series(sqs_list[-1]) |
|
403
|
1 |
|
results['units']['unit_annotation_score'] = pd.Series(srs) |
|
404
|
1 |
|
results['workers']['wqs'] = pd.Series(wqs_list[-1]) |
|
405
|
1 |
|
results['workers']['wwa'] = pd.Series(wwa_list[-1]) |
|
406
|
1 |
|
results['workers']['wsa'] = pd.Series(wsa_list[-1]) |
|
407
|
1 |
|
if not config.open_ended_task: |
|
408
|
1 |
|
results['annotations']['aqs'] = pd.Series(rqs_list[-1]) |
|
409
|
|
|
|
|
410
|
1 |
|
results['units']['uqs_initial'] = pd.Series(sqs_list[1]) |
|
411
|
1 |
|
results['units']['unit_annotation_score_initial'] = pd.Series(srs_initial) |
|
412
|
1 |
|
results['workers']['wqs_initial'] = pd.Series(wqs_list[1]) |
|
413
|
1 |
|
results['workers']['wwa_initial'] = pd.Series(wwa_list[1]) |
|
414
|
1 |
|
results['workers']['wsa_initial'] = pd.Series(wsa_list[1]) |
|
415
|
1 |
|
if not config.open_ended_task: |
|
416
|
1 |
|
results['annotations']['aqs_initial'] = pd.Series(rqs_list[1]) |
|
417
|
|
|
return results |
|
418
|
|
|
|