Training with Streaming Annotation

02/11/2020
by   Tongtao Zhang, et al.
0

In this paper, we address a practical scenario where training data is released in a sequence of small-scale batches and annotation in earlier phases has lower quality than the later counterparts. To tackle the situation, we utilize a pre-trained transformer network to preserve and integrate the most salient document information from the earlier batches while focusing on the annotation (presumably with higher quality) from the current batch. Using event extraction as a case study, we demonstrate in the experiments that our proposed framework can perform better than conventional approaches (the improvement ranges from 3.6 to 14.9 noise in the early annotation; and our approach spares 19.1 to the best conventional method.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset