Abstract
Introducing Shallow Syntactic Information within the Graph-based Dependency Parsing
Nikolay Paev, Kiril Ivanov Simov, Petya Osenova
The paper presents a new BERT model, finetuned for parsing of Bulgarian texts. This model is extended with a new neural network layer in order to incorporate shallow syntactic information during the training phase. The results show statistically significant improvement over the baseline. Thus, the addition of syntactic knowledge- even partial- makes the model better. Also, some error analysis has been conducted on the results from the parsers. Although the architecture has been designed and tested for Bulgarian, it is also scalable for other languages. This scalability was shown here with some experiments and evaluation on an English treebank with a comparable size.