-
Notifications
You must be signed in to change notification settings - Fork 267
SLING with many frames on a small dataset #446
Comments
Just as other deep models, the SLING parser training is "data hungry". It needs a fair amount of training data, and more data is better data! We are working on a silver annotation pipeline, which can take a Wikipedia and generate synthetic training data using the Wikidata knowledge base and a bunch of heuristics. PS: Please notice that the SLING project has moved to https://github.com/ringgaard/sling. |
Thanks for the pointers.
|
The current parser is just supervised learning with cross-entropy loss. While the silver data pipeline makes use of the knowledge base, the parser training does not (yet) use the knowledge base during training. The silver annotations are noisy, especially when it comes to recall. There are many facts mentioned in Wikipedia that are not known in Wikidata, so the ROLE score is not very high when measured on the silver data. However, the silver parser is only meant to be the first step in the parser training. We are working on adding support for reinforcement learning (RL) to the parser trainer. After the parser has been trained on the silver data, the next phase is to use RL to "fine-tune" the parser. The silver parser is then used for sampling during the RL training. The hope is that the silver model will assign some probability mass for the correct (unknown) golden annotations and that the correct annotation will get a higher reward than the silver annotations. We have a plausibility model that we hope can be used as part of the reward function. I am not sure I understand what you mean by "custom logical relationships among frames". You can add you own annotators to the silver annotation pipeline which can add or modify the annotations. You could try to run the training in two phases. First the silver training, and then train on your own training data set. It might work because the silver training will pretrain the contextual embeddings, but you will have to try to do the experiment to see how it works. There is an option to start training from an existing model which you can use. One thing to be aware of is that you will have to add all the roles you need. Currently the role set is either determined from the training data (caspar) or a fixed set (knolex) that is currently hard-coded. |
These questions are about using SLING on a small data set.
Motivation for using SLING: SLING appears to be unique in providing a good framework for specifying and training on arbitrary entity relationships that originate from text, while still allowing entities that are not directly tied to specific tokens.
However, on limited data, my confidence in SLING is based on the following assumptions:
I'd appreciate any insights you have or corrections to my understanding. Thank you.
The text was updated successfully, but these errors were encountered: