sequence tagging for NER for ULMFiT
to replicate result:
you can download the data/ folder from here, and put it in root directory.
I am currently doing experiments in jupyter notebook coNLL_three_layer.ipynb
-
concat both forward and backward outputs from language model
W_LM = [W_forward, W_backward] -
feeding word vectors from GloVe to a BiLSTM and get output
W_glove -
concatenating these outputs
W = [W_glove, W_LM] -
feeding
Wto another BiLSTM to get final result.
F1 score of 76.
(need to improve by fine tuning parameters, see how the toks are preprocessed, adding char embedding, adding CRF layer.
-
which layer of lanuage model should be used for Sequence tagging problem
-
how to build a better language model for sequence tagging
Regularizing and Optimizing LSTM Language Models
deep contextualized word representations
End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF
Semi-supervised sequence tagging with bidirectional language models