-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NER training is very slow #1973
Comments
Re: how to enable a gpu, previous thread here |
That code doesn't use minibatching, in order to keep the example simple. For larger training tasks, try using the |
@honnibal Ok, can I convert the TRAIN_DATA in the spacy's json format somehow? |
The easiest way at the moment would be to put the data into BILUO format and use the |
@honnibal i am testing the same code with minibatches and it works good now (with CPU). |
Great! |
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
Hello,
i am training a new NER model with this code: https:/explosion/spaCy/blob/master/examples/training/train_new_entity_type.py
i have noticed that the training is very very slow.
Doing a test with 2500 documents (500-800 tokens long) i only can see few iterations after 8 hours.
Ok, i am running it via CPU but my pc is good.
(i can try with GPU, is now possible to enable it?)
is really the training of a new model that slow? I have around 50k documents, it will took forever.
Can i optimize it somehow?
Thanks!
Your Environment
The text was updated successfully, but these errors were encountered: