-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can word vectors have an impact on Textcat? #4009
Comments
Thanks for the report, this does look like a bug. I should have a workaround for you shortly. |
The root cause is that the mechanism by which spaCy decides to use the pretrained vectors is pretty messy, which has led to a number of bugs. In the In many examples, we call The workaround is to add the keyword argument to the call: This bug should be resolved by making sure the components' |
I still didn't get it to work, calling |
Hi, |
The Could you add some more code to show how you're loading your model and trying to train it? |
Thanks for the answer, here is the training code adapted from the example shown on Spacy's website:
I tried calling |
Hi, could I get any update on this? Maybe just some short code example of how it is supposed to work? |
@honnibal : could this be due to the method [EDIT] (GH is being annoying): should we remove that parameter all together and only rely on |
@romlatron : apologies for the late follow-up, this will be fixed in the next version, cf #5004. |
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
I have a model with NER and Textcat components, using custom word vectors.
While the impact of the vectors is clear on the NER, there doesn't seem to be any difference on the Textcat whether it is loaded with or without the vocabulary (and thus the vectors).
Is there a way to get the best of the word vectors to improve my textcat component?
Which page or section is this issue related to?
https://spacy.io/api/textcategorizer
The text was updated successfully, but these errors were encountered: