-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cPickle.UnpicklingError: unpickling stack underflow #1447
Comments
I believe that The method |
@jayantj @prakhar2b wdyt? |
@gojomo yes, As for OOV word synthesis, what do you mean by As of now, |
Yes, with the from gensim.models.wrappers.fasttext import FastText
model = FastText.load_fasttext_format('/path/to/model') # without the .bin/.vec extension With the from gensim.models.keyedvectors import KeyedVectors
model = KeyedVectors.load_word2vec_format('/path/to/model.vec') # with the .vec extension |
@jayantj Thank, let me try first with the |
@prakhar2b My "not sure" comment was regarding to some discussion I saw on another issue or PR in progress, perhaps the one that's also discussing whether the discarding-of-untrained-ngrams is a necessary optimization – I had the impression our calculation might be diverging from the original FB fasttext on some (perhaps just OOV) words. (And even if that's defensible, because the untrained ngrams are still just random vectors, it might not be the 'right thing' overall because it may violate user expectations that whether loaded into original FT code, or gensim FT code, OOV words get the same vectors from the same loaded model.) |
We definitely want to follow whatever the original FT does -- the path of least surprise for anyone migrating / trying both. |
I get this error while loading
wiki.en.vec
from FastText pre-trained Word2Vec model. See here for this model.loaded with
I'm using
The text was updated successfully, but these errors were encountered: