Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Architecture question #2

Open
mtourne opened this issue Oct 25, 2015 · 0 comments
Open

Architecture question #2

mtourne opened this issue Oct 25, 2015 · 0 comments

Comments

@mtourne
Copy link

mtourne commented Oct 25, 2015

Pretty cool stuff.
Reading the code I'm just wondering about why so many levels of indirection from indexes to word2vec sentence matrixes.

It's like parsing -> creation of an "alphabet" to map words to indexes -> creation of questions / answers as series of alphabet indexes -> creation of an alphabet index to word2vec mapping.
This also requiring a nn layer that will do the lookup index to word2vec vector, before the convolution.

Is there a reason to bother with indexes at all, and not transforming everything straight into a word2vec matrix either at parsing time or even before the feed forward phase ?
Seems like this way the code would be more tolerant to being fed new document pairs containing words that exist in the word2vec but not in the "alphabet" mapping.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant