-
Notifications
You must be signed in to change notification settings - Fork 83
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
One question regarding padding #4
Comments
Hello! That's correct. This is because the convolutional neural net (CNN) processes the padded character vectors of equal length. Alternatively, one could split the CNN input into equal-size batches (e.g. here). Each input batch to the bi-directional LSTM has the same length, depending on how many words there are in a document. |
Thank you for responding. One follow up question. |
The maximum length was chosen after analyzing word lengths in the documents, such that no words are cut off. |
There is some confusion. In my understanding, padding is applied to only characters not to words. |
hi,
I see that you are padding the inputs to get equal length of 52. But, it seems padding is applied to only character inputs but not to words.
Sentences contains below:
I see that you have made batches of inputs with words of equal length. Is this the correct approach?
Can you pls let me know.
The text was updated successfully, but these errors were encountered: