-
-
Notifications
You must be signed in to change notification settings - Fork 143
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Where is the dataset for training? #7
Comments
Hi, @arijit1410 Sorry, I cannot show the whole dataset (according to some reason). Hope this helps! |
Thanks for helping out quickly. One more question, what purpose does the 'context.txt' file have? |
The |
Oh so we don't require it we are loading pre-trained word embeddings? |
Yes. |
@RandolphVI can you please give us an insight on the sample data format in data_sample.json, what exactly. is "features_content", "labels_index" and "labels_num". An early reply will be greatly appreciated. |
For instance, you have two sentences:
Now, you have to:
So the Hope this helps! |
One more thing needed clarification @RandolphVI , can you please give information about the directory in which we have to place train.json, test.json, and validation.json. I was successfully able to train the dataset but testing it is giving several issues. |
Like this:
|
Thanks for an early reply @RandolphVI. I was successfully able to train my dataset but when I tested on the TestSet, the predictions file predicted the same label for each and every data point and that too was consistent across all the models, due to which I am getting same precision, recall, and F score for all the models. @RandolphVI what do you think could the reason for this? I have attached predictions |
Sorry for replying so late. Did you figure it out? |
@RandolphVI It's not changing much, to be honest during the training phase. I changed some parameters like the threshold to see there is some issue in labels predictions logic, but that seems to be working fine. This means the issue could be during the training phase itself. Changing models lead to change in the evaluation metrics, but still, every model is predicting the same set of labels for each and every data point in the test set. This seems quite strange don't you think @RandolphVI. |
@RandolphVI ,When run test_cnn.py, |
@Emmanuelgiwu |
@RandolphVI I seem to be facing the same issue as @akash418. The model predicts all test data as the same set of labels every time. Even during eval cycles, all the metrics except ROC-AUC are quite bad. This leads me to think that the issue is during the training cycle itself as @akash418 had mentioned. Have you faced this issue in your runs? @akash418 Were you able to work around this? |
No description provided.
The text was updated successfully, but these errors were encountered: