-
Notifications
You must be signed in to change notification settings - Fork 169
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to impove predict result? #2
Comments
Thanks for trying out this sample code. When doing model tuning, typically I will first look at the training curve and validation curve, to see whether the model overfits or underfits. From there, I will tune accordingly to address the problem. Moreover, the sample code published here does not include the slot label dependency modeling as mentioned in the code comment. If there are strong dependencies among your labels/tags, connecting label/tag output back to RNN hidden state might help. |
It seems that test fl score and accuracy is good.
PS: I don't know if you know chinese, if you can , may i leave messeage in chinese ?:) |
@HadoopIt I have some questions as below regarding the data set and implementation on github repo as below.
If you have free time, I request you to please answer my questions above. I want to learn about these. Thanks for your time and concern. |
@liaicheng To get the F1 score calculated properly from conlleval.pl, do we need to have some fixed format for 'slots' ? Because I didn't have the same format in my dataset slots, and somehow its printing F1 scores = 0.00 (but actually F1 score is not zero when I manually check the predicted slots in test_results) |
@HadoopIt @liaicheng Have anyone tried using pre-trained embedding word vector to improve the accuracy ? how to do this and where to add this embedding in this project ? Thanks! |
@hariom-yadaw actually i process the slot data as the sample show, so i can get the f1 score. |
@hariom-yadaw Slot labels for words that do not appear in training set might be inferred from the structure in the sequence, e.g. structure like "flight from A to B". Pre-trained word embedding can be fed to the graph when you do tf.session.run, just like feeding other input values. More details: |
Thanks a lot @HadoopIt In case we use pre-trained embedding, do we need to change other things in code like vocab will be taken from pre-trained file and not from train.seq.in ? |
@HadoopIt Also I get weird results as below when I run it on my dataset. the Also why the accuracy doesn't follow monotonous behaviour ? It decreases and decreases in between. what could be the reason behind this ? I get good tagging but somehow global step 300 step-time 0.08. Training perplexity 1.53 My Test data set.test.seq.in test.seq.out(slot tagging) ( |
HI, i used your code and train a model. When predict with my test data, intent result seems good, but tagging task seems worse comparing to other mate.
I changed some flags parameters, such as double 'batch_size',double 'word_embedding_size', double 'max_training_steps',double 'num_layers',no use. Can you give me other tip? :)
thanks!
The text was updated successfully, but these errors were encountered: