-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
han-transformer not work #91
Comments
I think you are run on toy task, and loss is decreasing. if you want coverage, you can use real task, |
I run p1_HierarchicalAttention_model.py ,its work fine .and i just want to try transformer,but i dosen't work,so i try you test function,but it dose not work too? |
got it. transformer need to fix a problem. before that you can use this: https://github.com/google-research/bert/blob/master/modeling.py |
thanks |
I think transformer work better now. as someone find a bug which was fixed: |
hey,that's me . and i just try your transofrmer code.it's work. but i campare to textcnn, just as the same conclusion in your github(https://github.com/brightmart/text_classification). transformer isn't better than textcnn ,may be my dataset is not large enough. But, i think transformer may be more potential. I will try pre-train like bert and find whether it's help. |
good to hear from you. please keep me update. |
i have tried p1_HierarchicalAttention_model_transformer.py with your test function,but the loss dose not converge。

The text was updated successfully, but these errors were encountered: