Skip to content

JD-AI-Research-NLP/P3LM

Repository files navigation

P3LM

The code for the EMNLP2022(Findings) paper:P3LM: Probabilistically Permuted Prophet Language Modeling for Generative Pre-Training image

Requirements

  1. python==3.6.8
  2. pytorch==1.5.0+cu101

GLGE Data

download here

Models on GLGE Used in the Paper

download here

Our Results

image

Reproduce the Reported Results

  1. Pre-process the GLGE data and put it in the ./P2DeNet/glge/ folder.
  2. Download the trained models, and place them in ./P2DeNet/glge/models/[DATASET]/ respectively.
  3. Run the following command:
python finetune.generation.py test [DATASET NAME]

Finetune the Models Yourself

python finetune.generation.py train [DATASET NAME]

Reference

Thanks for your citation:

@inproceedings{wang-etal-2020-learning-decouple,
    title = "Learning to Decouple Relations: Few-Shot Relation Classification with Entity-Guided Attention and Confusion-Aware Training",
    author = "Bao, Junwei  and
      Wang, Yifan  and
      Ying, Jiangyong  and
      Gong, Yeyun  and
      Zhao, Jing  and
      Wu, Youzheng  and
      He, Xiaodong  and
      Zhou, Bowen",
    booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
    year = "2021",
    abstract = "Conventional left-to-right (L2R) generative pre-training methods face two issues during decoding: limited to unidirectional target sequence modeling, and constrained on strong local dependencies. In this paper, we propose P2DeNet, a permutation over prophet decoding net, which strengthens the modeling of bi-directional information and long token dependencies in target sequences, for generative pre-training. Specifically, P2DeNet learns to generate tokens in permuted order upon an order-aware transformer decoder, as well as the corresponding future N tokens with a multi-stream attention mechanism. Extensive experiments are conducted on the GLGE benchmark, which includes four datasets for summarization, two for question generation, one for conversational question answering, and one for dialog response generation, where P2DeNet achieves state-of-the-art results compared with published methods.",
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published