Skip to content

xingweiqu/NLPCC-2024-Shared-Task-9

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

NLPCC-2024-Shared-Task-9

Chinese Metaphor Generation Task

Task Introduction

This task is designed to generate Chinese metaphors using machine learning techniques by either effectively identifying the ground or vehicle in the metaphoric relation. It is divided into two subtasks:

SUBTASK 1. Metaphor Generation, which involves creating a metaphor from a provided tuple consisting of TENOR, GROUND, and VEHICLE. The goal here is to synthesize a metaphor that aptly connects the subject (TENOR) with the object (VEHICLE), guided by the concept of the GROUND.

SUBTASK 2. Metaphor Components Identification, aimed at extracting the TENORs, GROUNDs, and VEHICLEs from a metaphorical sentence. This component requires the identification of metaphor elements that correspond to the specified grounds.

Data Description & Rules

Train Data

In our training data we provide the format like:

    {
        "metaphor_id": 0,
        "context": "天空随意飘着几朵白云 棉花糖一样如你般的纯净 打开手机看到今天天气晴 如果有可能想带你去远行 某年某月某日几点零几",
        "tenor": "白云",
        "vehicle": "棉花糖",
        "ground": "白色的外形",
        "sub_id": 0
    },

we allow you to use extra data apart from our provided training data

Validation & Test Data:

We will provide you with multiple-choice questions, each containing four options. Based on the descriptions of two tasks, task 1 and task 2, please select the option that most closely aligns with the given metaphor.

Subtask 1: Data Format and Files

Data format details and file links for Subtask 1 are provided below:

  • Validation File: The validation data includes the following fields: 本体, 喻体, 共性, 正确选项, 选项A, 选项B, 选项C, 选项D.
    View the Validation File

  • Test File: The test data includes these fields: 本体, 喻体, 共性, 选项A, 选项B, 选项C, 选项D. You will need to determine the correct option among the four provided.
    View the Test File

Subtask 2: Data Format and Files

Data format details and file links for Subtask 2 are provided below:

  • Validation File: The validation data includes the following fields: 比喻句, 正确答案, 答案A_本体, 答案A_喻体, 答案A_共性, 答案B_本体, 答案B_喻体, 答案B_共性, 答案C_本体, 答案C_喻体, 答案C_共性, 答案D_本体, 答案D_喻体, 答案D_共性.
    View the Validation File

  • Test File: The test data includes these fields: 比喻句, 答案A_本体, 答案A_喻体, 答案A_共性, 答案B_本体, 答案B_喻体, 答案B_共性, 答案C_本体, 答案C_喻体, 答案C_共性, 答案D_本体, 答案D_喻体, 答案D_共性. You will need to determine the correct option among the four provided.
    View the Test File

Evaluation

For each task we provide two tracks to evaluation:

  • Track1 : LLM Track

We encourage the use of large models to directly generate the options. You may use your own prompt, but please use a common prompt during the answer phase: "The answer is {}."

  • Track2: Rule based Track

You can also use traditional linguistic rules or machine learning-based methods to directly compare and derive conclusions for options A, B, C, and D.

For more information related to this dataset, please refer to our paper: CMDAG: A Chinese Metaphor Dataset with Annotated Grounds as CoT for Boosting Metaphor Generation. If there are any differences between the paper and this page, the content of this page should prevail.

Submission & Evaluation

For submission, the following materials should be packaged as one zip file, the final result must be a json file named with format yourteamid_task{1or2}_track{1or2}.json and sent to quxingwei25@gmail.com:

  • Submission File: For the Large Language Model Track, we allow you to give us the prompt based generations like. we will extract first letter from [A,B,C,D] as shown:
[
    {
        "metaphor_id": 0,
        "answer": "The answer is A",
    },
    {
        "metaphor_id": 1,
        "answer": "The answer is B",
    },
    ...
]

For the Rule based Track, The answer must be the only choice from [A,B,C,D] as shown:

[
    {
        "metaphor_id": 0,
        "answer": "A",
    },
    {
        "metaphor_id": 1,
        "answer": "B",
    },
    ...
]

  • Code: The code folder should contain all the codes of data augmentation, data processing, model training and model inference.

Evaluation

The top 3 participating teams in each task and track will be certificated by NLPCC and CCF-NLP.

Benchmark Results

We provide benchmark results for the MAP-NEO and CT-LLM models on two subtasks. You can find the example code & results file for these models in their respective repositories:

Results

Model Subtask 1 Subtask 2
MAP-NEO 48.4 37.8
CT-LLM 22.8 5.6

Leaderboard

Subtask 1 Track 1

Final result:

排名 队伍名A榜B榜总分
1 kangreen98.898.098.4
2 ShaunTheSheep96.29696.1
3 YNU-HPCC96.695.295.9

Subtask 1 Track 2

Final result:

排名 队伍名A榜B榜总分
1 YNU-HPCC98.497.497.9

Subtask 2 Track 1

Final result:

排名 队伍名A榜B榜总分
1 YNU-HPCC96.693.695.1
2 ZZU_NLP93.891.8292.8
3 ShaunTheSheep92.292.8192.5
4 kangreen92.891.892.3

Subtask 2 Track 2

Final result:

排名 队伍名A榜B榜总分
1 YNU-HPCC9593.294.1

Contact & Citation

If your publication employs our dataset, please cite the following article:

@article{Shao2024CMDAGAC,
  title={CMDAG: A Chinese Metaphor Dataset with Annotated Grounds as CoT for Boosting Metaphor Generation},
  author={Yujie Shao and Xinrong Yao and Xingwei Qu and Chenghua Lin and Shi Wang and Stephen W. Huang and Ge Zhang and Jie Fu},
  journal={ArXiv},
  year={2024},
  volume={abs/2402.13145},
  url={https://api.semanticscholar.org/CorpusID:267759689}
}

If you have any questions about this task, please email to quxingwei25@gmail.com.

About

Chinese Metaphor Generation Task

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages