Towards Probing Speech-Specific Risks in Large Multimodal Models: A Taxonomy, Benchmark, and Insights
The repo is for EMNLP 2024 paper: Towards Probing Speech-Specific Risks in Large Multimodal Models: A Taxonomy, Benchmark, and Insights
Our speech-specific risk taxonomy includes 8 risk categories under hostility (malicious sarcasm and threats), malicious imitation (age, gender, ethnicity), and stereotypical biases (age, gender, ethnicity).
Due to the safeguards and limitation of existing TTS system, we generate synthetic speech for four risk sub-categories: malicious sarcasm, age, gender, and ethnicity stereotypical biases.
We adopt Yes/No question and Multi-choice question as prompts, detailed in Table 9 of our paper.
We evaluate the capability of five advanced speech LMMs in detecting speech-specific risks, including Qwen-audio-chat, SALMONN-7B/13B, WavLLM, and Gemini-1.5-Pro. Please deploy models/APIs based on the corresponding offical instructions.
We provide an example evaluation in Qwen-Audio-Chat-Sarcasm.
The data access will be granted via submitting a form indicating the researchers’ affiliation and the intention of use. Access the dataset
@article{yang2024towards,
title={Towards Probing Speech-Specific Risks in Large Multimodal Models: A Taxonomy, Benchmark, and Insights},
author={Yang, Hao and Qu, Lizhen and Shareghi, Ehsan and Haffari, Gholamreza},
journal={arXiv preprint arXiv:2406.17430},
year={2024}
}