| | --- |
| | dataset_info: |
| | - config_name: role_playing |
| | features: |
| | - name: ID |
| | dtype: int64 |
| | - name: text_0 |
| | dtype: string |
| | - name: text_1 |
| | dtype: string |
| | - name: audio_0 |
| | dtype: |
| | audio: |
| | sampling_rate: 16000 |
| | - name: audio_1 |
| | dtype: |
| | audio: |
| | sampling_rate: 16000 |
| | - name: source |
| | dtype: string |
| | - name: speaker1 |
| | dtype: string |
| | - name: speaker2 |
| | dtype: string |
| | splits: |
| | - name: test |
| | num_bytes: 182310504.0 |
| | num_examples: 20 |
| | download_size: 148908359 |
| | dataset_size: 182310504.0 |
| | - config_name: voice_instruction_following |
| | features: |
| | - name: ID |
| | dtype: int64 |
| | - name: text_1 |
| | dtype: string |
| | - name: text_2 |
| | dtype: string |
| | - name: audio_1 |
| | dtype: |
| | audio: |
| | sampling_rate: 16000 |
| | - name: audio_2 |
| | dtype: |
| | audio: |
| | sampling_rate: 16000 |
| | splits: |
| | - name: test |
| | num_bytes: 36665909.0 |
| | num_examples: 20 |
| | download_size: 35109899 |
| | dataset_size: 36665909.0 |
| | configs: |
| | - config_name: role_playing |
| | data_files: |
| | - split: test |
| | path: role_playing/test-* |
| | - config_name: voice_instruction_following |
| | data_files: |
| | - split: test |
| | path: voice_instruction_following/test-* |
| | --- |
| | # StyleSet |
| |
|
| | **WARNING**: This dataset contains some profane words. |
| |
|
| | **A spoken language benchmark for evaluating speaking-style-related speech generation** |
| | Released in our paper, [Audio-Aware Large Language Models as Judges for Speaking Styles](https://arxiv.org/abs/2506.05984) |
| |
|
| | This dataset is released by NTU Speech Lab under the MIT license. |
| |
|
| |  |
| |
|
| | --- |
| |
|
| | ## Tasks |
| |
|
| | 1. **Voice Style Instruction Following** |
| | - Reproduce a given sentence verbatim. |
| | - Match specified prosodic styles (emotion, volume, pace, emphasis, pitch, non-verbal cues). |
| |
|
| | 2. **Role Playing** |
| | - Continue a two-turn dialogue prompt in character. |
| | - Generate the next utterance with appropriate prosody and style. |
| | - The dataset is modified from IEMOCAP with the consent of the authors. Please refer to [IEMOCAP](https://sail.usc.edu/iemocap/) for details and the original data of IEMOCAP. We do not redistribute the data here. |
| |
|
| | --- |
| |
|
| | ## Evaluation |
| |
|
| | We use ALLM-as-a-judge for evaluation. Currently, we found that `gemini-2.5-pro-0506` reaches the best agreement with human evaluators. |
| | The complete evaluation prompt and evaluation pipelines can be found in Table 3 to Table 5 in our paper. |
| |
|
| |
|
| | ## Citation |
| |
|
| | If you use StyleSet or find ALLM-as-a-judge useful, please cite our paper by |
| | ``` |
| | @misc{chiang2025audioawarelargelanguagemodels, |
| | title={Audio-Aware Large Language Models as Judges for Speaking Styles}, |
| | author={Cheng-Han Chiang and Xiaofei Wang and Chung-Ching Lin and Kevin Lin and Linjie Li and Radu Kopetz and Yao Qian and Zhendong Wang and Zhengyuan Yang and Hung-yi Lee and Lijuan Wang}, |
| | year={2025}, |
| | eprint={2506.05984}, |
| | archivePrefix={arXiv}, |
| | primaryClass={eess.AS}, |
| | url={https://arxiv.org/abs/2506.05984}, |
| | } |
| | ``` |