lugan's picture
Update README.md
8d30f66 verified
---
license: mit
task_categories:
- audio-classification
language:
- en
- zh
tags:
- keyword-spotting
- tts-synthetic
- multilingual
- speech-commands
- tinyml
- edge-ai
---
# SynTTS-Commands: A Multilingual Synthetic Speech Command Dataset
<!-- Badges Row -->
![Python](https://img.shields.io/badge/python-3.8%2B-blue)
![Domain](https://img.shields.io/badge/domain-Speech--TinyML-orange)
![Size](https://img.shields.io/badge/size-35.76%20GB-purple)
![Utterances](https://img.shields.io/badge/utterances-384K-yellow)
![Speakers](https://img.shields.io/badge/speakers-8,100-lightblue)
[![arXiv](https://img.shields.io/badge/arXiv-2511.07821-b31b1b.svg)](https://arxiv.org/abs/2511.07821)
[![Benchmarks](https://img.shields.io/badge/πŸ€—%20Hugging%20Face-Benchmarks-ffd21e)](https://huggingface.co/lugan/SynTTS-Commands-Media-Benchmarks)
[![Code](https://img.shields.io/badge/GitHub-Repository-181717)](https://github.com/lugan113/SynTTS-Commands-Official)
[![License](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
## πŸ“– Introduction
**SynTTS-Commands** is a large-scale, multilingual synthetic speech command dataset specifically designed for low-power Keyword Spotting (KWS) and speech command recognition tasks. As presented in the paper [SynTTS-Commands: A Public Dataset for On-Device KWS via TTS-Synthesized Multilingual Speech](https://huggingface.co/papers/2511.07821), this dataset is generated using advanced Text-to-Speech (TTS) technologies, aiming to address the scarcity of high-quality training data in the fields of TinyML and Edge AI.
### 🎯 Core Features
- **Multilingual Coverage**: Includes bilingual speech commands in both **Chinese** and **English**.
- **High-Quality Synthesis**: Generated via advanced TTS technology, ensuring high naturalness in speech.
- **Speaker Diversity**: Incorporates multiple acoustic feature sources to ensure a rich variety of speaker styles.
- **Real-World Scenarios**: Commands are designed for practical applications, including smart homes, in-car systems, and multimedia control.
- **Rigorous Quality Assurance**: All speech data has been screened via ASR models combined with manual human verification.
## πŸ“Š Dataset Overview
### Statistics
The **SynTTS-Commands-Media-Dataset** contains a total of **384,621 speech samples**, covering **48 distinct multimedia control commands**. It is divided into four subsets with the following distribution:
| Subset | Speakers | Commands | Samples | Duration (hrs) | Size (GB) |
|------|----------|--------|----------|------------|----------|
| Free-ST-Chinese | 855 | 25 | 21,214 | 6.82 | 2.19 |
| Free-ST-English | 855 | 23 | 19,228 | 4.88 | 1.57 |
| VoxCeleb1&2-Chinese | 7,245 | 25 | 180,331 | 58.03 | 18.6 |
| VoxCeleb1&2-English | 7,245 | 23 | 163,848 | 41.6 | 13.4 |
| **Total** | **8,100** | **48** | **384,621** | **111.33** | **35.76** |
### Dataset Highlights
- **Massive Scale**: Totaling **111.33 hours** and **35.76 GB** of synthetic speech data, making it one of the largest synthetic speech command datasets for academic research.
- **Extensive Speaker Diversity**: Covers **8,100 unique speakers**, spanning various accent groups, age ranges, and recording conditions.
- **Multi-Dimensional Research Support**: The four-subset structure enables research into cross-lingual speaker adaptation, speaker diversity effects, and acoustic robustness in different recording environments.
- **Application-Oriented**: Specifically focused on multimedia playback control scenarios, providing high-quality training data for real-world deployment.
### Directory Structure
```text
SynTTS-Commands-Media-Dataset/
β”œβ”€β”€ Free_ST_Chinese/ # 21,214 Chinese media control samples (855 speakers)
β”œβ”€β”€ Free_ST_English/ # 19,228 English media control samples (855 speakers)
β”œβ”€β”€ VoxCeleb1&2_Chinese/ # 180,331 Chinese media control samples (7,245 speakers)
β”œβ”€β”€ VoxCeleb1&2_English/ # 163,848 English media control samples (7,245 speakers)
β”œβ”€β”€ reviewed_bad/ # Rejected speech samples (failed quality audit)
β”œβ”€β”€ splits_by_language/ # Dataset splits organized by language
β”‚ β”œβ”€β”€ train/ # Training set
β”‚ β”œβ”€β”€ val/ # Validation set
β”‚ └── test/ # Test set
└── comprehensive_metadata.csv # Complete metadata file
```
🎯 Media Command Categories
English Media Control Commands (23 Classes)
Playback Control: "Play", "Pause", "Resume", "Play from start", "Repeat song"
Navigation: "Previous track", "Next track", "Last song", "Skip song", "Jump to first track"
Volume Control: "Volume up", "Volume down", "Mute", "Set volume to 50%", "Max volume"
Communication: "Answer call", "Hang up", "Decline call"
Wake Words: "Hey Siri", "OK Google", "Hey Google", "Alexa", "Hi Bixby"
Chinese Media Control Commands (25 Classes)
Playback Control: "ζ’­ζ”Ύ", "ζš‚εœ", "η»§η»­ζ’­ζ”Ύ", "δ»Žε€΄ζ’­ζ”Ύ", "单曲εΎͺ环"
Navigation: "δΈŠδΈ€ι¦–", "δΈ‹δΈ€ι¦–", "δΈŠδΈ€ζ›²", "δΈ‹δΈ€ζ›²", "θ·³εˆ°η¬¬δΈ€ι¦–", "ζ’­ζ”ΎδΈŠδΈ€εΌ δΈ“θΎ‘"
Volume Control: "ε’žε€§ιŸ³ι‡", "ε‡ε°ιŸ³ι‡", "ι™ιŸ³", "ιŸ³ι‡θ°ƒεˆ°50%", "ιŸ³ι‡ζœ€ε€§"
Communication: "ζŽ₯听甡话", "ζŒ‚ζ–­η”΅θ―", "ζ‹’ζŽ₯ζ₯η”΅"
Wake Words: "小爱同学", "Hello 小智", "小艺小艺", "ε—¨ δΈ‰ζ˜Ÿε°θ΄", "小度小度", "倩猫精灡"
## πŸ“ˆ Benchmark Results and Analysis
We present a comprehensive benchmark of **six representative acoustic models** on the SynTTS-Commands-Media Dataset across both English (EN) and Chinese (ZH) subsets. All models are evaluated in terms of **classification accuracy**, **cross-entropy loss**, and **parameter count**, providing insights into the trade-offs between performance and model complexity in multilingual voice command recognition.
### Performance Summary
| Model | EN Loss | EN Accuracy | EN Params | ZH Loss | ZH Accuracy | ZH Params |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| **MicroCNN** | 0.2304 | 93.22% | 4,189 | 0.5579 | 80.14% | 4,255 |
| **DS-CNN** | 0.0166 | 99.46% | 30,103 | 0.0677 | 97.18% | 30,361 |
| **TC-ResNet** | 0.0347 | 98.87% | 68,431 | 0.0884 | 96.56% | 68,561 |
| **CRNN** | **0.0163** | **99.50%** | 1.08M | 0.0636 | **97.42%** | 1.08M |
| **MobileNet-V1** | 0.0167 | **99.50%** | 2.65M | **0.0552** | 97.92% | 2.65M |
| **EfficientNet** | 0.0182 | 99.41% | 4.72M | 0.0701 | 97.93% | 4.72M |
## πŸ—ΊοΈ Roadmap & Future Expansion
We are expanding SynTTS-Commands beyond multimedia to support broader Edge AI applications.
πŸ‘‰ **[Click here to view our detailed Future Work Plan & Command List](https://github.com/lugan113/SynTTS-Commands-Official/blob/main/Future_Work_Plan.md)**
Our upcoming domains include:
* 🏠 **Smart Home:** Far-field commands for lighting and appliances.
* πŸš— **In-Vehicle:** Robust commands optimized for high-noise driving environments.
* πŸš‘ **Urgent Assistance:** Safety-critical keywords (e.g., "Call 911", "Help me") focusing on high recall.
We invite the community to review our [Command Roadmap](https://github.com/lugan113/SynTTS-Commands-Official/blob/main/Future_Work_Plan.md) and suggest additional keywords!
πŸ“œ Citation
If you use this dataset in your research, please cite our paper:
@misc{gan2025synttscommands,
title={SynTTS-Commands: A Public Dataset for On-Device KWS via TTS-Synthesized Multilingual Speech},
author={Lu Gan and Xi Li},
year={2025},
eprint={2511.07821},
archivePrefix={arXiv},
primaryClass={cs.SD},
url={https://arxiv.org/abs/2511.07821},
doi={10.48550/arXiv.2511.07821}
}