gender
stringclasses 2
values | speaker_id
stringclasses 40
values | sentence_id
stringclasses 100
values | text
stringclasses 1
value | duration
float32 3.18
26.3
| audio.throat_microphone
audioduration (s) 3.18
26.3
| audio.acoustic_microphone
audioduration (s) 3.18
26.3
|
---|---|---|---|---|---|---|
female
|
p01
|
u00
| 6.0285 | |||
female
|
p01
|
u01
| 11.06125 | |||
female
|
p01
|
u02
| 11.171875 | |||
female
|
p01
|
u03
| 7.653625 | |||
female
|
p01
|
u04
| 11.89825 | |||
female
|
p01
|
u05
| 5.963625 | |||
female
|
p01
|
u06
| 7.766375 | |||
female
|
p01
|
u07
| 5.56875 | |||
female
|
p01
|
u08
| 14.994375 | |||
female
|
p01
|
u09
| 7.83575 | |||
female
|
p01
|
u10
| 7.23525 | |||
female
|
p01
|
u11
| 7.5725 | |||
female
|
p01
|
u12
| 7.780875 | |||
female
|
p01
|
u13
| 11.45175 | |||
female
|
p01
|
u14
| 4.6665 | |||
female
|
p01
|
u15
| 9.484125 | |||
female
|
p01
|
u16
| 6.623 | |||
female
|
p01
|
u17
| 5.2165 | |||
female
|
p01
|
u18
| 7.47925 | |||
female
|
p01
|
u19
| 11.64125 | |||
female
|
p01
|
u20
| 8.684 | |||
female
|
p01
|
u21
| 8.664875 | |||
female
|
p01
|
u22
| 5.425625 | |||
female
|
p01
|
u23
| 6.817625 | |||
female
|
p01
|
u24
| 7.896125 | |||
female
|
p01
|
u25
| 5.472875 | |||
female
|
p01
|
u26
| 10.403125 | |||
female
|
p01
|
u27
| 7.51925 | |||
female
|
p01
|
u28
| 9.746 | |||
female
|
p01
|
u29
| 5.4225 | |||
female
|
p01
|
u30
| 7.74225 | |||
female
|
p01
|
u31
| 6.56225 | |||
female
|
p01
|
u32
| 8.161 | |||
female
|
p01
|
u33
| 7.07225 | |||
female
|
p01
|
u34
| 9.55475 | |||
female
|
p01
|
u35
| 8.881 | |||
female
|
p01
|
u36
| 6.08425 | |||
female
|
p01
|
u37
| 6.013 | |||
female
|
p01
|
u38
| 7.026875 | |||
female
|
p01
|
u39
| 6.25325 | |||
female
|
p01
|
u40
| 5.618125 | |||
female
|
p01
|
u41
| 5.9185 | |||
female
|
p01
|
u42
| 7.741 | |||
female
|
p01
|
u43
| 7.404125 | |||
female
|
p01
|
u44
| 10.541625 | |||
female
|
p01
|
u45
| 11.061 | |||
female
|
p01
|
u46
| 9.948 | |||
female
|
p01
|
u47
| 10.618375 | |||
female
|
p01
|
u48
| 5.9795 | |||
female
|
p01
|
u49
| 11.829125 | |||
female
|
p01
|
u50
| 9.635 | |||
female
|
p01
|
u51
| 5.29225 | |||
female
|
p01
|
u52
| 11.203125 | |||
female
|
p01
|
u53
| 10.152 | |||
female
|
p01
|
u54
| 10.818625 | |||
female
|
p01
|
u55
| 12.3605 | |||
female
|
p01
|
u56
| 9.904375 | |||
female
|
p01
|
u57
| 9.206125 | |||
female
|
p01
|
u58
| 9.0105 | |||
female
|
p01
|
u59
| 12.3045 | |||
female
|
p01
|
u60
| 8.598875 | |||
female
|
p01
|
u61
| 11.822 | |||
female
|
p01
|
u62
| 10.00925 | |||
female
|
p01
|
u63
| 11.319875 | |||
female
|
p01
|
u64
| 4.864 | |||
female
|
p01
|
u65
| 6.293375 | |||
female
|
p01
|
u66
| 9.54075 | |||
female
|
p01
|
u67
| 8.720625 | |||
female
|
p01
|
u68
| 6.16625 | |||
female
|
p01
|
u69
| 10.760375 | |||
female
|
p01
|
u70
| 11.085875 | |||
female
|
p01
|
u71
| 5.86525 | |||
female
|
p01
|
u72
| 11.7195 | |||
female
|
p01
|
u73
| 8.623375 | |||
female
|
p01
|
u74
| 7.34325 | |||
female
|
p01
|
u75
| 7.78925 | |||
female
|
p01
|
u76
| 12.6935 | |||
female
|
p01
|
u77
| 9.783875 | |||
female
|
p01
|
u78
| 5.778875 | |||
female
|
p01
|
u79
| 10.57925 | |||
female
|
p01
|
u80
| 6.784625 | |||
female
|
p01
|
u81
| 13.771125 | |||
female
|
p01
|
u82
| 7.41775 | |||
female
|
p01
|
u83
| 8.12925 | |||
female
|
p01
|
u84
| 10.72025 | |||
female
|
p01
|
u85
| 7.54075 | |||
female
|
p01
|
u86
| 9.01325 | |||
female
|
p01
|
u87
| 8.141375 | |||
female
|
p01
|
u88
| 8.717625 | |||
female
|
p01
|
u89
| 7.910125 | |||
female
|
p01
|
u90
| 9.30025 | |||
female
|
p01
|
u91
| 9.84025 | |||
female
|
p01
|
u92
| 6.492375 | |||
female
|
p01
|
u93
| 5.896875 | |||
female
|
p01
|
u94
| 5.0035 | |||
female
|
p01
|
u95
| 5.19625 | |||
female
|
p01
|
u96
| 5.812875 | |||
female
|
p01
|
u97
| 5.0385 | |||
female
|
p01
|
u98
| 8.753375 | |||
female
|
p01
|
u99
| 9.655875 |
TAPS: Throat and Acoustic Paired Speech Dataset
1. DATASET SUMMARY
The Throat and Acoustic Paired Speech (TAPS) dataset is a standardized corpus designed for deep learning-based speech enhancement, specifically targeting throat microphone recordings. Throat microphones effectively suppress background noise but suffer from high-frequency attenuation due to the low-pass filtering effect of the skin and tissue. The dataset provides paired recordings from 60 native Korean speakers, captured simultaneously using a throat microphone (accelerometer-based) and an acoustic microphone. This dataset facilitates speech enhancement research by enabling the development of models that recover lost high-frequency components and improve intelligibility. Additionally, we introduce a mismatch correction technique to align signals from the two microphones, which enhances model training.
2. Dataset Usage
To use the TAPS dataset, follow the steps below:
2.1 Loading the dataset
You can load the dataset from Hugging Face as follows:
from datasets import load_dataset
dataset = load_dataset("yskim3271/Throat_and_Acoustic_Pairing_Speech_Dataset")
print(dataset)
Example output:
DatasetDict({
train: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 4000
})
dev: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 1000
})
test: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 1000
})
})
2.2 Accessing a sample
Each dataset entry consists of metadata and paired audio recordings. You can access a sample as follows:
sample = dataset["train"][0] # Get the first sample
print(f"Gender: {sample['gender']}")
print(f"Speaker ID: {sample['speaker_id']}")
print(f"Sentence ID: {sample['sentence_id']}")
print(f"Text: {sample['text']}")
print(f"Duration: {sample['duration']} sec")
print(f"Throat Microphone Audio Path: {sample['audio.throat_microphone']['path']}")
print(f"Acoustic Microphone Audio Path: {sample['audio.acoustic_microphone']['path']}")
3. Links and Details
- Project website: Link
- Point of contact: Yunsik Kim ([email protected])
- Collected by: Intelligent Semiconductor and Wearable Devices (ISWD) of the Pohang University of Science and Technology (POSTECH)
- Language: Korean
- Download size: 7.03 GB
- Total audio duration: 15.3 hours
- Number of speech utterances: 6,000
4. Citataion
The BibTeX entry for the dataset is currently being prepared.
5. DATASET STRUCTURE & STATISTICS
- Training Set (40 speakers, 4,000 utterances, 10.2 hours)
- Development Set (10 speakers, 1,000 utterances, 2.5 hours)
- Test Set (10 speakers, 1,000 utterances, 2.6 hours)
- Each set is gender-balanced (50% male, 50% female).
- No speaker overlap across train/dev/test sets.
Dataset Type | Train | Dev | Test |
---|---|---|---|
Number of Speakers | 40 | 10 | 10 |
Number of male speakers | 20 | 5 | 5 |
Mean / standard deviation of the speaker age | 28.5 / 7.3 | 25.6 / 3.0 | 26.2 / 1.4 |
Number of utterances | 4,000 | 1,000 | 1,000 |
Total length of utterances (hours) | 10.2 | 2.5 | 2.6 |
Max / average / min length of utterances (s) | 26.3 / 9.1 / 3.2 | 17.9 / 9.0 / 3.3 | 16.6 / 9.3 / 4.2 |
6. DATA FIELDS
Each dataset entry contains:
gender
: Speaker’s gender (male/female).speaker_id
: Unique speaker identifier (e.g., "p01").sentence_id
: Utterance index (e.g., "u30").text
: Transcription (provided only for test set).duration
: Length of the audio sample.audio.throat_microphone
: Throat microphone signal.audio.acoustic_microphone
: Acoustic microphone signal.
7. DATASET CREATION
7.1 Hardware System for Audio Data Collection
The hardware system simultaneously records signals from a throat microphone and an acoustic microphone, ensuring synchronization.
- Throat microphone: The TDK IIM-42652 MEMS accelerometer captures neck surface vibrations (8 kHz, 16-bit resolution).
- Acoustic microphone: The CUI Devices CMM-4030D-261 MEMS microphone records audio (16 kHz, 24-bit resolution) and is integrated into a peripheral board.
- MCU and data transmission: The STM32F301C8T6 MCU processes signals via SPI (throat microphone) and I²S (acoustic microphone). Data is transmitted to a laptop in real-time through UART communication.
7.2 Sensors Positioning and Recording Environment
- Throat microphone placement: Attached to the supraglottic area of the neck.
- Acoustic microphone position: 30 cm in front of the speaker.
- Recording conditions: Conducted in a controlled, semi-soundproof environment to minimize ambient noise.
- Head rest: A headrest was used to maintain a consistent head position during recording.
- Nylon filter: A nylon pop filter was placed between the speaker and the acoustic microphone to minimize plosive sounds.
- Scripts for Utterances: Sentences were displayed on a screen for participants to read.
7.3 Python-based Software for Data Recordings
The custom-built software facilitates real-time data recording, monitoring, and synchronization of throat and acoustic microphone signals.
- Interface overview: The software displays live waveforms, spectrograms, and synchronization metrics (e.g., SNR, shift values).
- Shift analysis: Visualizations include a shift plot to monitor synchronization between the microphones and a shift histogram for statistical analysis.
- Recording control: Users can manage recordings using controls for file navigation (e.g., Prev, Next, Skip).
- Real-time feedback: Signal quality is monitored with metrics like SNR and synchronization shifts.
7.4 Recorded Audio Data Post-Processing
- Noise reduction: Applied Demucs model to suppress background noise in the acoustic microphone recordings.
- Mismatch correction: Aligned throat and acoustic microphone signals using cross-correlation.
- Silent segment trimming: Removed leading/trailing silence.
7.5 Personal and Sensitive Information
- No personally identifiable information is included.
- Ethical approval: Institutional Review Board (IRB) approval from POSTECH.
- Consent: All participants provided written informed consent.
8. POTENTIAL APPLICATIONS OF THE DATASET
The TAPS dataset enables various speech processing tasks, including:
- Speech enhancement: Improving the intelligibility and quality of throat microphone recordings by recovering high-frequency components.
- Automatic speech recognition (ASR): Enhancing throat microphone speech for better transcription accuracy in noisy environments.
- Speaker Verification: Exploring the effectiveness of throat microphone recordings for identity verification in challenging acoustic environments.
- Downloads last month
- 51