gender
stringclasses 2
values | speaker_id
stringclasses 40
values | sentence_id
stringclasses 100
values | text
stringclasses 1
value | duration
float32 3.18
26.3
| audio.throat_microphone
audioduration (s) 3.18
26.3
| audio.acoustic_microphone
audioduration (s) 3.18
26.3
|
---|---|---|---|---|---|---|
female | p01 | u00 | 6.0285 | |||
female | p01 | u01 | 11.06125 | |||
female | p01 | u02 | 11.171875 | |||
female | p01 | u03 | 7.653625 | |||
female | p01 | u04 | 11.89825 | |||
female | p01 | u05 | 5.963625 | |||
female | p01 | u06 | 7.766375 | |||
female | p01 | u07 | 5.56875 | |||
female | p01 | u08 | 14.994375 | |||
female | p01 | u09 | 7.83575 | |||
female | p01 | u10 | 7.23525 | |||
female | p01 | u11 | 7.5725 | |||
female | p01 | u12 | 7.780875 | |||
female | p01 | u13 | 11.45175 | |||
female | p01 | u14 | 4.6665 | |||
female | p01 | u15 | 9.484125 | |||
female | p01 | u16 | 6.623 | |||
female | p01 | u17 | 5.2165 | |||
female | p01 | u18 | 7.47925 | |||
female | p01 | u19 | 11.64125 | |||
female | p01 | u20 | 8.684 | |||
female | p01 | u21 | 8.664875 | |||
female | p01 | u22 | 5.425625 | |||
female | p01 | u23 | 6.817625 | |||
female | p01 | u24 | 7.896125 | |||
female | p01 | u25 | 5.472875 | |||
female | p01 | u26 | 10.403125 | |||
female | p01 | u27 | 7.51925 | |||
female | p01 | u28 | 9.746 | |||
female | p01 | u29 | 5.4225 | |||
female | p01 | u30 | 7.74225 | |||
female | p01 | u31 | 6.56225 | |||
female | p01 | u32 | 8.161 | |||
female | p01 | u33 | 7.07225 | |||
female | p01 | u34 | 9.55475 | |||
female | p01 | u35 | 8.881 | |||
female | p01 | u36 | 6.08425 | |||
female | p01 | u37 | 6.013 | |||
female | p01 | u38 | 7.026875 | |||
female | p01 | u39 | 6.25325 | |||
female | p01 | u40 | 5.618125 | |||
female | p01 | u41 | 5.9185 | |||
female | p01 | u42 | 7.741 | |||
female | p01 | u43 | 7.404125 | |||
female | p01 | u44 | 10.541625 | |||
female | p01 | u45 | 11.061 | |||
female | p01 | u46 | 9.948 | |||
female | p01 | u47 | 10.618375 | |||
female | p01 | u48 | 5.9795 | |||
female | p01 | u49 | 11.829125 | |||
female | p01 | u50 | 9.635 | |||
female | p01 | u51 | 5.29225 | |||
female | p01 | u52 | 11.203125 | |||
female | p01 | u53 | 10.152 | |||
female | p01 | u54 | 10.818625 | |||
female | p01 | u55 | 12.3605 | |||
female | p01 | u56 | 9.904375 | |||
female | p01 | u57 | 9.206125 | |||
female | p01 | u58 | 9.0105 | |||
female | p01 | u59 | 12.3045 | |||
female | p01 | u60 | 8.598875 | |||
female | p01 | u61 | 11.822 | |||
female | p01 | u62 | 10.00925 | |||
female | p01 | u63 | 11.319875 | |||
female | p01 | u64 | 4.864 | |||
female | p01 | u65 | 6.293375 | |||
female | p01 | u66 | 9.54075 | |||
female | p01 | u67 | 8.720625 | |||
female | p01 | u68 | 6.16625 | |||
female | p01 | u69 | 10.760375 | |||
female | p01 | u70 | 11.085875 | |||
female | p01 | u71 | 5.86525 | |||
female | p01 | u72 | 11.7195 | |||
female | p01 | u73 | 8.623375 | |||
female | p01 | u74 | 7.34325 | |||
female | p01 | u75 | 7.78925 | |||
female | p01 | u76 | 12.6935 | |||
female | p01 | u77 | 9.783875 | |||
female | p01 | u78 | 5.778875 | |||
female | p01 | u79 | 10.57925 | |||
female | p01 | u80 | 6.784625 | |||
female | p01 | u81 | 13.771125 | |||
female | p01 | u82 | 7.41775 | |||
female | p01 | u83 | 8.12925 | |||
female | p01 | u84 | 10.72025 | |||
female | p01 | u85 | 7.54075 | |||
female | p01 | u86 | 9.01325 | |||
female | p01 | u87 | 8.141375 | |||
female | p01 | u88 | 8.717625 | |||
female | p01 | u89 | 7.910125 | |||
female | p01 | u90 | 9.30025 | |||
female | p01 | u91 | 9.84025 | |||
female | p01 | u92 | 6.492375 | |||
female | p01 | u93 | 5.896875 | |||
female | p01 | u94 | 5.0035 | |||
female | p01 | u95 | 5.19625 | |||
female | p01 | u96 | 5.812875 | |||
female | p01 | u97 | 5.0385 | |||
female | p01 | u98 | 8.753375 | |||
female | p01 | u99 | 9.655875 |
TAPS: Throat and Acoustic Paired Speech Dataset
1. DATASET SUMMARY
The Throat and Acoustic Paired Speech (TAPS) dataset is a standardized corpus designed for deep learning-based speech enhancement, specifically targeting throat microphone recordings. Throat microphones effectively suppress background noise but suffer from high-frequency attenuation due to the low-pass filtering effect of the skin and tissue. The dataset provides paired recordings from 60 native Korean speakers, captured simultaneously using a throat microphone (accelerometer-based) and an acoustic microphone. This dataset facilitates speech enhancement research by enabling the development of models that recover lost high-frequency components and improve intelligibility. Additionally, we introduce a mismatch correction technique to align signals from the two microphones, which enhances model training.
2. Dataset Usage
To use the TAPS dataset, follow the steps below:
2.1 Loading the dataset
You can load the dataset from Hugging Face as follows:
from datasets import load_dataset
dataset = load_dataset("yskim3271/Throat_and_Acoustic_Pairing_Speech_Dataset")
print(dataset)
Example output:
DatasetDict({
train: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 4000
})
dev: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 1000
})
test: Dataset({
features: ['gender', 'speaker_id', 'sentence_id', 'text', 'duration', 'audio.throat_microphone', 'audio.acoustic_microphone'],
num_rows: 1000
})
})
2.2 Accessing a sample
Each dataset entry consists of metadata and paired audio recordings. You can access a sample as follows:
sample = dataset["train"][0] # Get the first sample
print(f"Gender: {sample['gender']}")
print(f"Speaker ID: {sample['speaker_id']}")
print(f"Sentence ID: {sample['sentence_id']}")
print(f"Text: {sample['text']}")
print(f"Duration: {sample['duration']} sec")
print(f"Throat Microphone Audio Path: {sample['audio.throat_microphone']['path']}")
print(f"Acoustic Microphone Audio Path: {sample['audio.acoustic_microphone']['path']}")
3. Links and Details
- Project website: Link
- Point of contact: Yunsik Kim ([email protected])
- Collected by: Intelligent Semiconductor and Wearable Devices (ISWD) of the Pohang University of Science and Technology (POSTECH)
- Language: Korean
- Download size: 7.03 GB
- Total audio duration: 15.3 hours
- Number of speech utterances: 6,000
4. Citataion
The BibTeX entry for the dataset is currently being prepared.
5. DATASET STRUCTURE & STATISTICS
- Training Set (40 speakers, 4,000 utterances, 10.2 hours)
- Development Set (10 speakers, 1,000 utterances, 2.5 hours)
- Test Set (10 speakers, 1,000 utterances, 2.6 hours)
- Each set is gender-balanced (50% male, 50% female).
- No speaker overlap across train/dev/test sets.
Dataset Type | Train | Dev | Test |
---|---|---|---|
Number of Speakers | 40 | 10 | 10 |
Number of male speakers | 20 | 5 | 5 |
Mean / standard deviation of the speaker age | 28.5 / 7.3 | 25.6 / 3.0 | 26.2 / 1.4 |
Number of utterances | 4,000 | 1,000 | 1,000 |
Total length of utterances (hours) | 10.2 | 2.5 | 2.6 |
Max / average / min length of utterances (s) | 26.3 / 9.1 / 3.2 | 17.9 / 9.0 / 3.3 | 16.6 / 9.3 / 4.2 |
6. DATA FIELDS
Each dataset entry contains:
gender
: Speaker’s gender (male/female).speaker_id
: Unique speaker identifier (e.g., "p01").sentence_id
: Utterance index (e.g., "u30").text
: Transcription (provided only for test set).duration
: Length of the audio sample.audio.throat_microphone
: Throat microphone signal.audio.acoustic_microphone
: Acoustic microphone signal.
7. DATASET CREATION
7.1 Hardware System for Audio Data Collection
The hardware system simultaneously records signals from a throat microphone and an acoustic microphone, ensuring synchronization.
- Throat microphone: The TDK IIM-42652 MEMS accelerometer captures neck surface vibrations (8 kHz, 16-bit resolution).
- Acoustic microphone: The CUI Devices CMM-4030D-261 MEMS microphone records audio (16 kHz, 24-bit resolution) and is integrated into a peripheral board.
- MCU and data transmission: The STM32F301C8T6 MCU processes signals via SPI (throat microphone) and I²S (acoustic microphone). Data is transmitted to a laptop in real-time through UART communication.
7.2 Sensors Positioning and Recording Environment
- Throat microphone placement: Attached to the supraglottic area of the neck.
- Acoustic microphone position: 30 cm in front of the speaker.
- Recording conditions: Conducted in a controlled, semi-soundproof environment to minimize ambient noise.
- Head rest: A headrest was used to maintain a consistent head position during recording.
- Nylon filter: A nylon pop filter was placed between the speaker and the acoustic microphone to minimize plosive sounds.
- Scripts for Utterances: Sentences were displayed on a screen for participants to read.
7.3 Python-based Software for Data Recordings
The custom-built software facilitates real-time data recording, monitoring, and synchronization of throat and acoustic microphone signals.
- Interface overview: The software displays live waveforms, spectrograms, and synchronization metrics (e.g., SNR, shift values).
- Shift analysis: Visualizations include a shift plot to monitor synchronization between the microphones and a shift histogram for statistical analysis.
- Recording control: Users can manage recordings using controls for file navigation (e.g., Prev, Next, Skip).
- Real-time feedback: Signal quality is monitored with metrics like SNR and synchronization shifts.
7.4 Recorded Audio Data Post-Processing
- Noise reduction: Applied Demucs model to suppress background noise in the acoustic microphone recordings.
- Mismatch correction: Aligned throat and acoustic microphone signals using cross-correlation.
- Silent segment trimming: Removed leading/trailing silence.
7.5 Personal and Sensitive Information
- No personally identifiable information is included.
- Ethical approval: Institutional Review Board (IRB) approval from POSTECH.
- Consent: All participants provided written informed consent.
8. POTENTIAL APPLICATIONS OF THE DATASET
The TAPS dataset enables various speech processing tasks, including:
- Speech enhancement: Improving the intelligibility and quality of throat microphone recordings by recovering high-frequency components.
- Automatic speech recognition (ASR): Enhancing throat microphone speech for better transcription accuracy in noisy environments.
- Speaker Verification: Exploring the effectiveness of throat microphone recordings for identity verification in challenging acoustic environments.
- Downloads last month
- 194