Datasets:

Formats:
parquet
Languages:
English
ArXiv:
DOI:
Libraries:
Datasets
Dask
License:
File size: 14,750 Bytes
a063a9e
ae77b12
a063a9e
 
 
 
 
 
f540cd7
 
a063a9e
 
 
 
 
 
 
 
 
 
 
 
 
6c75fd0
a063a9e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6c75fd0
7af4149
a063a9e
7af4149
 
a063a9e
6c75fd0
7af4149
6c75fd0
 
a063a9e
 
 
 
 
 
 
 
 
4a46b48
 
aac9a89
4a46b48
 
 
 
 
1ba6e94
 
 
 
 
 
 
a063a9e
90b2d94
4a46b48
0630098
1ba6e94
ae77b12
1ba6e94
 
e620632
0630098
 
6b8bf7b
90b2d94
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9778af7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
90b2d94
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1ba6e94
 
90b2d94
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ecbe24d
 
 
 
 
 
 
 
 
 
 
 
 
90b2d94
 
 
 
 
 
 
 
905ecee
90b2d94
905ecee
 
 
 
 
 
 
 
 
90b2d94
905ecee
 
 
 
90b2d94
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d3ca3f8
 
 
 
82b5fa7
 
 
ae77b12
 
 
 
 
 
 
 
82b5fa7
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
---
arxiv: 2510.23141
dataset_info:
  features:
  - name: audio
    dtype:
      audio:
        sampling_rate: 16000
  - name: filename
    dtype: string
  - name: room
    dtype: string
  - name: room_description
    dtype: string
  - name: room_volume
    dtype: float32
  - name: source
    dtype: string
  - name: source_position
    list: float32
  - name: receiver
    dtype: string
  - name: receiver_position
    list: float32
  - name: direct_path_length_m
    dtype: float32
  - name: rir_format
    dtype: string
  - name: T30
    list: float32
  - name: EDT
    list: float32
  - name: C50
    list: float32
  - name: absorption
    list: float32
  - name: center_frequencies
    list: float32
  - name: librispeech_split
    dtype: string
  - name: librispeech_file
    dtype: string
  - name: transcript
    dtype: string
  splits:
  - name: speech_mono
    num_bytes: 3084427987
    num_examples: 3085
  - name: speech_6ch
    num_bytes: 18498009661
    num_examples: 3085
  - name: speech_hoa8
    num_bytes: 249702475180
    num_examples: 3085
  download_size: 273472049365
  dataset_size: 271284912828
configs:
- config_name: default
  data_files:
  - split: speech_mono
    path: data/speech_mono-*
  - split: speech_6ch
    path: data/speech_6ch-*
  - split: speech_hoa8
    path: data/speech_hoa8-*
language:
- en
license: cc-by-4.0
task_categories:
- automatic-speech-recognition
pretty_name: Treble10-Speech
size_categories:
- 1K<n<10K
tags:
- audio
- speech
- acoustics
source_datasets:
- treble-technologies/Treble10-RIR
- openslr/librispeech_asr
---

# **Treble10-Speech (16 kHz)**

## Dataset Description
 - **Paper:** https://arxiv.org/abs/2510.23141
 - **Point of contact:** [email protected]

[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1pSJHitYAOGIv2uzr7kGNDxqVflWvTQ2c?usp=sharing)


The **Treble10-Speech** dataset is a dataset for automatic speech recognition (ASR), containing pre-convolved speech files using high fidelity room-acoustic simulations from the [Treble10-RIR dataset](https://huggingface.co/datasets/treble-technologies/Treble10-RIR) with 10 different furnished rooms: 2 bathrooms, 2 bedrooms, 2 living rooms with hallway, 2 living rooms without hallway, 2 meeting rooms. 
The room volumes range between 14 and 46 m3, resulting in reverberation times between 0.17 and 0.84 s.

## Examples: accessing a reverberant mono speech file:
```python
from datasets import load_dataset, Audio
import matplotlib.pyplot as plt
import numpy as np
ds = load_dataset(
    "treble-technologies/Treble10-Speech",
    split="speech_mono",
    streaming=True,
)
ds = ds.cast_column("audio", Audio())
# Read the samples from the TorchCodec decoder object:
rec = next(iter(ds))
samples = rec["audio"].get_all_samples()
speech_mono = samples.data
sr = samples.sample_rate
print(f"Mono speech has this shape: {speech_mono.shape}, and a sampling rate of {sr} Hz.")
# We can access and compare individual channels from the mono device like this
t_axis = np.arange(speech_mono.shape[1]) / sr
plt.figure()
plt.plot(t_axis, speech_mono.numpy().T, label="Mono speech")
plt.xlabel("Time (s)")
plt.ylabel("Amplitude")
plt.legend()
plt.show()
```

## Example: Accessing a reverberant speech file encoded to 8th-order Ambisonics
```python
from datasets import load_dataset, Audio
import io, soundfile as sf

# Load dataset in streaming mode
ds = load_dataset("treble-technologies/Treble10-Speech", split="speech_hoa8", streaming=True)

# Disable automatic decoding (we'll do it manually)
ds = ds.cast_column("audio", Audio(decode=False))

# Get one sample from the iterator
sample = next(iter(ds))

# Fetch raw audio bytes
audio_bytes = sample["audio"]["bytes"]

# Some older datasets may not have "bytes", so fall back to reading from the file
if audio_bytes is None:
    # Use huggingface's file object directly
    with sample["audio"]["path"].open("rb") as f:
        audio_bytes = f.read()

# Decode the HOA audio directly from memory
rir_hoa, sr = sf.read(io.BytesIO(audio_bytes))
print(f"Loaded HOA audio: shape={rir_hoa.shape}, sr={sr}")
```

## Example: Accessing a reverberant speech file at the microphones of a 6-channel device
```python
from datasets import load_dataset, Audio
import matplotlib.pyplot as plt
import numpy as np
ds = load_dataset(
    "treble-technologies/Treble10-Speech",
    split="speech_6ch",
    streaming=True,
)
ds = ds.cast_column("audio", Audio())
# Read the samples from the TorchCodec decoder object:
rec = next(iter(ds))
samples = rec["audio"].get_all_samples()
speech_6ch = samples.data
sr = samples.sample_rate
print(f"6 channel speech has this shape: {speech_6ch.shape}, and a sampling rate of {sr} Hz.")
# We can access and compare individual channels from the 6ch device like this
speech0 = speech_6ch[0]  # mic 0
speech4 = speech_6ch[4]  # mic 4
t_axis = np.arange(speech0.shape[0]) / sr
plt.figure()
plt.plot(t_axis, speech0.numpy(), label="Microphone 0")
plt.plot(t_axis, speech4.numpy(), label="Microphone 4")
plt.xlabel("Time (s)")
plt.ylabel("Amplitude")
plt.legend()
plt.show()
```

## Dataset Details
The dataset contains three subsets:  
- **Treble10-Speech-mono**: This subset contains reverberant mono speech files, obtained by convolving dry speech signals with mono room impulse responses (RIRs). In each room, RIRs are available between 5 sound sources and several receivers. The receivers are placed along horizontal receiver grids with 0.5 m resolution at three heights (0.5 m, 1.0 m, 1.5 m). The validity of all source and receiver positions is checked to ensure that none of them intersects with the room geometry or furniture. 
- **Treble10-Speech-hoa8**: This subset contains reverberant speech files encoded in 8th-order Ambisonics. These reverberant speech files are obtained by convolving dry speech signals with 8th-order Ambisonics RIRs. The sound sources and receivers are identical to the Speech-mono subset. 
- **Treble10-Speech-6ch**: For this subset, a 6-channel cylindrical device is placed at the receiver positions from the Speech-mono subset. RIRs are then acquired between the 5 sound sources from above and each of the 6 device microphones. In other words, there is a 6-channel DeviceRIR for each source-receiver combination of the Speech-mono subset. Each channel of the DeviceRIR is then convolved with the same dry speech signal, resulting in a 6-channel reverberant speech signal. This 6-channel reverberant speech signal resembles the recordings you would obtain when placing that 6-channel device at the corresponding receiver position and recording speech played back at the source position.

All RIRs (mono/HOA/device) that were used to generate reverberant speech for this dataset were simulated with the Treble SDK. We use a hybrid simulation paradigm that combines a numerical wave-based solver (discontinuous Galerkin finite element method, DG-FEM) at low to midrange frequencies with geometrical acoustics (GA) simulations at high frequencies. For this dataset, the transition frequency between the wave-based and the GA simulation is set at 5 kHz. The resulting hybrid RIRs are broadband signals with a 32 kHz sampling rate, thus covering the entire frequency range of the signal and containing audio content up to 16 kHz.  

All dry speech files that were used to generate reverberant speech files through convolution with the above RIRs were taken from the _test_ splits of the [LibriSpeech corpus](https://www.openslr.org/12).
As the dry speech files were sampled at 16 kHz, the RIRs were downsampled while generating the Treble10-Speech set. You can create your own 32kHz speech samples by downloading the [Treble10-RIR](https://huggingface.co/datasets/treble-technologies/Treble10-RIR) dataset and convolving them with audio signals of your choice. 
## Uses

Use cases such as far-field automatic speech recognition (ASR), speech enhancement, dereverberation, and source separation benefit greatly from the **Treble10-Speech** dataset. To illustrate this, consider the contrast between near-field and far-field ASR. In near-field setups, such as smartphones or headsets, the microphone is close to the speaker, capturing a clean signal dominated by the direct sound. In far-field scenarios, as in smart speakers or conference-room devices, the microphone is several meters away, and the recorded signal becomes a complex blend of direct sound, reverberation, and background noise. This difference is not merely spatial but physical: in far-field conditions, sound waves reflect off walls, diffract around objects, and decay over time, all of which are captured by the RIR. To achieve robust performance in such environments, ASR and related models must be trained on datasets that accurately represent these intricate acoustic interactions—precisely what **Treble10-Speech** provides. Similarly, the performance of such systems can only be reliably determined when evaluating them on data that is accurate enough to model sound propagation in complex environments. 

## Dataset Structure

Each subset of **Treble10-Speech** corresponds to a different channel configuration of the simulated room impulse responses (RIRs). 
All subsets share the same metadata schema and organization.

|Split         | Description         | Channels |
|--------------|---------------------|----------|
|`speech_mono`    | Single-channel reverberant mono speech | 1 |
|`speech_hoa8`    | Reverberant speech encoded as 8th-order Ambisonics (ACN/SN3D format) | 81 |
|`speech_6ch`     | Reverberant speech at the microphones of a six-channel home audio device | 6 |


The six-channel device has microphones positioned at the following locatiosn relative to the center of the device:

|Channel|Position [m]|
|-------|--------|
|0      |[0.03, 0., 0.]|
|1      |[0.015.. 0.026., 0.]|
|2      |[-0.0145, 0.026, 0.]|
|3      |[-0.03, 0., 0.]|
|4      |[-0.015, -0.026, 0.]|
|5      |[0.015, -0.026, 0.]|



### File Contents

Each `.parquet` file contains the metadata for one subset (split) of the dataset.  
As this set of reverberant speech signals may be used for a variety of potential audio machine-learning tasks, we leave the actual segmentation of the data to the users. 
The metadata links each reverberant speech file to its corresponding dry speech file and includes detailed acoustic parameters.

| Column | Description |
|---------|-------------|
| **audio** | The convolved speech file. |
| **audio_filename** | Filename and relative path of the RIR WAV file. |
| **room** | Short room nickname (e.g., `Room1`, `Room5`). |
| **room_description** | Descriptive room type (e.g., `meeting_room`, `living_room`). |
| **room_volume** |  Volume of the room in cubic meters. |
| **source** | Label of the source. |
| **source_position** | 3D coordinates of the source in meters.  |
| **receiver** | Label of the receiver. |
| **receiver_position** | 3D coordinates of the receiver in meters.  |
| **direct_path_length** | Distance between source and receiver in meters.  |
| **rir_format** | Format of the RIR used (`mono`, `6ch`, or `hoa8`) |
| **Frequencies, EDT, T30, C50, Average Absorption** | Octave-band acoustic parameters. |
| **librispeech_split** | Source split of the dry speech signal. |
| **librispeech_file** | The file path and name of the dry signal as local to the Librispeech dataset. |
| **transcript** | The transcript of the utterance. |

## Acoustic Parameters
The RIRs that were used to generate the reverberant speech signals are presented with a few relevant acoustical parameters describing the acoustical field as sampled with the specific source/receiver pairs. 
### T30: Reverberation Time
T30 is a measure of how long a sound takes to fade away in a room after the sound source stops emitting noise.
It is a key measure of how reverberant a space is. 
Specifically, it's the time needed for the sound energy to drop by 60 decibels, estimated from the first 30 dBs of the decay.'
A short T30 correlates to a "dry" sounding room, like a small office or recording booth (ideally, under 0.2s). A long T30 correlates to a room that sounds "wet", such as a concert hall or parking garage (1.0s or more). 
### EDT: Early Decay Time
Early Decay Time is another measure of reverberation, but is calculated from the first 10 dB of energy decay. 
EDT is highly correlated with the psychoacoustic perception of reverberation, and can also provide information about the uniformity of the acoustic field within a space. 
If EDT is approximately equal to T30, the reverberation is approximately a single-slope decay. 
IF EDT is much shorter than T30, this indicates the existence of a double-slope energy decay, which may form when two rooms are acoustically coupled. 
### C50: Clarity Index (Speech)
C50 is an energy ratio between the early arriving sound (the first 50 milliseconds) to the late arrinng sound (from 50 milliseconds to the end of the RIR). 
C50 is typically used as a measure of the potential speech intelligibility and clarity of a room, as it quantifies how much the early sound is obscured by the room's reverberation. '
High C50 values (above 0dB) are typically considered to be ideal for clear and intelligible speech. 
Low C50 values (below 0dB) are typically considered to be difficult for speech clarity. 
## More Information
More information on the dataset can be found on the corresponding blog post.
## Licensing Information
The **Treble10-Speech** dataset combines two components with different licenses:
Speech recordings (dry signals) — sourced from the LibriSpeech corpus, licensed under the [Creative Commons Attribution 4.0 International License (CC-BY-4.0)](https://creativecommons.org/licenses/by/4.0/deed.en).
Acoustic impulse responses (RIRs) and acoustical metadata — originating from the Treble10-RIR dataset, licensed under the [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International license](https://creativecommons.org/licenses/by-nc-sa/4.0/).
The convolved (“wet”) speech recordings in this dataset are derivative works that combine both sources. As a result, they are governed by the CC-BY-4.0 license. The room impulse responses and all acoustical metadata associated with them remain governed by the (CC-BY-NC-SA-4.0).
## Citation Information

```
@misc{mullins2025treble10highqualitydatasetfarfield,
      title={Treble10: A high-quality dataset for far-field speech recognition, dereverberation, and enhancement}, 
      author={Sarabeth S. Mullins and Georg G\"otz and Eric Bezzam and Steven Zheng and Daniel Gert Nielsen},
      year={2025},
      eprint={2510.23141},
      archivePrefix={arXiv},
      primaryClass={eess.AS},
      url={https://arxiv.org/abs/2510.23141}, 
}
```