File size: 1,270 Bytes
149bcec
 
 
 
 
 
 
 
fca4445
149bcec
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7a67169
 
 
 
149bcec
37dc70e
149bcec
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
---
language:
- ar
- br
- cn
- zh
- de
- es
- en
- fa
- fil
- fr
- gr
- el
- gu
- hi
- it
- ja
- kz
- uk
- ru
- nl
- pt
- pl
- ca
- cs
- eo
- ko
- sv
- te
- tg
- tr
- uz
- vn
pipeline_tag: automatic-speech-recognition
tags:
- vosk
- stt
- tts
- speaker_indentification
- "speaker indentification"
- punctuation
---
# Vosk Models
Downloaded from: [**URL**](https://alphacephei.com/vosk/models)


Models
We have two types of models - big and small, small models are ideal for some limited task on mobile applications. They can run on smartphones, Raspberry Pi’s. They are also recommended for desktop applications. Small model typically is around 50Mb in size and requires about 300Mb of memory in runtime. Big models are for the high-accuracy transcription on the server. Big models require up to 16Gb in memory since they apply advanced AI algorithms. Ideally you run them on some high-end servers like i7 or latest AMD Ryzen. On AWS you can take a look on c5a machines and similar machines in other clouds.
Most small model allow dynamic vocabulary reconfiguration. Big models are static the vocabulary can not be modified in runtime.

Support me:  [**Boosty**](https://boosty.to/dreammine)  or  [**Donationalerts**](https://www.donationalerts.com/r/derur_dreammine)