Commit
·
ef350dd
1
Parent(s):
9a97d7f
extract zips, mel_24000_cpu.raw -> .bin, rm zips
Browse files- versatile_audio_sr_base_openvino_models.zip → audiosr_decoder.bin +2 -2
- audiosr_decoder.xml +0 -0
- versatile_audio_sr_ddpm_basic_openvino_models.zip → audiosr_encoder.bin +2 -2
- audiosr_encoder.xml +0 -0
- versatile_audio_sr_ddpm_speech_openvino_models.zip → basic/ddpm.bin +2 -2
- basic/ddpm.xml +0 -0
- mel_24000_cpu.bin +3 -0
- post_quant_conv.bin +3 -0
- post_quant_conv.xml +160 -0
- quant_conv.bin +3 -0
- quant_conv.xml +160 -0
- speech/ddpm.bin +3 -0
- speech/ddpm.xml +0 -0
- vae_feature_extract.bin +3 -0
- vae_feature_extract.xml +0 -0
- vocoder.bin +3 -0
- vocoder.xml +0 -0
versatile_audio_sr_base_openvino_models.zip → audiosr_decoder.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3f15dd621d5ffb191f1dbb5540b7c75301e8769ccbe062942b74a81e98ac2dd2
|
| 3 |
+
size 267047980
|
audiosr_decoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
versatile_audio_sr_ddpm_basic_openvino_models.zip → audiosr_encoder.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9eb70d3baff6a5859a8f05f6ab7b3a13d5d0b23540758c78ccf91b0ccd4f83d6
|
| 3 |
+
size 180020602
|
audiosr_encoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
versatile_audio_sr_ddpm_speech_openvino_models.zip → basic/ddpm.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:03fa2b86f718e23b0c81e1ef481c1d75c87d3d6eb949cdfe06ca351b8f509d8d
|
| 3 |
+
size 516390728
|
basic/ddpm.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mel_24000_cpu.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8dd24b6b1a81dc8f70ed479186fe66b1ccb4fc2dcffc3133dfd01daa93618c1d
|
| 3 |
+
size 1049600
|
post_quant_conv.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe865e2ad452ac5de682a7d9afecd685da2fca328ff453d0575fd75196f809c0
|
| 3 |
+
size 544
|
post_quant_conv.xml
ADDED
|
@@ -0,0 +1,160 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<?xml version="1.0"?>
|
| 2 |
+
<net name="torch_jit" version="11">
|
| 3 |
+
<layers>
|
| 4 |
+
<layer id="0" name="in" type="Parameter" version="opset1">
|
| 5 |
+
<data shape="1,16,128,32" element_type="f32" />
|
| 6 |
+
<output>
|
| 7 |
+
<port id="0" precision="FP32" names="in">
|
| 8 |
+
<dim>1</dim>
|
| 9 |
+
<dim>16</dim>
|
| 10 |
+
<dim>128</dim>
|
| 11 |
+
<dim>32</dim>
|
| 12 |
+
</port>
|
| 13 |
+
</output>
|
| 14 |
+
</layer>
|
| 15 |
+
<layer id="1" name="weight_compressed" type="Const" version="opset1">
|
| 16 |
+
<data element_type="f16" shape="16, 16, 1, 1" offset="0" size="512" />
|
| 17 |
+
<output>
|
| 18 |
+
<port id="0" precision="FP16">
|
| 19 |
+
<dim>16</dim>
|
| 20 |
+
<dim>16</dim>
|
| 21 |
+
<dim>1</dim>
|
| 22 |
+
<dim>1</dim>
|
| 23 |
+
</port>
|
| 24 |
+
</output>
|
| 25 |
+
</layer>
|
| 26 |
+
<layer id="2" name="weight" type="Convert" version="opset1">
|
| 27 |
+
<data destination_type="f32" />
|
| 28 |
+
<rt_info>
|
| 29 |
+
<attribute name="decompression" version="0" />
|
| 30 |
+
</rt_info>
|
| 31 |
+
<input>
|
| 32 |
+
<port id="0" precision="FP16">
|
| 33 |
+
<dim>16</dim>
|
| 34 |
+
<dim>16</dim>
|
| 35 |
+
<dim>1</dim>
|
| 36 |
+
<dim>1</dim>
|
| 37 |
+
</port>
|
| 38 |
+
</input>
|
| 39 |
+
<output>
|
| 40 |
+
<port id="1" precision="FP32" names="weight">
|
| 41 |
+
<dim>16</dim>
|
| 42 |
+
<dim>16</dim>
|
| 43 |
+
<dim>1</dim>
|
| 44 |
+
<dim>1</dim>
|
| 45 |
+
</port>
|
| 46 |
+
</output>
|
| 47 |
+
</layer>
|
| 48 |
+
<layer id="3" name="/Conv/WithoutBiases" type="Convolution" version="opset1">
|
| 49 |
+
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
| 50 |
+
<input>
|
| 51 |
+
<port id="0" precision="FP32">
|
| 52 |
+
<dim>1</dim>
|
| 53 |
+
<dim>16</dim>
|
| 54 |
+
<dim>128</dim>
|
| 55 |
+
<dim>32</dim>
|
| 56 |
+
</port>
|
| 57 |
+
<port id="1" precision="FP32">
|
| 58 |
+
<dim>16</dim>
|
| 59 |
+
<dim>16</dim>
|
| 60 |
+
<dim>1</dim>
|
| 61 |
+
<dim>1</dim>
|
| 62 |
+
</port>
|
| 63 |
+
</input>
|
| 64 |
+
<output>
|
| 65 |
+
<port id="2" precision="FP32">
|
| 66 |
+
<dim>1</dim>
|
| 67 |
+
<dim>16</dim>
|
| 68 |
+
<dim>128</dim>
|
| 69 |
+
<dim>32</dim>
|
| 70 |
+
</port>
|
| 71 |
+
</output>
|
| 72 |
+
</layer>
|
| 73 |
+
<layer id="4" name="Reshape_16_compressed" type="Const" version="opset1">
|
| 74 |
+
<data element_type="f16" shape="1, 16, 1, 1" offset="512" size="32" />
|
| 75 |
+
<output>
|
| 76 |
+
<port id="0" precision="FP16">
|
| 77 |
+
<dim>1</dim>
|
| 78 |
+
<dim>16</dim>
|
| 79 |
+
<dim>1</dim>
|
| 80 |
+
<dim>1</dim>
|
| 81 |
+
</port>
|
| 82 |
+
</output>
|
| 83 |
+
</layer>
|
| 84 |
+
<layer id="5" name="Reshape_16" type="Convert" version="opset1">
|
| 85 |
+
<data destination_type="f32" />
|
| 86 |
+
<rt_info>
|
| 87 |
+
<attribute name="decompression" version="0" />
|
| 88 |
+
</rt_info>
|
| 89 |
+
<input>
|
| 90 |
+
<port id="0" precision="FP16">
|
| 91 |
+
<dim>1</dim>
|
| 92 |
+
<dim>16</dim>
|
| 93 |
+
<dim>1</dim>
|
| 94 |
+
<dim>1</dim>
|
| 95 |
+
</port>
|
| 96 |
+
</input>
|
| 97 |
+
<output>
|
| 98 |
+
<port id="1" precision="FP32">
|
| 99 |
+
<dim>1</dim>
|
| 100 |
+
<dim>16</dim>
|
| 101 |
+
<dim>1</dim>
|
| 102 |
+
<dim>1</dim>
|
| 103 |
+
</port>
|
| 104 |
+
</output>
|
| 105 |
+
</layer>
|
| 106 |
+
<layer id="6" name="out" type="Add" version="opset1">
|
| 107 |
+
<data auto_broadcast="numpy" />
|
| 108 |
+
<input>
|
| 109 |
+
<port id="0" precision="FP32">
|
| 110 |
+
<dim>1</dim>
|
| 111 |
+
<dim>16</dim>
|
| 112 |
+
<dim>128</dim>
|
| 113 |
+
<dim>32</dim>
|
| 114 |
+
</port>
|
| 115 |
+
<port id="1" precision="FP32">
|
| 116 |
+
<dim>1</dim>
|
| 117 |
+
<dim>16</dim>
|
| 118 |
+
<dim>1</dim>
|
| 119 |
+
<dim>1</dim>
|
| 120 |
+
</port>
|
| 121 |
+
</input>
|
| 122 |
+
<output>
|
| 123 |
+
<port id="2" precision="FP32" names="out">
|
| 124 |
+
<dim>1</dim>
|
| 125 |
+
<dim>16</dim>
|
| 126 |
+
<dim>128</dim>
|
| 127 |
+
<dim>32</dim>
|
| 128 |
+
</port>
|
| 129 |
+
</output>
|
| 130 |
+
</layer>
|
| 131 |
+
<layer id="7" name="out/sink_port_0" type="Result" version="opset1">
|
| 132 |
+
<input>
|
| 133 |
+
<port id="0" precision="FP32">
|
| 134 |
+
<dim>1</dim>
|
| 135 |
+
<dim>16</dim>
|
| 136 |
+
<dim>128</dim>
|
| 137 |
+
<dim>32</dim>
|
| 138 |
+
</port>
|
| 139 |
+
</input>
|
| 140 |
+
</layer>
|
| 141 |
+
</layers>
|
| 142 |
+
<edges>
|
| 143 |
+
<edge from-layer="0" from-port="0" to-layer="3" to-port="0" />
|
| 144 |
+
<edge from-layer="1" from-port="0" to-layer="2" to-port="0" />
|
| 145 |
+
<edge from-layer="2" from-port="1" to-layer="3" to-port="1" />
|
| 146 |
+
<edge from-layer="3" from-port="2" to-layer="6" to-port="0" />
|
| 147 |
+
<edge from-layer="4" from-port="0" to-layer="5" to-port="0" />
|
| 148 |
+
<edge from-layer="5" from-port="1" to-layer="6" to-port="1" />
|
| 149 |
+
<edge from-layer="6" from-port="2" to-layer="7" to-port="0" />
|
| 150 |
+
</edges>
|
| 151 |
+
<rt_info>
|
| 152 |
+
<MO_version value="2024.4.0-16579-c3152d32c9c-releases/2024/4" />
|
| 153 |
+
<Runtime_version value="2024.4.0-16579-c3152d32c9c-releases/2024/4" />
|
| 154 |
+
<conversion_parameters>
|
| 155 |
+
<input_model value="DIR\post_quant_conv.onnx" />
|
| 156 |
+
<is_python_api_used value="False" />
|
| 157 |
+
</conversion_parameters>
|
| 158 |
+
<legacy_frontend value="False" />
|
| 159 |
+
</rt_info>
|
| 160 |
+
</net>
|
quant_conv.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a10eb725d47b58dd4edf72284681c78edcd56c83386e2665d3f7dcdf22ff96f
|
| 3 |
+
size 2112
|
quant_conv.xml
ADDED
|
@@ -0,0 +1,160 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<?xml version="1.0"?>
|
| 2 |
+
<net name="torch_jit" version="11">
|
| 3 |
+
<layers>
|
| 4 |
+
<layer id="0" name="h" type="Parameter" version="opset1">
|
| 5 |
+
<data shape="1,32,128,32" element_type="f32" />
|
| 6 |
+
<output>
|
| 7 |
+
<port id="0" precision="FP32" names="h">
|
| 8 |
+
<dim>1</dim>
|
| 9 |
+
<dim>32</dim>
|
| 10 |
+
<dim>128</dim>
|
| 11 |
+
<dim>32</dim>
|
| 12 |
+
</port>
|
| 13 |
+
</output>
|
| 14 |
+
</layer>
|
| 15 |
+
<layer id="1" name="weight_compressed" type="Const" version="opset1">
|
| 16 |
+
<data element_type="f16" shape="32, 32, 1, 1" offset="0" size="2048" />
|
| 17 |
+
<output>
|
| 18 |
+
<port id="0" precision="FP16">
|
| 19 |
+
<dim>32</dim>
|
| 20 |
+
<dim>32</dim>
|
| 21 |
+
<dim>1</dim>
|
| 22 |
+
<dim>1</dim>
|
| 23 |
+
</port>
|
| 24 |
+
</output>
|
| 25 |
+
</layer>
|
| 26 |
+
<layer id="2" name="weight" type="Convert" version="opset1">
|
| 27 |
+
<data destination_type="f32" />
|
| 28 |
+
<rt_info>
|
| 29 |
+
<attribute name="decompression" version="0" />
|
| 30 |
+
</rt_info>
|
| 31 |
+
<input>
|
| 32 |
+
<port id="0" precision="FP16">
|
| 33 |
+
<dim>32</dim>
|
| 34 |
+
<dim>32</dim>
|
| 35 |
+
<dim>1</dim>
|
| 36 |
+
<dim>1</dim>
|
| 37 |
+
</port>
|
| 38 |
+
</input>
|
| 39 |
+
<output>
|
| 40 |
+
<port id="1" precision="FP32" names="weight">
|
| 41 |
+
<dim>32</dim>
|
| 42 |
+
<dim>32</dim>
|
| 43 |
+
<dim>1</dim>
|
| 44 |
+
<dim>1</dim>
|
| 45 |
+
</port>
|
| 46 |
+
</output>
|
| 47 |
+
</layer>
|
| 48 |
+
<layer id="3" name="/Conv/WithoutBiases" type="Convolution" version="opset1">
|
| 49 |
+
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
| 50 |
+
<input>
|
| 51 |
+
<port id="0" precision="FP32">
|
| 52 |
+
<dim>1</dim>
|
| 53 |
+
<dim>32</dim>
|
| 54 |
+
<dim>128</dim>
|
| 55 |
+
<dim>32</dim>
|
| 56 |
+
</port>
|
| 57 |
+
<port id="1" precision="FP32">
|
| 58 |
+
<dim>32</dim>
|
| 59 |
+
<dim>32</dim>
|
| 60 |
+
<dim>1</dim>
|
| 61 |
+
<dim>1</dim>
|
| 62 |
+
</port>
|
| 63 |
+
</input>
|
| 64 |
+
<output>
|
| 65 |
+
<port id="2" precision="FP32">
|
| 66 |
+
<dim>1</dim>
|
| 67 |
+
<dim>32</dim>
|
| 68 |
+
<dim>128</dim>
|
| 69 |
+
<dim>32</dim>
|
| 70 |
+
</port>
|
| 71 |
+
</output>
|
| 72 |
+
</layer>
|
| 73 |
+
<layer id="4" name="Reshape_16_compressed" type="Const" version="opset1">
|
| 74 |
+
<data element_type="f16" shape="1, 32, 1, 1" offset="2048" size="64" />
|
| 75 |
+
<output>
|
| 76 |
+
<port id="0" precision="FP16">
|
| 77 |
+
<dim>1</dim>
|
| 78 |
+
<dim>32</dim>
|
| 79 |
+
<dim>1</dim>
|
| 80 |
+
<dim>1</dim>
|
| 81 |
+
</port>
|
| 82 |
+
</output>
|
| 83 |
+
</layer>
|
| 84 |
+
<layer id="5" name="Reshape_16" type="Convert" version="opset1">
|
| 85 |
+
<data destination_type="f32" />
|
| 86 |
+
<rt_info>
|
| 87 |
+
<attribute name="decompression" version="0" />
|
| 88 |
+
</rt_info>
|
| 89 |
+
<input>
|
| 90 |
+
<port id="0" precision="FP16">
|
| 91 |
+
<dim>1</dim>
|
| 92 |
+
<dim>32</dim>
|
| 93 |
+
<dim>1</dim>
|
| 94 |
+
<dim>1</dim>
|
| 95 |
+
</port>
|
| 96 |
+
</input>
|
| 97 |
+
<output>
|
| 98 |
+
<port id="1" precision="FP32">
|
| 99 |
+
<dim>1</dim>
|
| 100 |
+
<dim>32</dim>
|
| 101 |
+
<dim>1</dim>
|
| 102 |
+
<dim>1</dim>
|
| 103 |
+
</port>
|
| 104 |
+
</output>
|
| 105 |
+
</layer>
|
| 106 |
+
<layer id="6" name="moments" type="Add" version="opset1">
|
| 107 |
+
<data auto_broadcast="numpy" />
|
| 108 |
+
<input>
|
| 109 |
+
<port id="0" precision="FP32">
|
| 110 |
+
<dim>1</dim>
|
| 111 |
+
<dim>32</dim>
|
| 112 |
+
<dim>128</dim>
|
| 113 |
+
<dim>32</dim>
|
| 114 |
+
</port>
|
| 115 |
+
<port id="1" precision="FP32">
|
| 116 |
+
<dim>1</dim>
|
| 117 |
+
<dim>32</dim>
|
| 118 |
+
<dim>1</dim>
|
| 119 |
+
<dim>1</dim>
|
| 120 |
+
</port>
|
| 121 |
+
</input>
|
| 122 |
+
<output>
|
| 123 |
+
<port id="2" precision="FP32" names="moments">
|
| 124 |
+
<dim>1</dim>
|
| 125 |
+
<dim>32</dim>
|
| 126 |
+
<dim>128</dim>
|
| 127 |
+
<dim>32</dim>
|
| 128 |
+
</port>
|
| 129 |
+
</output>
|
| 130 |
+
</layer>
|
| 131 |
+
<layer id="7" name="moments/sink_port_0" type="Result" version="opset1">
|
| 132 |
+
<input>
|
| 133 |
+
<port id="0" precision="FP32">
|
| 134 |
+
<dim>1</dim>
|
| 135 |
+
<dim>32</dim>
|
| 136 |
+
<dim>128</dim>
|
| 137 |
+
<dim>32</dim>
|
| 138 |
+
</port>
|
| 139 |
+
</input>
|
| 140 |
+
</layer>
|
| 141 |
+
</layers>
|
| 142 |
+
<edges>
|
| 143 |
+
<edge from-layer="0" from-port="0" to-layer="3" to-port="0" />
|
| 144 |
+
<edge from-layer="1" from-port="0" to-layer="2" to-port="0" />
|
| 145 |
+
<edge from-layer="2" from-port="1" to-layer="3" to-port="1" />
|
| 146 |
+
<edge from-layer="3" from-port="2" to-layer="6" to-port="0" />
|
| 147 |
+
<edge from-layer="4" from-port="0" to-layer="5" to-port="0" />
|
| 148 |
+
<edge from-layer="5" from-port="1" to-layer="6" to-port="1" />
|
| 149 |
+
<edge from-layer="6" from-port="2" to-layer="7" to-port="0" />
|
| 150 |
+
</edges>
|
| 151 |
+
<rt_info>
|
| 152 |
+
<MO_version value="2024.4.0-16579-c3152d32c9c-releases/2024/4" />
|
| 153 |
+
<Runtime_version value="2024.4.0-16579-c3152d32c9c-releases/2024/4" />
|
| 154 |
+
<conversion_parameters>
|
| 155 |
+
<input_model value="DIR\quant_conv.onnx" />
|
| 156 |
+
<is_python_api_used value="False" />
|
| 157 |
+
</conversion_parameters>
|
| 158 |
+
<legacy_frontend value="False" />
|
| 159 |
+
</rt_info>
|
| 160 |
+
</net>
|
speech/ddpm.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3565cdfca685b7d6bbce89257dd62c54827bb5b5452ce5bf5439e271fa2d2c8d
|
| 3 |
+
size 516390728
|
speech/ddpm.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vae_feature_extract.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9cb961e646544a8de08cddef328aabe222e7e701794bb37193cd651f522a188f
|
| 3 |
+
size 180022744
|
vae_feature_extract.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vocoder.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:33a9c6375fe33bc652f8e22d4fc0b5317351f9acdaa7121bf92c9b1e594fdd8f
|
| 3 |
+
size 380563784
|
vocoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|