PhoenixStormJr commited on
Commit
0a99765
·
verified ·
1 Parent(s): 34f40f7

Update Changelog_CN.md

Browse files
Files changed (1) hide show
  1. Changelog_CN.md +67 -67
Changelog_CN.md CHANGED
@@ -1,80 +1,80 @@
1
- ### 20230618更新
2
- - v2增加32k48k两个新预训练模型
3
- - 修复非f0模型推理报错
4
- - 对于超过一小时的训练集的索引建立环节,自动kmeans缩小特征处理以加速索引训练、加入和查询
5
- - 附送一个人声转吉他玩具仓库
6
- - 数据处理剔除异常值切片
7
- - onnx导出选项卡
8
 
9
- 失败的实验:
10
- - ~~特征检索增加时序维度:寄,没啥效果~~
11
- - ~~特征检索增加PCAR降维可选项:寄,数据大用kmeans缩小数据量,数据小降维操作耗时比省下的匹配耗时还多~~
12
- - ~~支持onnx推理(附带仅推理的小压缩包):寄,生成nsf还是需要pytorch~~
13
- - ~~训练时在音高、gendereq、噪声等方面对输入进行随机增强:寄,没啥效果~~
14
 
15
- todolist
16
- - 接入小型声码器调研
17
- - 训练集音高识别支持crepe
18
- - crepe的精度支持和RVC-config同步
19
- - 对接F0编辑器
20
 
21
 
22
- ### 20230528更新
23
- - 增加v2jupyter notebook,韩文changelog,增加一些环境依赖
24
- - 增加呼吸、清辅音、齿音保护模式
25
- - 支持crepe-full推理
26
- - UVR5人声伴奏分离加上3个去延迟模型和MDX-Net去混响模型,增加HP3人声提取模型
27
- - 索引名称增加版本和实验名称
28
- - 人声伴奏分离、推理批量导出增加音频导出格式选项
29
- - 废弃32k模型的训练
30
 
31
- ### 20230513更新
32
- - 清除一键包内部老版本runtime内残留的infer_packuvr5_pack
33
- - 修复训练集预处理伪多进程的bug
34
- - 增加harvest识别音高可选通过中值滤波削弱哑音现象,可调整中值滤波半径
35
- - 导出音频增加后处理重采样
36
- - 训练n_cpu进程数从"仅调整f0提取"改为"调整数据预处理和f0提取"
37
- - 自动检测logs文件夹下的index路径,提供下拉列表功能
38
- - tab页增加"常见问题解答"(也可参考github-rvc-wiki
39
- - 相同路径的输入音频推理增加了音高缓存(用途:使用harvest音高提取,整个pipeline会经历漫长且重复的音高提取过程,如果不使用缓存,实验不同音色、索引、音高中值滤波半径参数的用户在第一次测试后的等待结果会非常痛苦)
40
 
41
- ### 20230514更新
42
- - 音量包络对齐输入混合(可以缓解“输入静音输出小幅度噪声”的问题。如果输入音频背景底噪大则不建议开启,默认不开启(值为1可视为不开启))
43
- - 支持按照指定频率保存提取的小模型(假如你想尝试不同epoch下的推理效果,但是不想保存所有大checkpoint并且每次都要ckpt手工处理提取小模型,这项功能会非常实用)
44
- - 通过设置环境变量解决服务端开了系统全局代理导致浏览器连接错误的问题
45
- - 支持v2预训练模型(目前只公开了40k版本进行测试,另外2个采样率还没有训练完全)
46
- - 推理前限制超过1的过大音量
47
- - 微调数据预处理参数
48
 
49
 
50
- ### 20230409更新
51
- - 修正训练参数,提升显卡平均利用率,A100最高从25%提升至90%左右,V100:50%->90%左右,2060S:60%->85%左右,P40:25%->95%左右,训练速度显著提升
52
- - 修正参数:总batch_size改为每张卡的batch_size
53
- - 修正total_epoch:最大限制100解锁至1000;默认10提升至默认20
54
- - 修复ckpt提取识别是否带音高错误导致推理异常的问题
55
- - 修复分布式训练每个rank都保存一次ckpt的问题
56
- - 特征提取进行nan特征过滤
57
- - 修复静音输入输出随机辅音or噪声的问题(老版模型需要重做训练集重训)
58
 
59
- ### 20230416更新
60
- - 新增本地实时变声迷你GUI,双击go-realtime-gui.bat启动
61
- - 训练推理均对<50Hz的频段进行滤波过滤
62
- - 训练推理音高提取pyworld最低音高从默认80下降至50,50-80hz间的男声低音不会哑
63
- - WebUI支持根据系统区域变更语言(现支持en_USja_JPzh_CNzh_HKzh_SGzh_TW,不支持的默认en_US
64
- - 修正部分显卡识别(例如V100-16G识别失败,P4识别失败)
65
 
66
- ### 20230428更新
67
- - 升级faiss索引设置,速度更快,质量更高
68
- - 取消total_npy依赖,后续分享模型不再需要填写total_npy
69
- - 解锁16系限制。4G显存GPU给到4G的推理设置。
70
- - 修复部分音频格式下UVR5人声伴奏分离的bug
71
- - 实时变声迷你gui增加对非40k与不懈怠音高模型的支持
72
 
73
- ### 后续计划:
74
- 功能:
75
- - 支持多人训练选项卡(至多4人)
76
 
77
- 底模:
78
- - 收集呼吸wav加入训练集修正呼吸变声电音的问题
79
- - 我们正在训练增加了歌声训练集的底模,未来会公开
80
 
 
1
+ ### 20230618 Update
2
+ - v2 adds two new pre-trained models, 32k and 48k
3
+ - Fix non-f0 model inference error
4
+ - For the indexing phase of training sets longer than one hour, automatically kmeans reduces feature processing to speed up index training, joining and querying
5
+ - Comes with a voice-to-guitar toy warehouse
6
+ - Data processing removes outlier slices
7
+ - onnx export tab
8
 
9
+ Failed experiments:
10
+ - ~~Feature retrieval adds time series dimension: sent, no effect~~
11
+ - ~~Feature retrieval adds PCAR dimension reduction option: sent, use kmeans to reduce the data volume for large data, and the dimension reduction operation takes more time than the saved matching time for small data~~
12
+ - ~~Support onnx reasoning (with a small compressed package for reasoning only): sent, pytorch is still needed to generate nsf~~
13
+ - ~~Randomly enhance the input in terms of pitch, gender, eq, noise, etc. during training: sent, no effect~~
14
 
15
+ todolist:
16
+ - Connect to small vocoder research
17
+ - Training set pitch recognition supports crepe
18
+ - Crepe's accuracy supports synchronization with RVC-config
19
+ - Connect to F0 editor
20
 
21
 
22
+ ### 20230528 Update
23
+ - Added v2 jupyter notebook, Korean changelog, and some environment dependencies
24
+ - Added breathing, clear consonant, and sibilant protection modes
25
+ - Support crepe-full reasoning
26
+ - UVR5 vocal accompaniment separation plus 3 de-delay models and MDX-Net dereverberation model, and added HP3 vocal extraction model
27
+ - Added version and experiment name to the index name
28
+ - Added audio export format options for vocal accompaniment separation and reasoning batch export
29
+ - Abandoned the training of 32k model
30
 
31
+ ### 20230513 Update
32
+ - Clear the remaining infer_pack and uvr5_pack in the old version of runtime in the one-click package
33
+ - Fixed the bug of pseudo multi-process in training set preprocessing
34
+ - Added harvest recognition pitch option to reduce mute phenomenon through median filtering, and adjustable median filter radius
35
+ - Export audio with post-processing resampling
36
+ - Changed the number of training n_cpu processes from "Adjust f0 extraction only" to "Adjust data preprocessing and f0 extraction"
37
+ - Automatically detect the index path under the logs folder and provide a drop-down list function
38
+ - Add "FAQ" to the tab page (also refer to github-rvc-wiki)
39
+ - Added pitch cache for input audio inference on the same path (Purpose: When using harvest pitch extraction, the entire pipeline will go through a long and repeated pitch extraction process. If cache is not used, users who experiment with different timbres, indices, and pitch median filter radius parameters will have a very painful wait for the results after the first test)
40
 
41
+ ### 20230514 Update
42
+ - Volume envelope alignment input mixing (can alleviate the problem of "input silence output small noise". If the input audio background noise is large, it is not recommended to turn it on. It is not turned on by default (the value is 1, which can be regarded as not turned on))
43
+ - Support saving the extracted small model at a specified frequency (if you want to try the inference effect under different epochs, but don't want to save all the large checkpoints and manually process the extracted small model every time, this function will be very useful)
44
+ - Solve the problem of browser connection errors caused by the server opening the system global proxy by setting environment variables
45
+ - Support v2 pre-trained models (currently only the 40k version is open for testing, and the other 2 sampling rates have not been fully trained)
46
+ - Limit the excessive volume of more than 1 before inference
47
+ - Fine-tune data preprocessing parameters
48
 
49
 
50
+ ### 20230409 Update
51
+ - Corrected training parameters, improved average graphics card utilization, A100 increased from 25% to about 90%, V100: 50%->90%, 2060S: 60%->85%, P40: 25%->95%, training speed significantly improved
52
+ - Corrected parameters: total batch_size changed to batch_size of each card
53
+ - Corrected total_epoch: maximum limit 100 unlocked to 1000; default 10 increased to default 20
54
+ - Fixed the problem of ckpt extraction recognition with pitch error causing inference abnormality
55
+ - Fixed the problem of ckpt being saved once for each rank in distributed training
56
+ - Feature extraction performs nan feature filtering
57
+ - Fixed the problem of silent input outputting random consonants or noise (old version model needs to redo training set retraining)
58
 
59
+ ### 20230416 Update
60
+ - Added local real-time voice change mini GUI, double-click go-realtime-gui.bat to start
61
+ - Training and inference all filter the frequency band <50Hz
62
+ - Training and inference pitch extraction pyworld's lowest pitch is reduced from the default 80 to 50, and the male bass between 50-80hz will not be muffled
63
+ - WebUI supports changing the language according to the system region (currently supports en_US, ja_JP, zh_CN, zh_HK, zh_SG, zh_TW, and the default en_US is not supported)
64
+ - Corrected some graphics card recognition (for example, V100-16G recognition failed, P4 recognition failed)
65
 
66
+ ### 20230428 Update
67
+ - Upgrade faiss index settings, faster and higher quality
68
+ - Cancel total_npy dependency, no need to fill in total_npy for subsequent model sharing
69
+ - Unlock 16 series restrictions. 4G memory GPU gives 4G inference settings.
70
+ - Fix the bug of UVR5 vocal accompaniment separation under some audio formats
71
+ - Real-time voice change mini gui adds support for non-40k and non-slack pitch models
72
 
73
+ ### Follow-up plan:
74
+ Function:
75
+ - Support multi-player training tab (up to 4 people)
76
 
77
+ Base model:
78
+ - Collect breathing wav and add it to the training set to fix the problem of breathing voice changing electronic music
79
+ - We are training the base model of the singing training set, which will be made public in the future
80