2907
Browse files- micro/config.json +2 -2
- micro/diffusion_pytorch_model.safetensors +2 -2
- samples/micro_192x384_0.jpg +2 -2
- samples/micro_256x384_0.jpg +2 -2
- samples/micro_320x384_0.jpg +2 -2
- samples/micro_384x192_0.jpg +2 -2
- samples/micro_384x256_0.jpg +2 -2
- samples/micro_384x320_0.jpg +2 -2
- samples/micro_384x384_0.jpg +2 -2
- samples/sdxl_192x384_0.jpg +3 -0
- samples/sdxl_256x384_0.jpg +3 -0
- samples/sdxl_320x384_0.jpg +3 -0
- samples/sdxl_384x192_0.jpg +3 -0
- samples/sdxl_384x256_0.jpg +3 -0
- samples/sdxl_384x320_0.jpg +3 -0
- src/micro_create.ipynb +2 -2
- train.py +6 -6
micro/config.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18891ea8f81c705422b12d7575493f5abb949d60f2814a989a436fde65e84c82
|
3 |
+
size 1873
|
micro/diffusion_pytorch_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9fa21e51b39a8f8f44036b0e271cddc65914ea3591cce6e0a35506425026f183
|
3 |
+
size 3929714960
|
samples/micro_192x384_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_256x384_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_320x384_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_384x192_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_384x256_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_384x320_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/micro_384x384_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxl_192x384_0.jpg
ADDED
![]() |
Git LFS Details
|
samples/sdxl_256x384_0.jpg
ADDED
![]() |
Git LFS Details
|
samples/sdxl_320x384_0.jpg
ADDED
![]() |
Git LFS Details
|
samples/sdxl_384x192_0.jpg
ADDED
![]() |
Git LFS Details
|
samples/sdxl_384x256_0.jpg
ADDED
![]() |
Git LFS Details
|
samples/sdxl_384x320_0.jpg
ADDED
![]() |
Git LFS Details
|
src/micro_create.ipynb
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:856a34f139aad0b40359836b62caa7a005ee898990e416a7a04a9ded0881496a
|
3 |
+
size 33982
|
train.py
CHANGED
@@ -27,12 +27,12 @@ import torch.nn.functional as F
|
|
27 |
# --------------------------- Параметры ---------------------------
|
28 |
ds_path = "datasets/384"
|
29 |
project = "micro"
|
30 |
-
batch_size =
|
31 |
base_learning_rate = 1e-4
|
32 |
min_learning_rate = 5e-5
|
33 |
-
num_epochs =
|
34 |
# samples/save per epoch
|
35 |
-
sample_interval_share =
|
36 |
use_wandb = True
|
37 |
save_model = True
|
38 |
use_decay = True
|
@@ -54,12 +54,12 @@ dispersive_temperature=0.5
|
|
54 |
dispersive_weight= 0.05
|
55 |
percentile_clipping = 95 # 8bit optim
|
56 |
betta2 = 0.97
|
57 |
-
eps = 1e-
|
58 |
clip_grad_norm = 1.0
|
59 |
steps_offset = 0 # Scheduler
|
60 |
limit = 0
|
61 |
checkpoints_folder = ""
|
62 |
-
mixed_precision = "
|
63 |
gradient_accumulation_steps = 1
|
64 |
accelerator = Accelerator(
|
65 |
mixed_precision=mixed_precision,
|
@@ -807,7 +807,7 @@ for epoch in range(start_epoch, start_epoch + num_epochs):
|
|
807 |
avg_loss = np.mean(batch_losses[-sample_interval:])
|
808 |
avg_tloss = np.mean(batch_tlosses[-sample_interval:])
|
809 |
avg_grad = torch.mean(torch.stack(batch_grads[-sample_interval:])).cpu().item()
|
810 |
-
print(f"Эпоха {epoch}, шаг {global_step}, средний лосс: {avg_loss:.6f}")
|
811 |
|
812 |
if save_model:
|
813 |
print("saving:",avg_loss < min_loss*save_barrier)
|
|
|
27 |
# --------------------------- Параметры ---------------------------
|
28 |
ds_path = "datasets/384"
|
29 |
project = "micro"
|
30 |
+
batch_size = 64
|
31 |
base_learning_rate = 1e-4
|
32 |
min_learning_rate = 5e-5
|
33 |
+
num_epochs = 50
|
34 |
# samples/save per epoch
|
35 |
+
sample_interval_share = 10
|
36 |
use_wandb = True
|
37 |
save_model = True
|
38 |
use_decay = True
|
|
|
54 |
dispersive_weight= 0.05
|
55 |
percentile_clipping = 95 # 8bit optim
|
56 |
betta2 = 0.97
|
57 |
+
eps = 1e-6
|
58 |
clip_grad_norm = 1.0
|
59 |
steps_offset = 0 # Scheduler
|
60 |
limit = 0
|
61 |
checkpoints_folder = ""
|
62 |
+
mixed_precision = "no" #"fp16"
|
63 |
gradient_accumulation_steps = 1
|
64 |
accelerator = Accelerator(
|
65 |
mixed_precision=mixed_precision,
|
|
|
807 |
avg_loss = np.mean(batch_losses[-sample_interval:])
|
808 |
avg_tloss = np.mean(batch_tlosses[-sample_interval:])
|
809 |
avg_grad = torch.mean(torch.stack(batch_grads[-sample_interval:])).cpu().item()
|
810 |
+
print(f"Эпоха {epoch}, шаг {global_step}, средний лосс: {avg_loss:.6f}, grad: {avg_grad:.6f}")
|
811 |
|
812 |
if save_model:
|
813 |
print("saving:",avg_loss < min_loss*save_barrier)
|