ep20
Browse files- samples/sdxs_448x768_0.jpg +2 -2
- samples/sdxs_512x768_0.jpg +2 -2
- samples/sdxs_576x768_0.jpg +2 -2
- samples/sdxs_640x768_0.jpg +2 -2
- samples/sdxs_704x768_0.jpg +2 -2
- samples/sdxs_768x448_0.jpg +2 -2
- samples/sdxs_768x512_0.jpg +2 -2
- samples/sdxs_768x576_0.jpg +2 -2
- samples/sdxs_768x640_0.jpg +2 -2
- samples/sdxs_768x704_0.jpg +2 -2
- samples/sdxs_768x768_0.jpg +2 -2
- sdxs/diffusion_pytorch_model.safetensors +1 -1
- train.py +3 -3
samples/sdxs_448x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_512x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_576x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_640x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_704x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x448_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x512_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x576_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x640_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x704_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
samples/sdxs_768x768_0.jpg
CHANGED
![]() |
Git LFS Details
|
![]() |
Git LFS Details
|
sdxs/diffusion_pytorch_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 7010934208
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac5b45c10bb18c53736919d2519722bd7ae65be46033ed7aca56a3c6fbc38816
|
3 |
size 7010934208
|
train.py
CHANGED
@@ -24,7 +24,7 @@ import bitsandbytes as bnb
|
|
24 |
# --------------------------- Параметры ---------------------------
|
25 |
save_path = "datasets/768" # "datasets/576" #"datasets/576p2" #"datasets/1152p2" #"datasets/576p2" #"datasets/dataset384_temp" #"datasets/dataset384" #"datasets/imagenet-1kk" #"datasets/siski576" #"datasets/siski384" #"datasets/siski64" #"datasets/mnist"
|
26 |
batch_size = 26 #26 #45 #11 #45 #555 #35 #7
|
27 |
-
base_learning_rate = 9e-6 #1e-5 #2.5e-5 #4e-6 #9.5e-7 #9e-7 #2e-6 #1e-6 #9e-7 #1e-6 #2e-6 #1e-6 #2e-6 #6e-6 #2e-6 #8e-7 #6e-6 #2e-5 #4e-5 #3e-5 #5e-5 #8e-5
|
28 |
num_epochs = 3 #2 #36 #18
|
29 |
project = "sdxs" #"sdxxxs"
|
30 |
use_wandb = True
|
@@ -504,8 +504,8 @@ for epoch in range(start_epoch, start_epoch + num_epochs):
|
|
504 |
target = scheduler.get_velocity(latents, noise, timesteps)
|
505 |
|
506 |
# Считаем лосс
|
507 |
-
|
508 |
-
loss = torch.nn.functional.mse_loss(noise_pred, target)
|
509 |
|
510 |
# Делаем backward через Accelerator
|
511 |
accelerator.backward(loss)
|
|
|
24 |
# --------------------------- Параметры ---------------------------
|
25 |
save_path = "datasets/768" # "datasets/576" #"datasets/576p2" #"datasets/1152p2" #"datasets/576p2" #"datasets/dataset384_temp" #"datasets/dataset384" #"datasets/imagenet-1kk" #"datasets/siski576" #"datasets/siski384" #"datasets/siski64" #"datasets/mnist"
|
26 |
batch_size = 26 #26 #45 #11 #45 #555 #35 #7
|
27 |
+
base_learning_rate = 8e-6 #9e-6 #1e-5 #2.5e-5 #4e-6 #9.5e-7 #9e-7 #2e-6 #1e-6 #9e-7 #1e-6 #2e-6 #1e-6 #2e-6 #6e-6 #2e-6 #8e-7 #6e-6 #2e-5 #4e-5 #3e-5 #5e-5 #8e-5
|
28 |
num_epochs = 3 #2 #36 #18
|
29 |
project = "sdxs" #"sdxxxs"
|
30 |
use_wandb = True
|
|
|
504 |
target = scheduler.get_velocity(latents, noise, timesteps)
|
505 |
|
506 |
# Считаем лосс
|
507 |
+
loss = torch.nn.functional.mse_loss(noise_pred.float(), target.float())
|
508 |
+
#loss = torch.nn.functional.mse_loss(noise_pred, target)
|
509 |
|
510 |
# Делаем backward через Accelerator
|
511 |
accelerator.backward(loss)
|