提交 3a608ead 编写于 作者: A a2569875

fix weights

上级 73de4b8f
......@@ -449,7 +449,12 @@ def lora_Linear_forward(self, input):
torch.nn.Linear_forward_before_lora = backup_Linear_forward
return result
return lycoris.lyco_Linear_forward(self, input)
clear_cache_lora(self, False)
if lora_ext.is_sd_1_5:
import networks
networks.network_restore_weights_from_backup(self)
networks.network_reset_cached_weight(self)
else:
clear_cache_lora(self, False)
if (not self.weight.is_cuda) and input.is_cuda: #if variables not on the same device (between cpu and gpu)
self_weight_cuda = self.weight.to(device=devices.device) #pass to GPU
to_del = self.weight
......@@ -488,7 +493,12 @@ def lora_Conv2d_forward(self, input):
return result
return lycoris.lyco_Conv2d_forward(self, input)
clear_cache_lora(self, False)
if lora_ext.is_sd_1_5:
import networks
networks.network_restore_weights_from_backup(self)
networks.network_reset_cached_weight(self)
else:
clear_cache_lora(self, False)
if (not self.weight.is_cuda) and input.is_cuda:
self_weight_cuda = self.weight.to(device=devices.device)
to_del = self.weight
......@@ -528,7 +538,12 @@ def lora_MultiheadAttention_forward(self, input):
return result
return lycoris.lyco_MultiheadAttention_forward(self, input)
clear_cache_lora(self, False)
if lora_ext.is_sd_1_5:
import networks
networks.network_restore_weights_from_backup(self)
networks.network_reset_cached_weight(self)
else:
clear_cache_lora(self, False)
if (not self.weight.is_cuda) and input.is_cuda:
self_weight_cuda = self.weight.to(device=devices.device)
to_del = self.weight
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册