mirror of
https://github.com/AUTOMATIC1111/stable-diffusion-webui.git
synced 2026-01-07 01:32:49 -08:00
prevent accidental creation of CLIP models in float32 type when user wants float16
This commit is contained in:
parent
7ee2114cd9
commit
b443fdcf76
2 changed files with 4 additions and 3 deletions
|
|
@ -61,9 +61,9 @@ class SD3Cond(torch.nn.Module):
|
|||
self.tokenizer = SD3Tokenizer()
|
||||
|
||||
with torch.no_grad():
|
||||
self.clip_g = SDXLClipG(CLIPG_CONFIG, device="cpu", dtype=torch.float32)
|
||||
self.clip_l = SDClipModel(layer="hidden", layer_idx=-2, device="cpu", dtype=torch.float32, layer_norm_hidden_state=False, return_projected_pooled=False, textmodel_json_config=CLIPL_CONFIG)
|
||||
self.t5xxl = T5XXLModel(T5_CONFIG, device="cpu", dtype=torch.float32)
|
||||
self.clip_g = SDXLClipG(CLIPG_CONFIG, device="cpu", dtype=devices.dtype)
|
||||
self.clip_l = SDClipModel(layer="hidden", layer_idx=-2, device="cpu", dtype=devices.dtype, layer_norm_hidden_state=False, return_projected_pooled=False, textmodel_json_config=CLIPL_CONFIG)
|
||||
self.t5xxl = T5XXLModel(T5_CONFIG, device="cpu", dtype=devices.dtype)
|
||||
|
||||
self.weights_loaded = False
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue