File "F:\FLUX 2\webui\backend\loader.py", line 61, in load_huggingface_component
assert isinstance(state_dict, dict) and len(state_dict) > 16, 'You do not have CLIP state dict!'
AssertionError: You do not have CLIP state dict!
You do not have CLIP state dict!
Loading Model: {'checkpoint_info': {'filename': 'F:\FLUX 2\webui\models\Stable-diffusion\chroma_v35DetailCalibrated.safetensors', 'hash': '07a82cde'}, 'additional_modules': ['F:\FLUX 2\webui\models\text_encoder\t5xxl_fp8_e4m3fn.safetensors', 'F:\FLUX 2\webui\models\VAE\ae.safetensors', 'F:\FLUX 2\webui\models\text_encoder\diffusion_pytorch_model.safetensors'], 'unet_storage_dtype': None}
StateDict Keys: {'transformer': 643, 'vae': 244, 'text_encoder_2': 220, 'ignore': 0}
Traceback (most recent call last):
File "F:\FLUX 2\webui\modules_forge\main_thread.py", line 30, in work
self.result = self.func(*self.args, **self.kwargs)
File "F:\FLUX 2\webui\modules\txt2img.py", line 131, in txt2img_function
processed = processing.process_images(p)
File "F:\FLUX 2\webui\modules\processing.py", line 836, in process_images
manage_model_and_prompt_cache(p)
File "F:\FLUX 2\webui\modules\processing.py", line 804, in manage_model_and_prompt_cache
p.sd_model, just_reloaded = forge_model_reload()
File "F:\FLUX 2\system\python\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "F:\FLUX 2\webui\modules\sd_models.py", line 504, in forge_model_reload
sd_model = forge_loader(state_dict, additional_state_dicts=additional_state_dicts)
File "F:\FLUX 2\system\python\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "F:\FLUX 2\webui\backend\loader.py", line 287, in forge_loader
component = load_huggingface_component(estimated_config, component_name, lib_name, cls_name, local_path, component_sd)
File "F:\FLUX 2\webui\backend\loader.py", line 61, in load_huggingface_component
assert isinstance(state_dict, dict) and len(state_dict) > 16, 'You do not have CLIP state dict!'
AssertionError: You do not have CLIP state dict!
You do not have CLIP state dict!

File "F:\FLUX 2\webui\backend\loader.py", line 61, in load_huggingface_component
assert isinstance(state_dict, dict) and len(state_dict) > 16, 'You do not have CLIP state dict!'
AssertionError: You do not have CLIP state dict!
You do not have CLIP state dict!
Loading Model: {'checkpoint_info': {'filename': 'F:\FLUX 2\webui\models\Stable-diffusion\chroma_v35DetailCalibrated.safetensors', 'hash': '07a82cde'}, 'additional_modules': ['F:\FLUX 2\webui\models\text_encoder\t5xxl_fp8_e4m3fn.safetensors', 'F:\FLUX 2\webui\models\VAE\ae.safetensors', 'F:\FLUX 2\webui\models\text_encoder\diffusion_pytorch_model.safetensors'], 'unet_storage_dtype': None}
StateDict Keys: {'transformer': 643, 'vae': 244, 'text_encoder_2': 220, 'ignore': 0}
Traceback (most recent call last):
File "F:\FLUX 2\webui\modules_forge\main_thread.py", line 30, in work
self.result = self.func(*self.args, **self.kwargs)
File "F:\FLUX 2\webui\modules\txt2img.py", line 131, in txt2img_function
processed = processing.process_images(p)
File "F:\FLUX 2\webui\modules\processing.py", line 836, in process_images
manage_model_and_prompt_cache(p)
File "F:\FLUX 2\webui\modules\processing.py", line 804, in manage_model_and_prompt_cache
p.sd_model, just_reloaded = forge_model_reload()
File "F:\FLUX 2\system\python\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "F:\FLUX 2\webui\modules\sd_models.py", line 504, in forge_model_reload
sd_model = forge_loader(state_dict, additional_state_dicts=additional_state_dicts)
File "F:\FLUX 2\system\python\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "F:\FLUX 2\webui\backend\loader.py", line 287, in forge_loader
component = load_huggingface_component(estimated_config, component_name, lib_name, cls_name, local_path, component_sd)
File "F:\FLUX 2\webui\backend\loader.py", line 61, in load_huggingface_component
assert isinstance(state_dict, dict) and len(state_dict) > 16, 'You do not have CLIP state dict!'
AssertionError: You do not have CLIP state dict!
You do not have CLIP state dict!