fix: disable extra T5 mask padding for Wan (#1375)

This commit is contained in:
leejet 2026-03-30 23:10:07 +08:00 committed by GitHub
parent 02dd5e5dd2
commit 8d878872d9
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 3 additions and 3 deletions

View File

@ -1313,14 +1313,14 @@ struct T5CLIPEmbedder : public Conditioner {
std::shared_ptr<T5Runner> t5;
size_t chunk_len = 512;
bool use_mask = false;
int mask_pad = 1;
int mask_pad = 0;
bool is_umt5 = false;
T5CLIPEmbedder(ggml_backend_t backend,
bool offload_params_to_cpu,
const String2TensorStorage& tensor_storage_map = {},
bool use_mask = false,
int mask_pad = 1,
int mask_pad = 0,
bool is_umt5 = false)
: use_mask(use_mask), mask_pad(mask_pad), t5_tokenizer(is_umt5) {
bool use_t5 = false;

View File

@ -493,7 +493,7 @@ public:
offload_params_to_cpu,
tensor_storage_map,
true,
1,
0,
true);
diffusion_model = std::make_shared<WanModel>(backend,
offload_params_to_cpu,