fix: disable extra T5 mask padding for Wan (#1375)

This commit is contained in:
leejet 2026-03-30 23:10:07 +08:00 committed by GitHub
parent 02dd5e5dd2
commit 8d878872d9
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 3 additions and 3 deletions

View File

@ -1313,14 +1313,14 @@ struct T5CLIPEmbedder : public Conditioner {
std::shared_ptr<T5Runner> t5; std::shared_ptr<T5Runner> t5;
size_t chunk_len = 512; size_t chunk_len = 512;
bool use_mask = false; bool use_mask = false;
int mask_pad = 1; int mask_pad = 0;
bool is_umt5 = false; bool is_umt5 = false;
T5CLIPEmbedder(ggml_backend_t backend, T5CLIPEmbedder(ggml_backend_t backend,
bool offload_params_to_cpu, bool offload_params_to_cpu,
const String2TensorStorage& tensor_storage_map = {}, const String2TensorStorage& tensor_storage_map = {},
bool use_mask = false, bool use_mask = false,
int mask_pad = 1, int mask_pad = 0,
bool is_umt5 = false) bool is_umt5 = false)
: use_mask(use_mask), mask_pad(mask_pad), t5_tokenizer(is_umt5) { : use_mask(use_mask), mask_pad(mask_pad), t5_tokenizer(is_umt5) {
bool use_t5 = false; bool use_t5 = false;

View File

@ -493,7 +493,7 @@ public:
offload_params_to_cpu, offload_params_to_cpu,
tensor_storage_map, tensor_storage_map,
true, true,
1, 0,
true); true);
diffusion_model = std::make_shared<WanModel>(backend, diffusion_model = std::make_shared<WanModel>(backend,
offload_params_to_cpu, offload_params_to_cpu,