[FlaxStableDiffusionPipeline] fix bug when nsfw is detected (#832)
fix nsfw bug
This commit is contained in:
@@ -291,7 +291,8 @@ class FlaxStableDiffusionPipeline(FlaxDiffusionPipeline):
|
|||||||
# block images
|
# block images
|
||||||
if any(has_nsfw_concept):
|
if any(has_nsfw_concept):
|
||||||
for i, is_nsfw in enumerate(has_nsfw_concept):
|
for i, is_nsfw in enumerate(has_nsfw_concept):
|
||||||
images[i] = np.asarray(images_uint8_casted[i])
|
if is_nsfw:
|
||||||
|
images[i] = np.asarray(images_uint8_casted[i])
|
||||||
|
|
||||||
images = images.reshape(num_devices, batch_size, height, width, 3)
|
images = images.reshape(num_devices, batch_size, height, width, 3)
|
||||||
else:
|
else:
|
||||||
|
|||||||
Reference in New Issue
Block a user