Spaces:
Running
on
Zero
Running
on
Zero
| from dataclasses import dataclass | |
| from typing import List, Optional, Union | |
| import numpy as np | |
| import PIL.Image | |
| from ...utils import ( | |
| BaseOutput, | |
| ) | |
| class StableDiffusionSafePipelineOutput(BaseOutput): | |
| """ | |
| Output class for Safe Stable Diffusion pipelines. | |
| Args: | |
| images (`List[PIL.Image.Image]` or `np.ndarray`) | |
| List of denoised PIL images of length `batch_size` or numpy array of shape `(batch_size, height, width, | |
| num_channels)`. PIL images or numpy array present the denoised images of the diffusion pipeline. | |
| nsfw_content_detected (`List[bool]`) | |
| List of flags denoting whether the corresponding generated image likely represents "not-safe-for-work" | |
| (nsfw) content, or `None` if safety checking could not be performed. | |
| images (`List[PIL.Image.Image]` or `np.ndarray`) | |
| List of denoised PIL images that were flagged by the safety checker any may contain "not-safe-for-work" | |
| (nsfw) content, or `None` if no safety check was performed or no images were flagged. | |
| applied_safety_concept (`str`) | |
| The safety concept that was applied for safety guidance, or `None` if safety guidance was disabled | |
| """ | |
| images: Union[List[PIL.Image.Image], np.ndarray] | |
| nsfw_content_detected: Optional[List[bool]] | |
| unsafe_images: Optional[Union[List[PIL.Image.Image], np.ndarray]] | |
| applied_safety_concept: Optional[str] | |