Dgenerate

Latest version: v4.4.1

Safety actively analyzes 681881 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 3 of 10

4.1.0

* Stable Cascade can now use long prompts with weighting via the ``compel`` and ``sd-embed`` prompt weighters, simply specify one of these two values to ``--prompt-weighter`` to enable the feature when using Stable Cascade.

* Built and published wheels for MacOS on Apple Silicon, dgenerate can now be used with ``--device mps`` on MacOS (arm64). See install instructions here: [MacOS Install (Apple Silicon Only)](https://github.com/Teriks/dgenerate?tab=readme-ov-file#macos-install-apple-silicon-only). This is very experimental.

* Console UI hotkeys dynamically adjust for useability when the UI is running on MacOS, I am sure that the hotkeys are very abnormal when compared to typical MacOS convention for text editors, but they all function and do not overlap with important system hotkeys. `Insert` (for multiline input) changes to `Command+i`, and `Ctrl+Space` (for run) changes to `Command+r`. All other hotkeys remain the same, including copy, paste, and cut.

* Added template function ``frange`` which is just ``range`` but for float ranges. This is useful for stepping through guidance scale values or LoRA scale values for example. It uses identical logic to ``range``, but the step value may be a float, e.g. ``frange(0.1, 1.0, 0.2) -> [0.1, 0.3, 0.5, 0.7, 0.9]``. The default step value is: ``0.1``

* torch 2.4.1

4.0.3

runwayml has removed their stable diffusion 1.5 repository from Hugging Face, breaking single file loads in the diffusers library and anything that uses it.

This release upgrades diffusers to 0.30.2 where the issue has been worked around by instead referencing `Lykon/dreamshaper-8`

All references to the runwayml model on Hugging Face have been removed from this project, including in documentation.

4.0.2

Add the arguments ``--flux-second-prompts``, ``--flux-max-sequence-length``, and ``--lora-fuse-scale`` to the argument
reconstruction used by ``--output-configs`` and ``--output-metadata``. These arguments were not being replicated in config output.

4.0.1

Add missing VAE tiling checkbox for Flux recipes in Console UI recipe form.

4.0.0

The command line interface remains backwards compatible for all supported model types, the major version change is mostly due to breaking library API changes and the complete removal of Flax/Jax support.

All support for Flax/Jax has been removed because there is only minimal support or adoption of it anywhere currently, and seemingly no work being done on it in the diffusers library. Also this alleviates quite a bit of dependency hell and possible issues caused by having torch and Flax/Jax in the same package as dependencies.

* Support for Flux in `txt2img` mode, add associated options: `--flux-second-prompts`, `--flux-max-sequence-length`. Single file loads from a `.safetensor` file is partially supported for Flux, you must specify the VAE and Text Encoders manually.

* Add Flux support to `sd-embed` prompt weighter internal plugin, prompt weighting syntax can be used with Flux when `--prompt-weighter sd-embed` is specified.

* New Console UI recipes for Flux that are optimized by default to run (slowly) on even old hardware or GPUs with less than 12gb of VRAM.

* Stable Diffusion 3 now supports per LoRA scale values.

* The option `--lora-fuse-scale` has been added to control the LoRA scale once all LoRA models have been fused together at their individual scale values. This controls at what scale the merged LoRA weights are fused into the main model, these weights possibly being merged from multiple LoRA models at different individual scales.

* Support loading a transformer model for new diffusion architectures based on transformer architecture (SD3 and Flux) via the argument `--transformer`, this new option is nearly identical to `--unet/--unet2`. Transformers can be quantized via the `quantize` URI argument using `optimum.quanto`. The quantize argument accepts a string value which is the datatype for quantization, for example `qint8`, `qfloat8`, etc. Acceptable values are listed in the `--help` text of the `--transformer` argument.

* Text encoders specified with `--text-encoders/--text-encoders2` can now be quantized via the `quantize` URI argument using `optimum.quanto`. Acceptable values are listed in the `--help` text of the `--text-encoders` argument.

* Support loading multiple IP Adapter models with `--ip-adapters`.

* Support loading an image encoder with `--image-encoder` for use with IP Adapters and Stable Cascade.

* New `--image-seeds` syntax for specifying IP Adapter images, see: [Specifying IP Adapters](https://github.com/Teriks/dgenerate/tree/master?tab=readme-ov-file#specifying-ip-adapters)

* Support loading multiple T2I Adapter models with `--t2i-adapters`, T2I Adapters function identically to Control Nets in terms of image input specification with ``--image-seeds``. They are essentially another form of control guidance model, and support Open Pose / Depth inputs etc. similar to Control Nets. T2I Adapters are a bit less resource intensive, but do not currently support `img2img` or `inpainting`, only `txt2img`.

* Support image input batching for `img2img` and `inpainting` mask images, allowing multiple input images to be processed on the GPU at once. New `--image-seeds` syntax feature added to support this, see: [Batching Input Images and Inpaint Masks](https://github.com/Teriks/dgenerate/tree/master?tab=readme-ov-file#batching-input-images-and-inpaint-masks)

* `--seed-image-processors` and `--mask-image-processors` now support the + syntax previously supported by `--control-image-processors` in order to allow for selective processing of input images when multiple `img2img` or `inpaint` masks are specified. This is covered in the [Image Processors](https://github.com/Teriks/dgenerate/tree/master?tab=readme-ov-file#image-processors) section: [Multiple control net images, and input image batching](https://github.com/Teriks/dgenerate/tree/master?tab=readme-ov-file#multiple-control-net-images-and-input-image-batching)

* Added the `resize` image processor which implements basic image resizing using the usual resampling algorithms in Pillow. This can be useful for resizing input images to Stable Cascade when multiple are provided, as there is not a mechanism to do so from `--image-seeds`

* Stable Cascade can now use the new input batching syntax of `--image-seeds` to process multiple input images as a style reference, Stable Cascade does not use multiple images for batching like other pipelines, but as image prompts.

* Improved `--help` output formatting and readability for `dgenerate` and `dgenerate --sub-command image-process`

* torch 2.4.0 and diffusers 0.30.0

3.10.4

Backport console shell process monitor from t2iadapter (4.0.0) branch.

This solves a console GUI hang when ``\exit`` is used or when the shell process exits unexpectedly on its own.

Page 3 of 10

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.