Skip to main content

Fine-tune Stable Diffusion for Textual Inversion with Determined AI

This guide, based on Determined AI's article Personalizing Stable Diffusion with Determined, explains how fine-tune a Stable Diffusion model on CoreWeave Cloud to do Textual Inversion, generating personalized images.

Stable Diffusion is the latest deep learning model to generate brilliant, eye-catching art based on simple input text. Built upon the ideas behind models such as DALL·E 2, Imagen, and LDM, Stable Diffusion is the first architecture in this class which is small enough to run on typical consumer-grade GPUs.


This guide assumes that the following are completed in advance.

Configure the experiment

In a text editor, open the files in /examples/diffusion/textual_inversion_stable_diffusion/. Change the values in finetune_const.yaml config file, as needed.

- object
- det-logo
- brain logo, sharp lines, connected circles, concept art
- det_logos

These values may be broken down as follows:

Value nameDescription
YOUR_HF_AUTH_TOKEN_HEREYour Hugging Face User Access Token
learnable_propertiesEither object or style, depending on whether you want to capture the object itself, or only its style

More than one concept at a time may be fine-tuned by appending relevant information to the lists:

  • concept_strs - the string used to refer to new concept in prompts
  • initializer_strs - a short phrase which roughly describes the concept of interest and provides a warm-start for fine-tuning

Clone the repository

To fine-tune images, first clone the Determined repository.

Acquire a Hugging Face User Access Token

Create a Hugging Face account if necessary, then generate a User Access Token. Accept the Stable Diffusion license at CompVis/stable-diffusion-v1-4 by clicking "Access repository".

Place images

In a new directory at the root of the repository, place the training images.

Submit the experiment

Submit the fine-tuning experiment by navigating to the root of the repository, then create the experiment by running det e create:

$ det e create finetune_const.yaml

To quickly test your configuration, add the HF_AUTH_TOKEN value. Leave all other values at their defaults, then submit the experiment without any further changes. A fine-tuning experiment using the Determined AI logo is set up in the repository by default.

After launching the experiment, navigate to the Web UI. Click the Logs tab to view training progress.

Stable Diffusion with Textual Inversion running on CoreWeave infrastructure using Determined AI

When the experiment completes, the checkpoints are exported to Object Storage (shown as s3).

Showcases all steps completed in the fine-tuning job

You can view Loss on the Overview tab. Here, you can choose Fork or Continue Trial if required to get better results.

Loss going down as more batches are iterated upon

Use the Checkpoints tab to view them.

All listed Checkpoints for the experiment

Use the Hyperparameters tab to see the values passed to the experiment.

HyperParameters for the fine-tuning experiment

You can visualize the results via TensorBoard.

TensorBoard showing inference prompts

Generate Images

After the fine-tuning experiment is complete, you can generate art with the newly-trained concept using Jupyter notebook or with large-scale generation through a Determined experiment. The former is useful for quick interactive experimentation, while the latter is useful for pure performance.

The Jupyter notebook workflow only requires three steps:

  1. Copy the User Access Token into the detsd-notebook.yaml config file, similar to the step in the previous section.
  2. Get the uuid of the desired Checkpoint by navigating to the Experiment page in the Web UI and either clicking on the Checkpoint's flag icon or inspecting the Checkpoints tab.
  3. Launch the textual_inversion.ipynb notebook and copy the Checkpoint uuid into the uuids list in the appropriate cell. You can also do this by executing the following command in the root of the repository:
$ det notebook start --config-file detsd-notebook.yaml --context .

The --context . argument loads the full contents of the repository into the JupyterLab instance, including the textual_inversion.ipynb repo itself and various supporting files.

In particular, these include a demonstration concept stored in which was extensively trained on Determined AI logos, which you can use instead of, or in addition to, one specified by uuid.

After you've prepared the textual_inversion.ipynb notebook, it can be run from top to bottom and the generated images will appear at the end.

Notebook open to run and download dependencies
Fetching the concept to generate images
Generate images from the concept

Here are some example images generated from the fine-tuning experiment:

Example 1
Example 2
Example 3

After you've found promising prompts and parameter settings with Jupyter notebook, images can be generated at scale by submitting a full-fledged experiment.

The generate_grid.yaml file loads pretrained concepts by uuid or local path and controls the generation process by specifying the prompts and parameter settings to scan over. All generated images are logged to Tensorboard for easy access, as shown previously.

By default, you can submit a two-GPU experiment that creates nearly 500 total images with the pre-trained demonstration in by inserting your authorization token, without other changes, then executing:

det e create generate_grid.yaml .


To learn more, see these resources: