---
license: creativeml-openrail-m
---
**Spider-Verse Diffusion**

This is the fine-tuned Stable Diffusion model trained on movie stills from Sony's Into the Spider-Verse.
Use the tokens **_spiderverse style_** in your prompts for the effect.

If you enjoy this model, please check out my other models on [Huggingface](https://huggingface.co/nitrosocke)

### 🧨 Diffusers

This model can be used just like any other Stable Diffusion model. For more information,
please have a look at the [Stable Diffusion](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).

You can also export the model to [ONNX](https://huggingface.co/docs/diffusers/optimization/onnx), [MPS](https://huggingface.co/docs/diffusers/optimization/mps) and/or [FLAX/JAX]().

```python
#!pip install diffusers transformers scipy torch
from diffusers import StableDiffusionPipeline
import torch

model_id = "nitrosocke/spider-verse-diffusion"
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
pipe = pipe.to("cuda")

prompt = "a magical princess with golden hair, spiderverse style"
image = pipe(prompt).images[0]

image.save("./magical_princess.png")
```

**Portraits rendered with the model:**
![Portrait Samples](https://huggingface.co/nitrosocke/spider-verse-diffusion/resolve/main/spiderverse-portraits-small.jpg)
**Sample images used for training:**
![Training Samples](https://huggingface.co/nitrosocke/spider-verse-diffusion/resolve/main/spiderverse-training-small.jpg)

This model was trained using the diffusers based dreambooth training and prior-preservation loss in 3.000 steps.