| |
|
| | --- |
| | license: creativeml-openrail-m |
| | base_model: SG161222/Realistic_Vision_V4.0 |
| | datasets: |
| | - recastai/LAION-art-EN-improved-captions |
| | tags: |
| | - stable-diffusion |
| | - stable-diffusion-diffusers |
| | - text-to-image |
| | - diffusers |
| | inference: true |
| | --- |
| | |
| | # Text-to-image Distillation |
| | |
| | This pipeline was distilled from **SG161222/Realistic_Vision_V4.0** on a Subset of **recastai/LAION-art-EN-improved-captions** dataset. Below are some example images generated with the tiny-sd model. |
| |
|
| |  |
| |
|
| |
|
| | This Pipeline is based upon [the paper](https://arxiv.org/pdf/2305.15798.pdf). Training Code can be found [here](https://github.com/segmind/distill-sd). |
| |
|
| | ## Pipeline usage |
| |
|
| | You can use the pipeline like so: |
| |
|
| | ```python |
| | from diffusers import DiffusionPipeline |
| | import torch |
| | |
| | pipeline = DiffusionPipeline.from_pretrained("segmind/tiny-sd", torch_dtype=torch.float16) |
| | prompt = "Portrait of a pretty girl" |
| | image = pipeline(prompt).images[0] |
| | image.save("my_image.png") |
| | ``` |
| |
|
| | ## Training info |
| |
|
| | These are the key hyperparameters used during training: |
| |
|
| | * Steps: 125000 |
| | * Learning rate: 1e-4 |
| | * Batch size: 32 |
| | * Gradient accumulation steps: 4 |
| | * Image resolution: 512 |
| | * Mixed-precision: fp16 |
| |
|
| | ## Speed Comparision |
| |
|
| | We have observed that the distilled models are upto 80% faster than the Base SD1.5 Models. Below is a comparision on an A100 80GB. |
| |
|
| |  |
| |  |
| |
|
| | [Here](https://github.com/segmind/distill-sd/blob/master/inference.py) is the code for benchmarking the speeds. |
| |
|
| |
|
| |
|
| |
|