Many sampling methods are available in AUTOMATIC1111. Euler a, Heun, DDIM… What are samplers? How do they work? What is the difference between them? Which one should you use? You will find the answers in this article.

We will discuss the samplers available in AUTOMATIC1111 Stable Diffusion GUI. You can use this GUI on Windows, Mac, or Google Colab.

## What is Sampling?

To produce an image, Stable Diffusion first generates a completely random image in the latent space. The **noise predictor** then estimates the noise of the image. The **predicted noise** is subtracted from the image. This process is repeated a dozen times. In the end, you get a clean image.

This **denoising process** is called **sampling** because Stable Diffusion generates a new sample image in each step. The method used in sampling is called the **sampler** or **sampling method**.

Sampling is just one part of the Stable Diffusion model. Read the article “How does Stable Diffusion work?” if you want to understand the whole model.

Below is a sampling process in action. The sampler gradually produces cleaner and cleaner images.

While the framework is the same, there are many different ways to carry out this denoising process. It is often a trade-off between speed and accuracy.

### Noise schedule

You must have noticed the noisy image gradually turns into a clear one. The **noise schedule** controls the **noise level at each sampling step**. The noise is highest at the first step and gradually reduces to zero at the last step.

At each step, the sampler’s job is to **produce an image with a noise level matching the noise schedule**.

What’s the effect of increasing the** number of sampling steps**? A smaller noise reduction between each step. This helps to reduce the truncation error of the sampling.

Compare the noise schedules of 15 steps and 30 steps below.

## Samplers overview

At the time of writing, there are 19 samplers available in AUTOMATIC1111. The number seems to be growing over time. What are the differences?

You will learn what they are in the later part of this article. The technical details can be overwhelming. So I include a birdseye view in this section. This should help you get a general idea of what they are.

### Old-School ODE solvers

Let’s knock out the easy ones first. Some of the samplers on the list were invented more than a hundred years ago. They are old-school solvers for ordinary differential equations (ODE).

**Euler**– The simplest possible solver.**Heun**– A more accurate but slower version of Euler.**LMS**(Linear multi-step method) – Same speed as Euler but (supposedly) more accurate.

### Ancestral samplers

Do you notice some sampler’s names have a single letter “a”?

- Euler a
- DPM2 a
- DPM++ 2S a
- DPM++ 2S a Karras

They are **ancestral samplers**. An ancestral sampler adds noise to the image at each sampling step. They are **stochastic** samplers because the sampling outcome has some randomness to it.

Be aware that many others are also stochastic samplers, even though their names do not have an “a” in them.

The drawback of using an ancestral sampler is that the image would not converge. Compare the images generated using Euler a and Euler below.

Images generated with Euler a do not converge at high sampling steps. In contrast, images from Euler converge well.

**For reproducibility, it is desirable to have the image converge**. If you want to generate slight variations, you should use variational seed.

### Karras noise schedule

The samplers with the label “**Karras**” use the **noise schedule** recommended in the Karras article. If you look carefully, you will see the noise step sizes are smaller near the end. They found that this improves the quality of images.

### DDIM and PLMS

DDIM (Denoising Diffusion Implicit Model) and PLMS (Pseudo Linear Multi-Step method) were the samplers shipped with the original Stable Diffusion v1. DDIM is one of the first samplers designed for diffusion models. PLMS is a newer and faster alternative to DDIM.

They are generally seen as outdated and not widely used anymore.

### DPM and DPM++

**DPM** (Diffusion probabilistic model solver) and **DPM++** are new samplers designed for diffusion models released in 2022. They represent a family of solvers of similar architecture.

**DPM** and** DPM2** are similar except for DPM2 being second order (More accurate but slower).

**DPM++** is an improvement over DPM.

**DPM adaptive** adjusts the step size adaptively. It can be slow since it doesn’t guarantee finishing within the number of sampling steps.

### UniPC

UniPC (Unified Predictor-Corrector) is a new sampler released in 2023. Inspired by the predictor-corrector method in ODE solvers, it can achieve high-quality image generation in 5-10 steps.

### k-diffusion

Finally, you may have heard the term **k-diffusion** and wondered what it means. It simply refers to Katherine Crowson’s k-diffusion GitHub repository and the samplers associated with it.

The repository implements the samplers studied in the Karras 2022 article.

Basically, all samplers in AUTOMATIC1111 except DDIM, PLMS, and UniPC are borrowed from k-diffusion.

## Evaluating samplers

How to pick a sampler? You will see some objective comparisons in this section to help you decide.

### Image Convergence

In this section, I will generate the same image using different samplers with up to 40 sampling steps. The last image at the 40th step is used as a reference for evaluating how quickly the sampling converges. The Euler method will be used as the reference.

#### Euler, DDIM, PLMS, LMS Karras and Heun

First, let’s look at the Euler, DDIM, PLMS, LMS Karras, and Heun as a group since they represent old-school ODE solvers or original diffusion solvers. DDIM converges at about the steps as Euler but with more variations. This is because it injects random noise during its sampling steps.

**PLMS** did not fare very well in this test.

**LMS Karras** seems to have a hard time converging and has stabilized at a higher baseline.

**Heun** converges faster but is two times slower since it is a 2nd order method. So we should compare Heun at 30 steps with Euler at 15 steps, for example.

#### Ancestral samplers

**If a stable, reproducible image is your goal, you should not use ancestral samplers**. All the ancestral samplers do not converge.

#### DPM and DPM2

**DPM fast** did not converge well. **DPM2** and **DPM2 Karras** performs better than Euler but again in the expense of being two times slower.

**DPM adaptive** performs deceptively well because it uses its own adaptive sampling steps. It can be very slow.

### DPM++ solvers

**DPM++ SDE** and **DPM++ SDE Karras** suffer the same shortcoming as ancestral samplers. They not only don’t converge, but the images also fluctuate significantly as the number of steps changes.

**DPM++ 2M** and **DPM++ 2M Karras** perform well. The Karras variant converges faster when the number of steps is high enough.

### UniPC

UniPC converges a bit slower than Euler, but not too bad.

### Speed

Although **DPM adaptive** performs well in convergence, it is also the slowest.

You may have noticed the rest of the rendering times fall into two groups, with the first group taking about the same time (about 1x), and the other group taking about twice as long (about 2x). This simply reflects the order of the solvers. 2nd order solvers, although more accurate, need to evaluate the denoising U-Net twice. So it is 2x slower.

### Quality

Of course, speed and convergence mean nothing if the images look crappy.

#### Final images

Let’s first look at samples of the image.

DPM++ fast failed pretty badly. Ancestral samples did not converge to the image that other samplers converged to.

Ancestral samplers tend to converge to an image of a kitten, while the deterministic ones tend to converge to a cat. There are no correct answers as long as they look good to you.

#### Perceptual quality

An image can still look good even if it hasn’t converged. Let’s look at how quickly each sampler can produce a high-quality image.

You will see perceptual quality as measured with BRISQUE (Blind/Referenceless Image Spatial Quality Evaluator). It is an evaluator for measuring the quality of natural images.

DDIM is doing surprisingly well here, capable of producing the highest quality image within the group in as few as 8 steps.

With one or two exceptions, all ancestral samplers perform similarly to Euler in generating quality images.

DPM2 samplers slightly outperform Euler.

**DPM++ SDE** and **DPM++ SDE Karras** performed the best in this quality test.

UniPC is slightly worse than Euler in low steps but comparable to it in high steps.

### So… which one is the best?

Here are my recommendations:

- If you want to use something fast, converging, new, and with decent quality, excellent choices are
**DPM++ 2M Karras**with 20 – 30 steps**UniPC**with 20-30 steps.

- If you want good quality images and don’t care about convergence, good choices are
**DPM++ SDE Karras**with 8-12 steps (Note: This is a slower sampler)**DDIM**with 10-15 steps.

- Avoid using any ancestral samplers if you prefer stable, reproducible images.
**Euler**and**Heun**are fine choices if you prefer something simple. Reduce the number of steps for Heun to save time.

## Samplers Explained

You will find information on samplers available in AUTOMATIC1111. The inner working of these samplers is quite mathematical in nature. I will only explain Euler (The simplest one) in detail. Many of them share elements of Euler.

### Euler

**Euler is the most straightforward sampler possible**. It is mathematically identical to Euler’s method for solving ordinary differential equations. It is entirely **deterministic**, meaning no random noise is added during sampling.

Below is the sampling step-by-step.

**Step 1**: Noise predictor estimates the noise image from the latent image.

**Step 2**: Calculate the amount of noise needed to be subtracted according to the **noise schedule**. That is the difference in noise between the current and the next step.

**Step 3**: Subtract the latent image by the normalized noise image (from step 1) multiplied by the amount of noise to be reduced (from step 2).

Repeat steps 1 to 3 until the end of the noise schedule.

#### Noise schedule

But how do you know the amount of noise in each step? Actually, this is something you tell the sampler.

**A noise schedule tells the sampler how much noise there should be at each step. **Why does the model need this information? The noise predictor estimates the noise in the latent image based on the total amount of noise supposed to be there. (This is how it was trained.)

There’s the highest amount of noise in the first step. The noise gradually decreases and is down to zero at the last step.

Changing the number of sampling steps changes the noise schedule. Effectively, the noise schedule gets smoother. A higher number of sampling steps has a smaller reduction in noise between any two steps. This helps to reduce truncation errors.

#### From random to deterministic sampling

Do you wonder why you can solve a random sampling problem with a deterministic ODE solver? This is called the **probability flow** formulation. Instead of solving for how a sample evolves, you solve for the evolution of its probability distribution. This is the same as solving for the probability distribution instead of sample trajectories in a stochastic process.

Compared with a drift process, these ODE solvers use the following mappings.

- Time → noise
- Time quantization → noise schedule
- Position → latent image
- Velocity → Predicted noise
- Initial position → Initial random latent image
- Final position → Final clear latent image

#### Sampling example

Below is an example of text-to-image using Euler’s method. The noise schedule dictates the noise level in each step. The sampler’s job is to reduce the noise by just the right amount in each step to match the noise schedule until it is zero at the last step.

### Euler a

**Euler ancestral (Euler a) sampler** is similar to Euler’s sampler. But at each step, it subtracts more noise than it should and adds some random noise back to match the noise schedule. The denoised image depends on the specific noise added in the previous steps. So it is an **ancestral sampler**, in the sense that the path the image denoises depends on the specific random noises added in each step. The result would be different if you were to do it again.

### DDIM

Denoising Diffusion Implicit Models (DDIM) is one of the first samplers for solving diffusion models. It is based on the idea that the image at each step can be approximated by adding the following three components.

- Final image
- Image direction pointing to the image at the current step
- Random noise

How do we know the final image before we get to the last step? The DDIM sampler approximates it with the denoised image. Similarly, the image direction is approximated by the noise estimated by the noise predictor.

### LMS and LMS Karras

Much like Euler’s method, the **linear multistep method (LMS)** is a standard method for solving ordinary differential equations. It aims at improving accuracy by clever use of the values of the previous time steps. AUTOMATIC1111 defaults to use up to 4 last values.

**LMS Karras** uses the Karras noise schedule.

### Heun

Heun’s method is a more accurate improvement to Euler’s method. But it needs to predict noise twice in each step, so it is twice as slow as Euler.

### DPM samplers

**Diffusion Probabilistic Model Solvers** (DPM-Solvers) belong to a family of newly developed solvers for diffusion models. They are the following solvers in AUTOMATIC1111.

- DPM2
- DPM2 Karras
- DPM2 a
- DPM2 a Karras
- DPM Fast
- DPM adaptive
- DPM Karras

**DPM2** is the DPM-Solver-2 (Algorithm 1) of the DPM-Solver article. The solver is accurate up to the second order.

**DPM2 Karras** is identical to DPM2 except for using the Karras noise scheduler.

**DPM2 a** is almost identical to DPM2, except noise is added to each sampling step. This makes it an ancestral sampler.

**DPM2 a Karras** is almost identical to DPM2 a, except for using the Karras noise schedule.

**DPM Fast** is a variant of the DPM solver with a uniform noise schedule. It is accurate up to the first order. So it is twice as fast as DPM2.

**DPM adaptive** is a first-order DPM solver with an adaptive noise schedule. It ignores the number of steps you set and adaptively determines its own.

**DPM++** samplers are the improved versions of DPM.

### UniPC

UniPC (Unified Predictor Corrector method) is a diffusion sampler newly developed in 2023. It consists of two parts

- Unified predictor (UniP)
- Unified corrector (UniC)

It supports any solver and noise predictors.

## More readings

- k-diffusion GitHub page – Katherine Crowson’s diffusion library. Many samples in AUOMATIC1111 are borrowed from this repository.
- Elucidating the Design Space of Diffusion-Based Generative Models (Karras 2022) – k-diffusion implements the samplers described in this article.
- Progressive Distillation for Fast Sampling of Diffusion Models – Fast sampling progressive distillation can generate images in as few as 4 steps. It needs model-level training.
- Pseudo Numerical Methods for Diffusion Models on Manifolds (Liu 2022) – Introducing PLMS.
- DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps (Lu 2022) – Introducing DPM and DPM2 solvers.
- DPM-Solver++: Fast Solver for Guided Sampling of Diffusion Probabilistic Models (Lu 2022) – Introducing DPM ++ solvers.
- Denoising Diffusion Implicit Models (Song 2020) – The DDIM sampler.
- Score-based generative modeling through stochastic differential equations (Song 2020) – Introducing reverse diffusion. We have ODE solvers thanks to the authors’ observation of the alternative probability flow formulation.
- Pseudo Numerical Methods for Diffusion Models on Manifolds (Liu 2022) – The PLMS sampler.
- This Reddit comment offers an excellent summary of samplers.

there is a new DPM++ 2M SDE Karras, how good is it compared to others? Thank you.

Hey, THANK YOU!!!

Interesting article with a lot of new information for me, but in practice this did not really get into how they react in real life to multiple parameters. Sometimes, to offset high CFG values, much more steps than expected need to be used, and while many of the latter samplers can create good output in 15-25 steps, often I find that small details are better and there’s “more information” in the image with higher step counts of up to 60-80 steps, especially for higher CFG values around 12-16.

If a particular sampling method reaches the low point in the curve at x steps, will it still reach that same point x at a higher image resolution, or with a different text prompt? And is there any reason to set the steps count higher than X? I see so many examples on different sites where they set the steps to like 70 or higher. Is there any benefit in quality when the steps is set to the max?

Thanks,

Eric

Generally speaking, you should not deviate too much from the native resolution (512px for v1 models). Higher resolution should be achieved using an upscaler.

Once the image converges or reaches a state of constant changes (with ancestral samplers), there’s no point in running more steps. Running more steps is technically more accurate, but we don’t know or don’t care about the true solution anyway.

Hi!

In your section on recommendations, you list several things, in particular, though, item 1:

“If you want to use something fast, converging, new, and with decent quality, excellent choices are ”

What seems to be missing is “If you don’t care about fast, but do want converging, new and with best quality,” or something like that.

Speed is of no relevance to me at all. I am most interested in getting the highest quality result.

Having read the article, my interpretation of your data is that the best are DPM++ SDE and DPM++ SDE Karras. Does that seem right?

Sounds reasonable to me.

Thank you so much for all your hard work! This is my new favorite site for all things SD. This guide explained everything perfectly.

Welcome!

just wanted to say thanks for all the helpful guides… always seem to pop up when I’m searching for something, and I really appreciate it!

You are welcome!

The Euler method (in K diffusion) has its own paper, as the “algorithm 2” of the “momentum sampler”. See if it still matches the Euler method you’ve descrived.

Paper: Soft Diffusion: Score Matching for General Corruptions

https://arxiv.org/abs/2209.05442

Thanks for pointing this paper out. Looks like the Euler method as used in A1111 is the algorithm 1 – Navie sampler but without the noise term (even more naive…)

Thanks for the article. I really enjoy it

Thank you very much, very detailed explanation, I learned a lot.

This blog guided me out of the sampler chaos. Thank you for sharing!

Thank goodness someone has finally clarified these samplers – thanks! (not that I *fully* understand, but this article certainly helps a lot). So for doing ‘comparitive’ tests between samplers, all other settings (including seed) being equal, it’s probably best to avoid the ancestral samplers. That way you can compare your (identical) final images on an ‘apples to apples’ basis.

Thanks a lot. This article let me know my wrong choice of simpler…