Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reducing inference timings for Sd2.1 Base model #64

Open
pratos opened this issue Dec 22, 2022 · 1 comment
Open

Reducing inference timings for Sd2.1 Base model #64

pratos opened this issue Dec 22, 2022 · 1 comment

Comments

@pratos
Copy link

pratos commented Dec 22, 2022

I managed to shave off inference timings for SD2.1 by a few seconds for 512x512 (50 steps) and 768x768 (50 Steps).

Using just few additions:

torch.backends.cudnn.benchmark = True
torch.backends.cuda.matmul.allow_tf32 = True

pipe = StableDiffusionPipeline.from_pretrained(
            MODEL_ID,
            cache_dir=MODEL_CACHE,
            local_files_only=True,
        )
pipe = pipe.to("cuda")

pipe.enable_xformers_memory_efficient_attention()
pipe.enable_vae_slicing()

Overall output didn't suffer coz of this. Getting crisp images. Wanted to know how do I create a PR to add these? And are there any tests around this?

Here are the inferences:

@pratos
Copy link
Author

pratos commented Dec 22, 2022

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant