Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement the g-diffuser in/outpainting methods #309

Closed
ilikenwf opened this issue Sep 12, 2022 · 1 comment
Closed

Implement the g-diffuser in/outpainting methods #309

ilikenwf opened this issue Sep 12, 2022 · 1 comment
Labels
enhancement New feature or request

Comments

@ilikenwf
Copy link

Current Implementation by @parlance-zz
https://github.com/parlance-zz/g-diffuser-bot/tree/g-diffuser-bot-diffuserslib-beta

Examples:
https://old.reddit.com/r/StableDiffusion/comments/xbjnnu/huge_outpainting_in_1_step_without_erased_colors/

The Author's Description:

Explanation: Getting good results in/out-painting with stable diffusion can be challenging. Although there are simpler effective solutions for in-painting, out-painting can be especially challenging because there is no color data in the masked area to help prompt the generator. Ideally, even for in-painting we'd like work effectively without that data as well. Provided here is my take on a potential solution to this problem.

By taking a fourier transform of the masked src img we get a function that tells us the presence and orientation of each feature scale in the unmasked src. Shaping the init/seed noise for in/outpainting to the same distribution of feature scales, orientations, and positions increases output coherence by helping keep features aligned. This technique is applicable to any continuous generation task such as audio or video, each of which can be conceptualized as a series of out-painting steps where the last half of the input "frame" is erased. For multi-channel data such as color or stereo sound the "color tone" or histogram of the seed noise can be matched to improve quality (using scikit-image currently) This method is quite robust and has the added benefit of being fast independently of the size of the out-painted area. The effects of this method include things like helping the generator integrate the pre-existing view distance and camera angle.

@AUTOMATIC1111 AUTOMATIC1111 added the enhancement New feature or request label Sep 12, 2022
@ilikenwf
Copy link
Author

This and #301 are duplicates, I'll close this one.

Sashimimochi pushed a commit to Sashimimochi/stable-diffusion-webui that referenced this issue Apr 7, 2023
Sashimimochi pushed a commit to Sashimimochi/stable-diffusion-webui that referenced this issue Apr 7, 2023
nne998 pushed a commit to fjteam/stable-diffusion-webui that referenced this issue Sep 26, 2023
Atry pushed a commit to Atry/stable-diffusion-webui that referenced this issue Jul 9, 2024
Update tilediffusion.py to Fix 'Attribute Not Found' error.
Atry pushed a commit to Atry/stable-diffusion-webui that referenced this issue Jul 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants