diff --git a/docs/source/ko/index.mdx b/docs/source/ko/index.mdx index d01dff5c5e00..a83dd0d0b29e 100644 --- a/docs/source/ko/index.mdx +++ b/docs/source/ko/index.mdx @@ -16,48 +16,82 @@ specific language governing permissions and limitations under the License.

-# 🧨 Diffusers - -πŸ€— DiffusersλŠ” μ‚¬μ „ν•™μŠ΅λœ λΉ„μ „ 및 μ˜€λ””μ˜€ ν™•μ‚° λͺ¨λΈμ„ μ œκ³΅ν•˜κ³ , μΆ”λ‘  및 ν•™μŠ΅μ„ μœ„ν•œ λͺ¨λ“ˆμ‹ 도ꡬ μƒμž 역할을 ν•©λ‹ˆλ‹€. - -보닀 μ •ν™•ν•˜κ²Œ, πŸ€— DiffusersλŠ” λ‹€μŒμ„ μ œκ³΅ν•©λ‹ˆλ‹€: - -- 단 λͺ‡ μ€„μ˜ μ½”λ“œλ‘œ 좔둠을 μ‹€ν–‰ν•  수 μžˆλŠ” μ΅œμ‹  ν™•μ‚° νŒŒμ΄ν”„λΌμΈμ„ μ œκ³΅ν•©λ‹ˆλ‹€. ([**Using Diffusers**](./using-diffusers/conditional_image_generation)λ₯Ό μ‚΄νŽ΄λ³΄μ„Έμš”) μ§€μ›λ˜λŠ” λͺ¨λ“  νŒŒμ΄ν”„λΌμΈκ³Ό ν•΄λ‹Ή 논문에 λŒ€ν•œ κ°œμš”λ₯Ό 보렀면 [**Pipelines**](#pipelines)을 μ‚΄νŽ΄λ³΄μ„Έμš”. -- μΆ”λ‘ μ—μ„œ 속도 vs ν’ˆμ§ˆμ˜ μ ˆμΆ©μ„ μœ„ν•΄ μƒν˜Έκ΅ν™˜μ μœΌλ‘œ μ‚¬μš©ν•  수 μžˆλŠ” λ‹€μ–‘ν•œ λ…Έμ΄μ¦ˆ μŠ€μΌ€μ€„λŸ¬λ₯Ό μ œκ³΅ν•©λ‹ˆλ‹€. μžμ„Έν•œ λ‚΄μš©μ€ [**Schedulers**](./api/schedulers/overview)λ₯Ό μ°Έκ³ ν•˜μ„Έμš”. -- UNetκ³Ό 같은 μ—¬λŸ¬ μœ ν˜•μ˜ λͺ¨λΈμ„ end-to-end ν™•μ‚° μ‹œμŠ€ν…œμ˜ ꡬ성 μš”μ†Œλ‘œ μ‚¬μš©ν•  수 μžˆμŠ΅λ‹ˆλ‹€. μžμ„Έν•œ λ‚΄μš©μ€ [**Models**](./api/models)을 μ°Έκ³ ν•˜μ„Έμš”. -- κ°€μž₯ μΈκΈ°μžˆλŠ” ν™•μ‚° λͺ¨λΈ ν…ŒμŠ€ν¬λ₯Ό ν•™μŠ΅ν•˜λŠ” 방법을 λ³΄μ—¬μ£ΌλŠ” μ˜ˆμ œλ“€μ„ μ œκ³΅ν•©λ‹ˆλ‹€. μžμ„Έν•œ λ‚΄μš©μ€ [**Training**](./training/overview)λ₯Ό μ°Έκ³ ν•˜μ„Έμš”. - -## 🧨 Diffusers νŒŒμ΄ν”„λΌμΈ - -λ‹€μŒ ν‘œμ—λŠ” κ³΅μ‹œμ μœΌλ‘œ μ§€μ›λ˜λŠ” λͺ¨λ“  νŒŒμ΄ν”„λΌμΈ, κ΄€λ ¨ λ…Όλ¬Έ, 직접 μ‚¬μš©ν•΄ λ³Ό 수 μžˆλŠ” Colab λ…ΈνŠΈλΆ(μ‚¬μš© κ°€λŠ₯ν•œ 경우)이 μš”μ•½λ˜μ–΄ μžˆμŠ΅λ‹ˆλ‹€. - -| Pipeline | Paper | Tasks | Colab -|---|---|:---:|:---:| -| [alt_diffusion](./api/pipelines/alt_diffusion) | [**AltDiffusion**](https://arxiv.org/abs/2211.06679) | Image-to-Image Text-Guided Generation | -| [audio_diffusion](./api/pipelines/audio_diffusion) | [**Audio Diffusion**](https://github.com/teticio/audio-diffusion.git) | Unconditional Audio Generation | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/teticio/audio-diffusion/blob/master/notebooks/audio_diffusion_pipeline.ipynb) -| [cycle_diffusion](./api/pipelines/cycle_diffusion) | [**Cycle Diffusion**](https://arxiv.org/abs/2210.05559) | Image-to-Image Text-Guided Generation | -| [dance_diffusion](./api/pipelines/dance_diffusion) | [**Dance Diffusion**](https://github.com/williamberman/diffusers.git) | Unconditional Audio Generation | -| [ddpm](./api/pipelines/ddpm) | [**Denoising Diffusion Probabilistic Models**](https://arxiv.org/abs/2006.11239) | Unconditional Image Generation | -| [ddim](./api/pipelines/ddim) | [**Denoising Diffusion Implicit Models**](https://arxiv.org/abs/2010.02502) | Unconditional Image Generation | -| [latent_diffusion](./api/pipelines/latent_diffusion) | [**High-Resolution Image Synthesis with Latent Diffusion Models**](https://arxiv.org/abs/2112.10752)| Text-to-Image Generation | -| [latent_diffusion](./api/pipelines/latent_diffusion) | [**High-Resolution Image Synthesis with Latent Diffusion Models**](https://arxiv.org/abs/2112.10752)| Super Resolution Image-to-Image | -| [latent_diffusion_uncond](./api/pipelines/latent_diffusion_uncond) | [**High-Resolution Image Synthesis with Latent Diffusion Models**](https://arxiv.org/abs/2112.10752) | Unconditional Image Generation | -| [paint_by_example](./api/pipelines/paint_by_example) | [**Paint by Example: Exemplar-based Image Editing with Diffusion Models**](https://arxiv.org/abs/2211.13227) | Image-Guided Image Inpainting | -| [pndm](./api/pipelines/pndm) | [**Pseudo Numerical Methods for Diffusion Models on Manifolds**](https://arxiv.org/abs/2202.09778) | Unconditional Image Generation | -| [score_sde_ve](./api/pipelines/score_sde_ve) | [**Score-Based Generative Modeling through Stochastic Differential Equations**](https://openreview.net/forum?id=PxTIG12RRHS) | Unconditional Image Generation | -| [score_sde_vp](./api/pipelines/score_sde_vp) | [**Score-Based Generative Modeling through Stochastic Differential Equations**](https://openreview.net/forum?id=PxTIG12RRHS) | Unconditional Image Generation | -| [stable_diffusion](./api/pipelines/stable_diffusion/text2img) | [**Stable Diffusion**](https://stability.ai/blog/stable-diffusion-public-release) | Text-to-Image Generation | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/training_example.ipynb) -| [stable_diffusion](./api/pipelines/stable_diffusion/img2img) | [**Stable Diffusion**](https://stability.ai/blog/stable-diffusion-public-release) | Image-to-Image Text-Guided Generation | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/image_2_image_using_diffusers.ipynb) -| [stable_diffusion](./api/pipelines/stable_diffusion/inpaint) | [**Stable Diffusion**](https://stability.ai/blog/stable-diffusion-public-release) | Text-Guided Image Inpainting | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/in_painting_with_stable_diffusion_using_diffusers.ipynb) -| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [**Stable Diffusion 2**](https://stability.ai/blog/stable-diffusion-v2-release) | Text-to-Image Generation | -| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [**Stable Diffusion 2**](https://stability.ai/blog/stable-diffusion-v2-release) | Text-Guided Image Inpainting | -| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [**Stable Diffusion 2**](https://stability.ai/blog/stable-diffusion-v2-release) | Text-Guided Super Resolution Image-to-Image | -| [stable_diffusion_safe](./api/pipelines/stable_diffusion_safe) | [**Safe Stable Diffusion**](https://arxiv.org/abs/2211.05105) | Text-Guided Generation | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/ml-research/safe-latent-diffusion/blob/main/examples/Safe%20Latent%20Diffusion.ipynb) -| [stochastic_karras_ve](./api/pipelines/stochastic_karras_ve) | [**Elucidating the Design Space of Diffusion-Based Generative Models**](https://arxiv.org/abs/2206.00364) | Unconditional Image Generation | -| [unclip](./api/pipelines/unclip) | [Hierarchical Text-Conditional Image Generation with CLIP Latents](https://arxiv.org/abs/2204.06125) | Text-to-Image Generation | -| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Text-to-Image Generation | -| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Image Variations Generation | -| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Dual Image and Text Guided Generation | -| [vq_diffusion](./api/pipelines/vq_diffusion) | [Vector Quantized Diffusion Model for Text-to-Image Synthesis](https://arxiv.org/abs/2111.14822) | Text-to-Image Generation | - -**μ°Έκ³ **: νŒŒμ΄ν”„λΌμΈμ€ ν•΄λ‹Ή λ¬Έμ„œμ— μ„€λͺ…λœ λŒ€λ‘œ ν™•μ‚° μ‹œμŠ€ν…œμ„ μ‚¬μš©ν•œ 방법에 λŒ€ν•œ κ°„λ‹¨ν•œ μ˜ˆμž…λ‹ˆλ‹€. + +# Diffusers + +πŸ€— DiffusersλŠ” 이미지, μ˜€λ””μ˜€, 심지어 λΆ„μžμ˜ 3D ꡬ쑰λ₯Ό μƒμ„±ν•˜κΈ° μœ„ν•œ μ΅œμ²¨λ‹¨ 사전 ν›ˆλ ¨λœ diffusion λͺ¨λΈμ„ μœ„ν•œ λΌμ΄λΈŒλŸ¬λ¦¬μž…λ‹ˆλ‹€. κ°„λ‹¨ν•œ μΆ”λ‘  μ†”λ£¨μ…˜μ„ μ°Ύκ³  μžˆλ“ , 자체 diffusion λͺ¨λΈμ„ ν›ˆλ ¨ν•˜κ³  μ‹Άλ“ , πŸ€— DiffusersλŠ” 두 κ°€μ§€ λͺ¨λ‘λ₯Ό μ§€μ›ν•˜λŠ” λͺ¨λ“ˆμ‹ νˆ΄λ°•μŠ€μž…λ‹ˆλ‹€. 저희 λΌμ΄λΈŒλŸ¬λ¦¬λŠ” [μ„±λŠ₯보닀 μ‚¬μš©μ„±](conceptual/philosophy#usability-over-performance), [κ°„νŽΈν•¨λ³΄λ‹€ λ‹¨μˆœν•¨](conceptual/philosophy#simple-over-easy), 그리고 [좔상화보닀 μ‚¬μš©μž μ§€μ • κ°€λŠ₯μ„±](conceptual/philosophy#tweakable-contributorfriendly-over-abstraction)에 쀑점을 두고 μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€. + +이 λΌμ΄λΈŒλŸ¬λ¦¬μ—λŠ” μ„Έ κ°€μ§€ μ£Όμš” ꡬ성 μš”μ†Œκ°€ μžˆμŠ΅λ‹ˆλ‹€: + +- λͺ‡ μ€„μ˜ μ½”λ“œλ§ŒμœΌλ‘œ μΆ”λ‘ ν•  수 μžˆλŠ” μ΅œμ²¨λ‹¨ [diffusion νŒŒμ΄ν”„λΌμΈ](api/pipelines/overview). +- 생성 속도와 ν’ˆμ§ˆ κ°„μ˜ κ· ν˜•μ„ λ§žμΆ”κΈ° μœ„ν•΄ μƒν˜Έκ΅ν™˜μ μœΌλ‘œ μ‚¬μš©ν•  수 μžˆλŠ” [λ…Έμ΄μ¦ˆ μŠ€μΌ€μ€„λŸ¬](api/schedulers/overview). +- λΉŒλ”© λΈ”λ‘μœΌλ‘œ μ‚¬μš©ν•  수 있고 μŠ€μΌ€μ€„λŸ¬μ™€ κ²°ν•©ν•˜μ—¬ 자체적인 end-to-end diffusion μ‹œμŠ€ν…œμ„ λ§Œλ“€ 수 μžˆλŠ” 사전 ν•™μŠ΅λœ [λͺ¨λΈ](api/models). + +
+
+
Tutorials
+

결과물을 μƒμ„±ν•˜κ³ , λ‚˜λ§Œμ˜ diffusion μ‹œμŠ€ν…œμ„ κ΅¬μΆ•ν•˜κ³ , ν™•μ‚° λͺ¨λΈμ„ ν›ˆλ ¨ν•˜λŠ” 데 ν•„μš”ν•œ κΈ°λ³Έ κΈ°μˆ μ„ λ°°μ›Œλ³΄μ„Έμš”. πŸ€— Diffusersλ₯Ό 처음 μ‚¬μš©ν•˜λŠ” 경우 μ—¬κΈ°μ—μ„œ μ‹œμž‘ν•˜λŠ” 것이 μ’‹μŠ΅λ‹ˆλ‹€!

+
+
How-to guides
+

νŒŒμ΄ν”„λΌμΈ, λͺ¨λΈ, μŠ€μΌ€μ€„λŸ¬λ₯Ό λ‘œλ“œν•˜λŠ” 데 도움이 λ˜λŠ” μ‹€μš©μ μΈ κ°€μ΄λ“œμž…λ‹ˆλ‹€. λ˜ν•œ νŠΉμ • μž‘μ—…μ— νŒŒμ΄ν”„λΌμΈμ„ μ‚¬μš©ν•˜κ³ , 좜λ ₯ 생성 방식을 μ œμ–΄ν•˜κ³ , μΆ”λ‘  속도에 맞게 μ΅œμ ν™”ν•˜κ³ , λ‹€μ–‘ν•œ ν•™μŠ΅ 기법을 μ‚¬μš©ν•˜λŠ” 방법도 배울 수 μžˆμŠ΅λ‹ˆλ‹€.

+
+
Conceptual guides
+

λΌμ΄λΈŒλŸ¬λ¦¬κ°€ μ™œ 이런 λ°©μ‹μœΌλ‘œ μ„€κ³„λ˜μ—ˆλŠ”μ§€ μ΄ν•΄ν•˜κ³ , 라이브러리 μ΄μš©μ— λŒ€ν•œ 윀리적 κ°€μ΄λ“œλΌμΈκ³Ό μ•ˆμ „ κ΅¬ν˜„μ— λŒ€ν•΄ μžμ„Ένžˆ μ•Œμ•„λ³΄μ„Έμš”.

+
+
Reference
+

πŸ€— Diffusers 클래슀 및 λ©”μ„œλ“œμ˜ μž‘λ™ 방식에 λŒ€ν•œ 기술 μ„€λͺ….

+
+
+
+ +## Supported pipelines + +| Pipeline | Paper/Repository | Tasks | +|---|---|:---:| +| [alt_diffusion](./api/pipelines/alt_diffusion) | [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) | Image-to-Image Text-Guided Generation | +| [audio_diffusion](./api/pipelines/audio_diffusion) | [Audio Diffusion](https://github.com/teticio/audio-diffusion.git) | Unconditional Audio Generation | +| [controlnet](./api/pipelines/stable_diffusion/controlnet) | [Adding Conditional Control to Text-to-Image Diffusion Models](https://arxiv.org/abs/2302.05543) | Image-to-Image Text-Guided Generation | +| [cycle_diffusion](./api/pipelines/cycle_diffusion) | [Unifying Diffusion Models' Latent Space, with Applications to CycleDiffusion and Guidance](https://arxiv.org/abs/2210.05559) | Image-to-Image Text-Guided Generation | +| [dance_diffusion](./api/pipelines/dance_diffusion) | [Dance Diffusion](https://github.com/williamberman/diffusers.git) | Unconditional Audio Generation | +| [ddpm](./api/pipelines/ddpm) | [Denoising Diffusion Probabilistic Models](https://arxiv.org/abs/2006.11239) | Unconditional Image Generation | +| [ddim](./api/pipelines/ddim) | [Denoising Diffusion Implicit Models](https://arxiv.org/abs/2010.02502) | Unconditional Image Generation | +| [if](./if) | [**IF**](./api/pipelines/if) | Image Generation | +| [if_img2img](./if) | [**IF**](./api/pipelines/if) | Image-to-Image Generation | +| [if_inpainting](./if) | [**IF**](./api/pipelines/if) | Image-to-Image Generation | +| [latent_diffusion](./api/pipelines/latent_diffusion) | [High-Resolution Image Synthesis with Latent Diffusion Models](https://arxiv.org/abs/2112.10752)| Text-to-Image Generation | +| [latent_diffusion](./api/pipelines/latent_diffusion) | [High-Resolution Image Synthesis with Latent Diffusion Models](https://arxiv.org/abs/2112.10752)| Super Resolution Image-to-Image | +| [latent_diffusion_uncond](./api/pipelines/latent_diffusion_uncond) | [High-Resolution Image Synthesis with Latent Diffusion Models](https://arxiv.org/abs/2112.10752) | Unconditional Image Generation | +| [paint_by_example](./api/pipelines/paint_by_example) | [Paint by Example: Exemplar-based Image Editing with Diffusion Models](https://arxiv.org/abs/2211.13227) | Image-Guided Image Inpainting | +| [pndm](./api/pipelines/pndm) | [Pseudo Numerical Methods for Diffusion Models on Manifolds](https://arxiv.org/abs/2202.09778) | Unconditional Image Generation | +| [score_sde_ve](./api/pipelines/score_sde_ve) | [Score-Based Generative Modeling through Stochastic Differential Equations](https://openreview.net/forum?id=PxTIG12RRHS) | Unconditional Image Generation | +| [score_sde_vp](./api/pipelines/score_sde_vp) | [Score-Based Generative Modeling through Stochastic Differential Equations](https://openreview.net/forum?id=PxTIG12RRHS) | Unconditional Image Generation | +| [semantic_stable_diffusion](./api/pipelines/semantic_stable_diffusion) | [Semantic Guidance](https://arxiv.org/abs/2301.12247) | Text-Guided Generation | +| [stable_diffusion_text2img](./api/pipelines/stable_diffusion/text2img) | [Stable Diffusion](https://stability.ai/blog/stable-diffusion-public-release) | Text-to-Image Generation | +| [stable_diffusion_img2img](./api/pipelines/stable_diffusion/img2img) | [Stable Diffusion](https://stability.ai/blog/stable-diffusion-public-release) | Image-to-Image Text-Guided Generation | +| [stable_diffusion_inpaint](./api/pipelines/stable_diffusion/inpaint) | [Stable Diffusion](https://stability.ai/blog/stable-diffusion-public-release) | Text-Guided Image Inpainting | +| [stable_diffusion_panorama](./api/pipelines/stable_diffusion/panorama) | [MultiDiffusion](https://multidiffusion.github.io/) | Text-to-Panorama Generation | +| [stable_diffusion_pix2pix](./api/pipelines/stable_diffusion/pix2pix) | [InstructPix2Pix: Learning to Follow Image Editing Instructions](https://arxiv.org/abs/2211.09800) | Text-Guided Image Editing| +| [stable_diffusion_pix2pix_zero](./api/pipelines/stable_diffusion/pix2pix_zero) | [Zero-shot Image-to-Image Translation](https://pix2pixzero.github.io/) | Text-Guided Image Editing | +| [stable_diffusion_attend_and_excite](./api/pipelines/stable_diffusion/attend_and_excite) | [Attend-and-Excite: Attention-Based Semantic Guidance for Text-to-Image Diffusion Models](https://arxiv.org/abs/2301.13826) | Text-to-Image Generation | +| [stable_diffusion_self_attention_guidance](./api/pipelines/stable_diffusion/self_attention_guidance) | [Improving Sample Quality of Diffusion Models Using Self-Attention Guidance](https://arxiv.org/abs/2210.00939) | Text-to-Image Generation Unconditional Image Generation | +| [stable_diffusion_image_variation](./stable_diffusion/image_variation) | [Stable Diffusion Image Variations](https://github.com/LambdaLabsML/lambda-diffusers#stable-diffusion-image-variations) | Image-to-Image Generation | +| [stable_diffusion_latent_upscale](./stable_diffusion/latent_upscale) | [Stable Diffusion Latent Upscaler](https://twitter.com/StabilityAI/status/1590531958815064065) | Text-Guided Super Resolution Image-to-Image | +| [stable_diffusion_model_editing](./api/pipelines/stable_diffusion/model_editing) | [Editing Implicit Assumptions in Text-to-Image Diffusion Models](https://time-diffusion.github.io/) | Text-to-Image Model Editing | +| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [Stable Diffusion 2](https://stability.ai/blog/stable-diffusion-v2-release) | Text-to-Image Generation | +| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [Stable Diffusion 2](https://stability.ai/blog/stable-diffusion-v2-release) | Text-Guided Image Inpainting | +| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [Depth-Conditional Stable Diffusion](https://github.com/Stability-AI/stablediffusion#depth-conditional-stable-diffusion) | Depth-to-Image Generation | +| [stable_diffusion_2](./api/pipelines/stable_diffusion_2) | [Stable Diffusion 2](https://stability.ai/blog/stable-diffusion-v2-release) | Text-Guided Super Resolution Image-to-Image | +| [stable_diffusion_safe](./api/pipelines/stable_diffusion_safe) | [Safe Stable Diffusion](https://arxiv.org/abs/2211.05105) | Text-Guided Generation | +| [stable_unclip](./stable_unclip) | Stable unCLIP | Text-to-Image Generation | +| [stable_unclip](./stable_unclip) | Stable unCLIP | Image-to-Image Text-Guided Generation | +| [stochastic_karras_ve](./api/pipelines/stochastic_karras_ve) | [Elucidating the Design Space of Diffusion-Based Generative Models](https://arxiv.org/abs/2206.00364) | Unconditional Image Generation | +| [text_to_video_sd](./api/pipelines/text_to_video) | [Modelscope's Text-to-video-synthesis Model in Open Domain](https://modelscope.cn/models/damo/text-to-video-synthesis/summary) | Text-to-Video Generation | +| [unclip](./api/pipelines/unclip) | [Hierarchical Text-Conditional Image Generation with CLIP Latents](https://arxiv.org/abs/2204.06125)(implementation by [kakaobrain](https://github.com/kakaobrain/karlo)) | Text-to-Image Generation | +| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Text-to-Image Generation | +| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Image Variations Generation | +| [versatile_diffusion](./api/pipelines/versatile_diffusion) | [Versatile Diffusion: Text, Images and Variations All in One Diffusion Model](https://arxiv.org/abs/2211.08332) | Dual Image and Text Guided Generation | +| [vq_diffusion](./api/pipelines/vq_diffusion) | [Vector Quantized Diffusion Model for Text-to-Image Synthesis](https://arxiv.org/abs/2111.14822) | Text-to-Image Generation | diff --git a/docs/source/ko/using-diffusers/conditional_image_generation.mdx b/docs/source/ko/using-diffusers/conditional_image_generation.mdx new file mode 100644 index 000000000000..5525ac990ca4 --- /dev/null +++ b/docs/source/ko/using-diffusers/conditional_image_generation.mdx @@ -0,0 +1,60 @@ + + +# 쑰건뢀 이미지 생성 + +[[open-in-colab]] + +쑰건뢀 이미지 생성을 μ‚¬μš©ν•˜λ©΄ ν…μŠ€νŠΈ ν”„λ‘¬ν”„νŠΈμ—μ„œ 이미지λ₯Ό 생성할 수 μžˆμŠ΅λ‹ˆλ‹€. ν…μŠ€νŠΈλŠ” μž„λ² λ”©μœΌλ‘œ λ³€ν™˜λ˜λ©°, μž„λ² λ”©μ€ λ…Έμ΄μ¦ˆμ—μ„œ 이미지λ₯Ό μƒμ„±ν•˜λ„λ‘ λͺ¨λΈμ„ μ‘°κ±΄ν™”ν•˜λŠ” 데 μ‚¬μš©λ©λ‹ˆλ‹€. + +[`DiffusionPipeline`]은 좔둠을 μœ„ν•΄ 사전 ν›ˆλ ¨λœ diffusion μ‹œμŠ€ν…œμ„ μ‚¬μš©ν•˜λŠ” κ°€μž₯ μ‰¬μš΄ λ°©λ²•μž…λ‹ˆλ‹€. + +λ¨Όμ € [`DiffusionPipeline`]의 μΈμŠ€ν„΄μŠ€λ₯Ό μƒμ„±ν•˜κ³  λ‹€μš΄λ‘œλ“œν•  νŒŒμ΄ν”„λΌμΈ [체크포인트](https://huggingface.co/models?library=diffusers&sort=downloads)λ₯Ό μ§€μ •ν•©λ‹ˆλ‹€. + +이 κ°€μ΄λ“œμ—μ„œλŠ” [잠재 Diffusion](https://huggingface.co/CompVis/ldm-text2im-large-256)κ³Ό ν•¨κ»˜ ν…μŠ€νŠΈ-이미지 생성에 [`DiffusionPipeline`]을 μ‚¬μš©ν•©λ‹ˆλ‹€: + +```python +>>> from diffusers import DiffusionPipeline + +>>> generator = DiffusionPipeline.from_pretrained("CompVis/ldm-text2im-large-256") +``` + +[`DiffusionPipeline`]은 λͺ¨λ“  λͺ¨λΈλ§, 토큰화, μŠ€μΌ€μ€„λ§ ꡬ성 μš”μ†Œλ₯Ό λ‹€μš΄λ‘œλ“œν•˜κ³  μΊμ‹œν•©λ‹ˆλ‹€. +이 λͺ¨λΈμ€ μ•½ 14μ–΅ 개의 νŒŒλΌλ―Έν„°λ‘œ κ΅¬μ„±λ˜μ–΄ 있기 λ•Œλ¬Έμ— GPUμ—μ„œ μ‹€ν–‰ν•  것을 κ°•λ ₯히 ꢌμž₯ν•©λ‹ˆλ‹€. +PyTorchμ—μ„œμ™€ λ§ˆμ°¬κ°€μ§€λ‘œ 생성기 객체λ₯Ό GPU둜 이동할 수 μžˆμŠ΅λ‹ˆλ‹€: + +```python +>>> generator.to("cuda") +``` + +이제 ν…μŠ€νŠΈ ν”„λ‘¬ν”„νŠΈμ—μ„œ `생성기`λ₯Ό μ‚¬μš©ν•  수 μžˆμŠ΅λ‹ˆλ‹€: + +```python +>>> image = generator("An image of a squirrel in Picasso style").images[0] +``` + +좜λ ₯값은 기본적으둜 [`PIL.Image`](https://pillow.readthedocs.io/en/stable/reference/Image.html?highlight=image#the-image-class) 객체둜 λž˜ν•‘λ©λ‹ˆλ‹€. + +ν˜ΈμΆœν•˜μ—¬ 이미지λ₯Ό μ €μž₯ν•  수 μžˆμŠ΅λ‹ˆλ‹€: + +```python +>>> image.save("image_of_squirrel_painting.png") +``` + +μ•„λž˜ 슀페이슀λ₯Ό μ‚¬μš©ν•΄λ³΄κ³  μ•ˆλ‚΄ 배율 λ§€κ°œλ³€μˆ˜λ₯Ό 자유둭게 μ‘°μ •ν•˜μ—¬ 이미지 ν’ˆμ§ˆμ— μ–΄λ–€ 영ν–₯을 λ―ΈμΉ˜λŠ”μ§€ 확인해 λ³΄μ„Έμš”! + + \ No newline at end of file diff --git a/docs/source/ko/using-diffusers/stable_diffusion_jax_how_to.mdx b/docs/source/ko/using-diffusers/stable_diffusion_jax_how_to.mdx index e5785374413c..ef2da6bdf902 100644 --- a/docs/source/ko/using-diffusers/stable_diffusion_jax_how_to.mdx +++ b/docs/source/ko/using-diffusers/stable_diffusion_jax_how_to.mdx @@ -8,7 +8,6 @@ http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ---> # JAX / Flaxμ—μ„œμ˜ 🧨 Stable Diffusion!