orion-dev-leon
		
	commited on
		
		
					Commit 
							
							·
						
						2c584d9
	
1
								Parent(s):
							
							2663371
								
update
Browse files- README.md +56 -0
- densepose/model_final_162be9.pkl +3 -0
- gitattributes +37 -0
- humanparsing/parsing_atr.onnx +3 -0
- humanparsing/parsing_lip.onnx +3 -0
- image_encoder/config.json +23 -0
- image_encoder/model.safetensors +3 -0
- model_index.json +39 -0
- openpose/ckpts/body_pose_model.pth +3 -0
- scheduler/scheduler_config.json +19 -0
- text_encoder/config.json +25 -0
- text_encoder/model.safetensors +3 -0
- text_encoder_2/config.json +25 -0
- text_encoder_2/model.safetensors +3 -0
- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +24 -0
- tokenizer/tokenizer_config.json +33 -0
- tokenizer/vocab.json +0 -0
- tokenizer_2/merges.txt +0 -0
- tokenizer_2/special_tokens_map.json +24 -0
- tokenizer_2/tokenizer_config.json +33 -0
- tokenizer_2/vocab.json +0 -0
- unet/config.json +78 -0
- unet/diffusion_pytorch_model.bin +3 -0
- unet_encoder/config.json +68 -0
- unet_encoder/diffusion_pytorch_model.safetensors +3 -0
- vae/config.json +32 -0
- vae/diffusion_pytorch_model.safetensors +3 -0
    	
        README.md
    ADDED
    
    | @@ -0,0 +1,56 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            ---
         | 
| 2 | 
            +
            base_model: stable-diffusion-xl-1.0-inpainting-0.1
         | 
| 3 | 
            +
            tags:
         | 
| 4 | 
            +
              - stable-diffusion-xl
         | 
| 5 | 
            +
              - inpainting
         | 
| 6 | 
            +
              - virtual try-on
         | 
| 7 | 
            +
            license: cc-by-nc-sa-4.0
         | 
| 8 | 
            +
            ---
         | 
| 9 | 
            +
             | 
| 10 | 
            +
             | 
| 11 | 
            +
             | 
| 12 | 
            +
            # Check out more codes on our [github repository](https://github.com/yisol/IDM-VTON)!
         | 
| 13 | 
            +
             | 
| 14 | 
            +
            # IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild
         | 
| 15 | 
            +
            This is an official implementation of paper 'Improving Diffusion Models for Authentic Virtual Try-on in the Wild'
         | 
| 16 | 
            +
            - [paper](https://arxiv.org/abs/2403.05139) 
         | 
| 17 | 
            +
            - [project page](https://idm-vton.github.io/) 
         | 
| 18 | 
            +
             | 
| 19 | 
            +
            🤗 Try our huggingface [Demo](https://huggingface.co/spaces/yisol/IDM-VTON)
         | 
| 20 | 
            +
             | 
| 21 | 
            +
             
         | 
| 22 | 
            +
             
         | 
| 23 | 
            +
             | 
| 24 | 
            +
             | 
| 25 | 
            +
            ## TODO LIST
         | 
| 26 | 
            +
             | 
| 27 | 
            +
             | 
| 28 | 
            +
            - [x] demo model
         | 
| 29 | 
            +
            - [x] inference code
         | 
| 30 | 
            +
            - [ ] training code
         | 
| 31 | 
            +
             | 
| 32 | 
            +
             | 
| 33 | 
            +
             | 
| 34 | 
            +
             | 
| 35 | 
            +
            ## Acknowledgements
         | 
| 36 | 
            +
             | 
| 37 | 
            +
            For the demo, GPUs are supported from [zerogpu](https://huggingface.co/zero-gpu-explorers), and auto masking generation codes are based on [OOTDiffusion](https://github.com/levihsu/OOTDiffusion) and [DCI-VTON](https://github.com/bcmi/DCI-VTON-Virtual-Try-On).  
         | 
| 38 | 
            +
            Parts of the code are based on [IP-Adapter](https://github.com/tencent-ailab/IP-Adapter).
         | 
| 39 | 
            +
             | 
| 40 | 
            +
             | 
| 41 | 
            +
             | 
| 42 | 
            +
            ## Citation
         | 
| 43 | 
            +
            ```
         | 
| 44 | 
            +
            @article{choi2024improving,
         | 
| 45 | 
            +
              title={Improving Diffusion Models for Virtual Try-on},
         | 
| 46 | 
            +
              author={Choi, Yisol and Kwak, Sangkyung and Lee, Kyungmin and Choi, Hyungwon and Shin, Jinwoo},
         | 
| 47 | 
            +
              journal={arXiv preprint arXiv:2403.05139},
         | 
| 48 | 
            +
              year={2024}
         | 
| 49 | 
            +
            }
         | 
| 50 | 
            +
            ```
         | 
| 51 | 
            +
             | 
| 52 | 
            +
            ## License
         | 
| 53 | 
            +
            The codes and checkpoints in this repository are under the [CC BY-NC-SA 4.0 license](https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode).
         | 
| 54 | 
            +
             | 
| 55 | 
            +
             | 
| 56 | 
            +
             | 
    	
        densepose/model_final_162be9.pkl
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91
         | 
| 3 | 
            +
            size 255757821
         | 
    	
        gitattributes
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            *.7z filter=lfs diff=lfs merge=lfs -text
         | 
| 2 | 
            +
            *.arrow filter=lfs diff=lfs merge=lfs -text
         | 
| 3 | 
            +
            *.bin filter=lfs diff=lfs merge=lfs -text
         | 
| 4 | 
            +
            *.bz2 filter=lfs diff=lfs merge=lfs -text
         | 
| 5 | 
            +
            *.ckpt filter=lfs diff=lfs merge=lfs -text
         | 
| 6 | 
            +
            *.ftz filter=lfs diff=lfs merge=lfs -text
         | 
| 7 | 
            +
            *.gz filter=lfs diff=lfs merge=lfs -text
         | 
| 8 | 
            +
            *.h5 filter=lfs diff=lfs merge=lfs -text
         | 
| 9 | 
            +
            *.joblib filter=lfs diff=lfs merge=lfs -text
         | 
| 10 | 
            +
            *.lfs.* filter=lfs diff=lfs merge=lfs -text
         | 
| 11 | 
            +
            *.mlmodel filter=lfs diff=lfs merge=lfs -text
         | 
| 12 | 
            +
            *.model filter=lfs diff=lfs merge=lfs -text
         | 
| 13 | 
            +
            *.msgpack filter=lfs diff=lfs merge=lfs -text
         | 
| 14 | 
            +
            *.npy filter=lfs diff=lfs merge=lfs -text
         | 
| 15 | 
            +
            *.npz filter=lfs diff=lfs merge=lfs -text
         | 
| 16 | 
            +
            *.onnx filter=lfs diff=lfs merge=lfs -text
         | 
| 17 | 
            +
            *.ot filter=lfs diff=lfs merge=lfs -text
         | 
| 18 | 
            +
            *.parquet filter=lfs diff=lfs merge=lfs -text
         | 
| 19 | 
            +
            *.pb filter=lfs diff=lfs merge=lfs -text
         | 
| 20 | 
            +
            *.pickle filter=lfs diff=lfs merge=lfs -text
         | 
| 21 | 
            +
            *.pkl filter=lfs diff=lfs merge=lfs -text
         | 
| 22 | 
            +
            *.pt filter=lfs diff=lfs merge=lfs -text
         | 
| 23 | 
            +
            *.pth filter=lfs diff=lfs merge=lfs -text
         | 
| 24 | 
            +
            *.rar filter=lfs diff=lfs merge=lfs -text
         | 
| 25 | 
            +
            *.safetensors filter=lfs diff=lfs merge=lfs -text
         | 
| 26 | 
            +
            saved_model/**/* filter=lfs diff=lfs merge=lfs -text
         | 
| 27 | 
            +
            *.tar.* filter=lfs diff=lfs merge=lfs -text
         | 
| 28 | 
            +
            *.tar filter=lfs diff=lfs merge=lfs -text
         | 
| 29 | 
            +
            *.tflite filter=lfs diff=lfs merge=lfs -text
         | 
| 30 | 
            +
            *.tgz filter=lfs diff=lfs merge=lfs -text
         | 
| 31 | 
            +
            *.wasm filter=lfs diff=lfs merge=lfs -text
         | 
| 32 | 
            +
            *.xz filter=lfs diff=lfs merge=lfs -text
         | 
| 33 | 
            +
            *.zip filter=lfs diff=lfs merge=lfs -text
         | 
| 34 | 
            +
            *.zst filter=lfs diff=lfs merge=lfs -text
         | 
| 35 | 
            +
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         | 
| 36 | 
            +
            assets/teaser.png filter=lfs diff=lfs merge=lfs -text
         | 
| 37 | 
            +
            assets/teaser2.png filter=lfs diff=lfs merge=lfs -text
         | 
    	
        humanparsing/parsing_atr.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:04c7d1d070d0e0ae943d86b18cb5aaaea9e278d97462e9cfb270cbbe4cd977f4
         | 
| 3 | 
            +
            size 266859305
         | 
    	
        humanparsing/parsing_lip.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8436e1dae96e2601c373d1ace29c8f0978b16357d9038c17a8ba756cca376dbc
         | 
| 3 | 
            +
            size 266863411
         | 
    	
        image_encoder/config.json
    ADDED
    
    | @@ -0,0 +1,23 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "./image_encoder",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "CLIPVisionModelWithProjection"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_dropout": 0.0,
         | 
| 7 | 
            +
              "dropout": 0.0,
         | 
| 8 | 
            +
              "hidden_act": "gelu",
         | 
| 9 | 
            +
              "hidden_size": 1280,
         | 
| 10 | 
            +
              "image_size": 224,
         | 
| 11 | 
            +
              "initializer_factor": 1.0,
         | 
| 12 | 
            +
              "initializer_range": 0.02,
         | 
| 13 | 
            +
              "intermediate_size": 5120,
         | 
| 14 | 
            +
              "layer_norm_eps": 1e-05,
         | 
| 15 | 
            +
              "model_type": "clip_vision_model",
         | 
| 16 | 
            +
              "num_attention_heads": 16,
         | 
| 17 | 
            +
              "num_channels": 3,
         | 
| 18 | 
            +
              "num_hidden_layers": 32,
         | 
| 19 | 
            +
              "patch_size": 14,
         | 
| 20 | 
            +
              "projection_dim": 1024,
         | 
| 21 | 
            +
              "torch_dtype": "float16",
         | 
| 22 | 
            +
              "transformers_version": "4.28.0.dev0"
         | 
| 23 | 
            +
            }
         | 
    	
        image_encoder/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:6ca9667da1ca9e0b0f75e46bb030f7e011f44f86cbfb8d5a36590fcd7507b030
         | 
| 3 | 
            +
            size 2528373448
         | 
    	
        model_index.json
    ADDED
    
    | @@ -0,0 +1,39 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "StableDiffusionXLInpaintPipeline",
         | 
| 3 | 
            +
              "_diffusers_version": "0.25.0",
         | 
| 4 | 
            +
              "_name_or_path": "stabilityai/stable-diffusion-xl-base-1.0",
         | 
| 5 | 
            +
              "force_zeros_for_empty_prompt": true,
         | 
| 6 | 
            +
              "requires_aesthetics_score": false,
         | 
| 7 | 
            +
              "scheduler": [
         | 
| 8 | 
            +
                "diffusers",
         | 
| 9 | 
            +
                "DDPMScheduler"
         | 
| 10 | 
            +
              ],
         | 
| 11 | 
            +
              "text_encoder": [
         | 
| 12 | 
            +
                "transformers",
         | 
| 13 | 
            +
                "CLIPTextModel"
         | 
| 14 | 
            +
              ],
         | 
| 15 | 
            +
              "feature_extractor": [
         | 
| 16 | 
            +
                "transformers",
         | 
| 17 | 
            +
                "CLIPImageProcessor"
         | 
| 18 | 
            +
              ],
         | 
| 19 | 
            +
              "text_encoder_2": [
         | 
| 20 | 
            +
                "transformers",
         | 
| 21 | 
            +
                "CLIPTextModelWithProjection"
         | 
| 22 | 
            +
              ],
         | 
| 23 | 
            +
              "tokenizer": [
         | 
| 24 | 
            +
                "transformers",
         | 
| 25 | 
            +
                "CLIPTokenizer"
         | 
| 26 | 
            +
              ],
         | 
| 27 | 
            +
              "tokenizer_2": [
         | 
| 28 | 
            +
                "transformers",
         | 
| 29 | 
            +
                "CLIPTokenizer"
         | 
| 30 | 
            +
              ],
         | 
| 31 | 
            +
              "unet": [
         | 
| 32 | 
            +
                "diffusers",
         | 
| 33 | 
            +
                "UNet2DConditionModel"
         | 
| 34 | 
            +
              ],
         | 
| 35 | 
            +
              "vae": [
         | 
| 36 | 
            +
                "diffusers",
         | 
| 37 | 
            +
                "AutoencoderKL"
         | 
| 38 | 
            +
              ]
         | 
| 39 | 
            +
            }
         | 
    	
        openpose/ckpts/body_pose_model.pth
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:25a948c16078b0f08e236bda51a385d855ef4c153598947c28c0d47ed94bb746
         | 
| 3 | 
            +
            size 209267595
         | 
    	
        scheduler/scheduler_config.json
    ADDED
    
    | @@ -0,0 +1,19 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "DDPMScheduler",
         | 
| 3 | 
            +
              "_diffusers_version": "0.21.0.dev0",
         | 
| 4 | 
            +
              "beta_end": 0.012,
         | 
| 5 | 
            +
              "beta_schedule": "scaled_linear",
         | 
| 6 | 
            +
              "beta_start": 0.00085,
         | 
| 7 | 
            +
              "clip_sample": false,
         | 
| 8 | 
            +
              "interpolation_type": "linear",
         | 
| 9 | 
            +
              "num_train_timesteps": 1000,
         | 
| 10 | 
            +
              "prediction_type": "epsilon",
         | 
| 11 | 
            +
              "sample_max_value": 1.0,
         | 
| 12 | 
            +
              "set_alpha_to_one": false,
         | 
| 13 | 
            +
              "skip_prk_steps": true,
         | 
| 14 | 
            +
              "steps_offset": 1,
         | 
| 15 | 
            +
              "timestep_spacing": "leading",
         | 
| 16 | 
            +
              "trained_betas": null,
         | 
| 17 | 
            +
              "use_karras_sigmas": false,
         | 
| 18 | 
            +
              "rescale_betas_zero_snr": true
         | 
| 19 | 
            +
            }
         | 
    	
        text_encoder/config.json
    ADDED
    
    | @@ -0,0 +1,25 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "/home/suraj_huggingface_co/.cache/huggingface/hub/models--stabilityai--stable-diffusion-xl-base-1.0/snapshots/bf714989e22c57ddc1c453bf74dab4521acb81d8/text_encoder",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "CLIPTextModel"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_dropout": 0.0,
         | 
| 7 | 
            +
              "bos_token_id": 0,
         | 
| 8 | 
            +
              "dropout": 0.0,
         | 
| 9 | 
            +
              "eos_token_id": 2,
         | 
| 10 | 
            +
              "hidden_act": "quick_gelu",
         | 
| 11 | 
            +
              "hidden_size": 768,
         | 
| 12 | 
            +
              "initializer_factor": 1.0,
         | 
| 13 | 
            +
              "initializer_range": 0.02,
         | 
| 14 | 
            +
              "intermediate_size": 3072,
         | 
| 15 | 
            +
              "layer_norm_eps": 1e-05,
         | 
| 16 | 
            +
              "max_position_embeddings": 77,
         | 
| 17 | 
            +
              "model_type": "clip_text_model",
         | 
| 18 | 
            +
              "num_attention_heads": 12,
         | 
| 19 | 
            +
              "num_hidden_layers": 12,
         | 
| 20 | 
            +
              "pad_token_id": 1,
         | 
| 21 | 
            +
              "projection_dim": 768,
         | 
| 22 | 
            +
              "torch_dtype": "float16",
         | 
| 23 | 
            +
              "transformers_version": "4.29.2",
         | 
| 24 | 
            +
              "vocab_size": 49408
         | 
| 25 | 
            +
            }
         | 
    	
        text_encoder/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:79f531155c765c22c89e23328793a2e91a1178070af961c57e2eae5f0509b65b
         | 
| 3 | 
            +
            size 492265879
         | 
    	
        text_encoder_2/config.json
    ADDED
    
    | @@ -0,0 +1,25 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "/home/suraj_huggingface_co/.cache/huggingface/hub/models--stabilityai--stable-diffusion-xl-base-1.0/snapshots/bf714989e22c57ddc1c453bf74dab4521acb81d8/text_encoder_2",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "CLIPTextModelWithProjection"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_dropout": 0.0,
         | 
| 7 | 
            +
              "bos_token_id": 0,
         | 
| 8 | 
            +
              "dropout": 0.0,
         | 
| 9 | 
            +
              "eos_token_id": 2,
         | 
| 10 | 
            +
              "hidden_act": "gelu",
         | 
| 11 | 
            +
              "hidden_size": 1280,
         | 
| 12 | 
            +
              "initializer_factor": 1.0,
         | 
| 13 | 
            +
              "initializer_range": 0.02,
         | 
| 14 | 
            +
              "intermediate_size": 5120,
         | 
| 15 | 
            +
              "layer_norm_eps": 1e-05,
         | 
| 16 | 
            +
              "max_position_embeddings": 77,
         | 
| 17 | 
            +
              "model_type": "clip_text_model",
         | 
| 18 | 
            +
              "num_attention_heads": 20,
         | 
| 19 | 
            +
              "num_hidden_layers": 32,
         | 
| 20 | 
            +
              "pad_token_id": 1,
         | 
| 21 | 
            +
              "projection_dim": 1280,
         | 
| 22 | 
            +
              "torch_dtype": "float16",
         | 
| 23 | 
            +
              "transformers_version": "4.29.2",
         | 
| 24 | 
            +
              "vocab_size": 49408
         | 
| 25 | 
            +
            }
         | 
    	
        text_encoder_2/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:283bb90f987a133dec11947571aca17692ed32f3fff708441ac8eedcfa4a040e
         | 
| 3 | 
            +
            size 2778702976
         | 
    	
        tokenizer/merges.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        tokenizer/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,24 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "bos_token": {
         | 
| 3 | 
            +
                "content": "<|startoftext|>",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": true,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "eos_token": {
         | 
| 10 | 
            +
                "content": "<|endoftext|>",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": true,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 17 | 
            +
              "unk_token": {
         | 
| 18 | 
            +
                "content": "<|endoftext|>",
         | 
| 19 | 
            +
                "lstrip": false,
         | 
| 20 | 
            +
                "normalized": true,
         | 
| 21 | 
            +
                "rstrip": false,
         | 
| 22 | 
            +
                "single_word": false
         | 
| 23 | 
            +
              }
         | 
| 24 | 
            +
            }
         | 
    	
        tokenizer/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,33 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "add_prefix_space": false,
         | 
| 3 | 
            +
              "bos_token": {
         | 
| 4 | 
            +
                "__type": "AddedToken",
         | 
| 5 | 
            +
                "content": "<|startoftext|>",
         | 
| 6 | 
            +
                "lstrip": false,
         | 
| 7 | 
            +
                "normalized": true,
         | 
| 8 | 
            +
                "rstrip": false,
         | 
| 9 | 
            +
                "single_word": false
         | 
| 10 | 
            +
              },
         | 
| 11 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 12 | 
            +
              "do_lower_case": true,
         | 
| 13 | 
            +
              "eos_token": {
         | 
| 14 | 
            +
                "__type": "AddedToken",
         | 
| 15 | 
            +
                "content": "<|endoftext|>",
         | 
| 16 | 
            +
                "lstrip": false,
         | 
| 17 | 
            +
                "normalized": true,
         | 
| 18 | 
            +
                "rstrip": false,
         | 
| 19 | 
            +
                "single_word": false
         | 
| 20 | 
            +
              },
         | 
| 21 | 
            +
              "errors": "replace",
         | 
| 22 | 
            +
              "model_max_length": 77,
         | 
| 23 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 24 | 
            +
              "tokenizer_class": "CLIPTokenizer",
         | 
| 25 | 
            +
              "unk_token": {
         | 
| 26 | 
            +
                "__type": "AddedToken",
         | 
| 27 | 
            +
                "content": "<|endoftext|>",
         | 
| 28 | 
            +
                "lstrip": false,
         | 
| 29 | 
            +
                "normalized": true,
         | 
| 30 | 
            +
                "rstrip": false,
         | 
| 31 | 
            +
                "single_word": false
         | 
| 32 | 
            +
              }
         | 
| 33 | 
            +
            }
         | 
    	
        tokenizer/vocab.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        tokenizer_2/merges.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        tokenizer_2/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,24 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "bos_token": {
         | 
| 3 | 
            +
                "content": "<|startoftext|>",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": true,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "eos_token": {
         | 
| 10 | 
            +
                "content": "<|endoftext|>",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": true,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": "!",
         | 
| 17 | 
            +
              "unk_token": {
         | 
| 18 | 
            +
                "content": "<|endoftext|>",
         | 
| 19 | 
            +
                "lstrip": false,
         | 
| 20 | 
            +
                "normalized": true,
         | 
| 21 | 
            +
                "rstrip": false,
         | 
| 22 | 
            +
                "single_word": false
         | 
| 23 | 
            +
              }
         | 
| 24 | 
            +
            }
         | 
    	
        tokenizer_2/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,33 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "add_prefix_space": false,
         | 
| 3 | 
            +
              "bos_token": {
         | 
| 4 | 
            +
                "__type": "AddedToken",
         | 
| 5 | 
            +
                "content": "<|startoftext|>",
         | 
| 6 | 
            +
                "lstrip": false,
         | 
| 7 | 
            +
                "normalized": true,
         | 
| 8 | 
            +
                "rstrip": false,
         | 
| 9 | 
            +
                "single_word": false
         | 
| 10 | 
            +
              },
         | 
| 11 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 12 | 
            +
              "do_lower_case": true,
         | 
| 13 | 
            +
              "eos_token": {
         | 
| 14 | 
            +
                "__type": "AddedToken",
         | 
| 15 | 
            +
                "content": "<|endoftext|>",
         | 
| 16 | 
            +
                "lstrip": false,
         | 
| 17 | 
            +
                "normalized": true,
         | 
| 18 | 
            +
                "rstrip": false,
         | 
| 19 | 
            +
                "single_word": false
         | 
| 20 | 
            +
              },
         | 
| 21 | 
            +
              "errors": "replace",
         | 
| 22 | 
            +
              "model_max_length": 77,
         | 
| 23 | 
            +
              "pad_token": "!",
         | 
| 24 | 
            +
              "tokenizer_class": "CLIPTokenizer",
         | 
| 25 | 
            +
              "unk_token": {
         | 
| 26 | 
            +
                "__type": "AddedToken",
         | 
| 27 | 
            +
                "content": "<|endoftext|>",
         | 
| 28 | 
            +
                "lstrip": false,
         | 
| 29 | 
            +
                "normalized": true,
         | 
| 30 | 
            +
                "rstrip": false,
         | 
| 31 | 
            +
                "single_word": false
         | 
| 32 | 
            +
              }
         | 
| 33 | 
            +
            }
         | 
    	
        tokenizer_2/vocab.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        unet/config.json
    ADDED
    
    | @@ -0,0 +1,78 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "UNet2DConditionModel",
         | 
| 3 | 
            +
              "_diffusers_version": "0.25.0",
         | 
| 4 | 
            +
              "_name_or_path": "valhalla/sdxl-inpaint-ema",
         | 
| 5 | 
            +
              "act_fn": "silu",
         | 
| 6 | 
            +
              "addition_embed_type": "text_time",
         | 
| 7 | 
            +
              "addition_embed_type_num_heads": 64,
         | 
| 8 | 
            +
              "addition_time_embed_dim": 256,
         | 
| 9 | 
            +
              "attention_head_dim": [
         | 
| 10 | 
            +
                5,
         | 
| 11 | 
            +
                10,
         | 
| 12 | 
            +
                20
         | 
| 13 | 
            +
              ],
         | 
| 14 | 
            +
              "attention_type": "default",
         | 
| 15 | 
            +
              "block_out_channels": [
         | 
| 16 | 
            +
                320,
         | 
| 17 | 
            +
                640,
         | 
| 18 | 
            +
                1280
         | 
| 19 | 
            +
              ],
         | 
| 20 | 
            +
              "center_input_sample": false,
         | 
| 21 | 
            +
              "class_embed_type": null,
         | 
| 22 | 
            +
              "class_embeddings_concat": false,
         | 
| 23 | 
            +
              "conv_in_kernel": 3,
         | 
| 24 | 
            +
              "conv_out_kernel": 3,
         | 
| 25 | 
            +
              "cross_attention_dim": 2048,
         | 
| 26 | 
            +
              "cross_attention_norm": null,
         | 
| 27 | 
            +
              "decay": 0.9999,
         | 
| 28 | 
            +
              "down_block_types": [
         | 
| 29 | 
            +
                "DownBlock2D",
         | 
| 30 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 31 | 
            +
                "CrossAttnDownBlock2D"
         | 
| 32 | 
            +
              ],
         | 
| 33 | 
            +
              "downsample_padding": 1,
         | 
| 34 | 
            +
              "dual_cross_attention": false,
         | 
| 35 | 
            +
              "encoder_hid_dim": 1280,
         | 
| 36 | 
            +
              "encoder_hid_dim_type": "ip_image_proj",
         | 
| 37 | 
            +
              "flip_sin_to_cos": true,
         | 
| 38 | 
            +
              "freq_shift": 0,
         | 
| 39 | 
            +
              "in_channels": 13,
         | 
| 40 | 
            +
              "inv_gamma": 1.0,
         | 
| 41 | 
            +
              "layers_per_block": 2,
         | 
| 42 | 
            +
              "mid_block_only_cross_attention": null,
         | 
| 43 | 
            +
              "mid_block_scale_factor": 1,
         | 
| 44 | 
            +
              "mid_block_type": "UNetMidBlock2DCrossAttn",
         | 
| 45 | 
            +
              "min_decay": 0.0,
         | 
| 46 | 
            +
              "norm_eps": 1e-05,
         | 
| 47 | 
            +
              "norm_num_groups": 32,
         | 
| 48 | 
            +
              "num_attention_heads": null,
         | 
| 49 | 
            +
              "num_class_embeds": null,
         | 
| 50 | 
            +
              "only_cross_attention": false,
         | 
| 51 | 
            +
              "optimization_step": 37000,
         | 
| 52 | 
            +
              "out_channels": 4,
         | 
| 53 | 
            +
              "power": 0.6666666666666666,
         | 
| 54 | 
            +
              "projection_class_embeddings_input_dim": 2816,
         | 
| 55 | 
            +
              "resnet_out_scale_factor": 1.0,
         | 
| 56 | 
            +
              "resnet_skip_time_act": false,
         | 
| 57 | 
            +
              "resnet_time_scale_shift": "default",
         | 
| 58 | 
            +
              "sample_size": 128,
         | 
| 59 | 
            +
              "time_cond_proj_dim": null,
         | 
| 60 | 
            +
              "time_embedding_act_fn": null,
         | 
| 61 | 
            +
              "time_embedding_dim": null,
         | 
| 62 | 
            +
              "time_embedding_type": "positional",
         | 
| 63 | 
            +
              "timestep_post_act": null,
         | 
| 64 | 
            +
              "transformer_layers_per_block": [
         | 
| 65 | 
            +
                1,
         | 
| 66 | 
            +
                2,
         | 
| 67 | 
            +
                10
         | 
| 68 | 
            +
              ],
         | 
| 69 | 
            +
              "up_block_types": [
         | 
| 70 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 71 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 72 | 
            +
                "UpBlock2D"
         | 
| 73 | 
            +
              ],
         | 
| 74 | 
            +
              "upcast_attention": null,
         | 
| 75 | 
            +
              "update_after_step": 0,
         | 
| 76 | 
            +
              "use_ema_warmup": false,
         | 
| 77 | 
            +
              "use_linear_projection": true
         | 
| 78 | 
            +
            }
         | 
    	
        unet/diffusion_pytorch_model.bin
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:046b775cb9bbc67635fc3b148bb03bfe00496ce2f9ce8488a82fdb388669a521
         | 
| 3 | 
            +
            size 11965769774
         | 
    	
        unet_encoder/config.json
    ADDED
    
    | @@ -0,0 +1,68 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "UNet2DConditionModel",
         | 
| 3 | 
            +
              "_diffusers_version": "0.19.0.dev0",
         | 
| 4 | 
            +
              "act_fn": "silu",
         | 
| 5 | 
            +
              "addition_embed_type_num_heads": 64,
         | 
| 6 | 
            +
              "addition_time_embed_dim": 256,
         | 
| 7 | 
            +
              "attention_head_dim": [
         | 
| 8 | 
            +
                5,
         | 
| 9 | 
            +
                10,
         | 
| 10 | 
            +
                20
         | 
| 11 | 
            +
              ],
         | 
| 12 | 
            +
              "block_out_channels": [
         | 
| 13 | 
            +
                320,
         | 
| 14 | 
            +
                640,
         | 
| 15 | 
            +
                1280
         | 
| 16 | 
            +
              ],
         | 
| 17 | 
            +
              "center_input_sample": false,
         | 
| 18 | 
            +
              "class_embed_type": null,
         | 
| 19 | 
            +
              "class_embeddings_concat": false,
         | 
| 20 | 
            +
              "conv_in_kernel": 3,
         | 
| 21 | 
            +
              "conv_out_kernel": 3,
         | 
| 22 | 
            +
              "cross_attention_dim": 2048,
         | 
| 23 | 
            +
              "cross_attention_norm": null,
         | 
| 24 | 
            +
              "down_block_types": [
         | 
| 25 | 
            +
                "DownBlock2D",
         | 
| 26 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 27 | 
            +
                "CrossAttnDownBlock2D"
         | 
| 28 | 
            +
              ],
         | 
| 29 | 
            +
              "downsample_padding": 1,
         | 
| 30 | 
            +
              "dual_cross_attention": false,
         | 
| 31 | 
            +
              "encoder_hid_dim": null,
         | 
| 32 | 
            +
              "encoder_hid_dim_type": null,
         | 
| 33 | 
            +
              "flip_sin_to_cos": true,
         | 
| 34 | 
            +
              "freq_shift": 0,
         | 
| 35 | 
            +
              "in_channels": 4,
         | 
| 36 | 
            +
              "layers_per_block": 2,
         | 
| 37 | 
            +
              "mid_block_only_cross_attention": null,
         | 
| 38 | 
            +
              "mid_block_scale_factor": 1,
         | 
| 39 | 
            +
              "mid_block_type": "UNetMidBlock2DCrossAttn",
         | 
| 40 | 
            +
              "norm_eps": 1e-05,
         | 
| 41 | 
            +
              "norm_num_groups": 32,
         | 
| 42 | 
            +
              "num_attention_heads": null,
         | 
| 43 | 
            +
              "num_class_embeds": null,
         | 
| 44 | 
            +
              "only_cross_attention": false,
         | 
| 45 | 
            +
              "out_channels": 4,
         | 
| 46 | 
            +
              "projection_class_embeddings_input_dim": 2816,
         | 
| 47 | 
            +
              "resnet_out_scale_factor": 1.0,
         | 
| 48 | 
            +
              "resnet_skip_time_act": false,
         | 
| 49 | 
            +
              "resnet_time_scale_shift": "default",
         | 
| 50 | 
            +
              "sample_size": 128,
         | 
| 51 | 
            +
              "time_cond_proj_dim": null,
         | 
| 52 | 
            +
              "time_embedding_act_fn": null,
         | 
| 53 | 
            +
              "time_embedding_dim": null,
         | 
| 54 | 
            +
              "time_embedding_type": "positional",
         | 
| 55 | 
            +
              "timestep_post_act": null,
         | 
| 56 | 
            +
              "transformer_layers_per_block": [
         | 
| 57 | 
            +
                1,
         | 
| 58 | 
            +
                2,
         | 
| 59 | 
            +
                10
         | 
| 60 | 
            +
              ],
         | 
| 61 | 
            +
              "up_block_types": [
         | 
| 62 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 63 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 64 | 
            +
                "UpBlock2D"
         | 
| 65 | 
            +
              ],
         | 
| 66 | 
            +
              "upcast_attention": null,
         | 
| 67 | 
            +
              "use_linear_projection": true
         | 
| 68 | 
            +
            }
         | 
    	
        unet_encoder/diffusion_pytorch_model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:357650fbfb3c7b4d94c1f5fd7664da819ad1ff5a839430484b4ec422d03f710a
         | 
| 3 | 
            +
            size 10270077736
         | 
    	
        vae/config.json
    ADDED
    
    | @@ -0,0 +1,32 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "AutoencoderKL",
         | 
| 3 | 
            +
              "_diffusers_version": "0.21.0.dev0",
         | 
| 4 | 
            +
              "_name_or_path": "madebyollin/sdxl-vae-fp16-fix",
         | 
| 5 | 
            +
              "act_fn": "silu",
         | 
| 6 | 
            +
              "block_out_channels": [
         | 
| 7 | 
            +
                128,
         | 
| 8 | 
            +
                256,
         | 
| 9 | 
            +
                512,
         | 
| 10 | 
            +
                512
         | 
| 11 | 
            +
              ],
         | 
| 12 | 
            +
              "down_block_types": [
         | 
| 13 | 
            +
                "DownEncoderBlock2D",
         | 
| 14 | 
            +
                "DownEncoderBlock2D",
         | 
| 15 | 
            +
                "DownEncoderBlock2D",
         | 
| 16 | 
            +
                "DownEncoderBlock2D"
         | 
| 17 | 
            +
              ],
         | 
| 18 | 
            +
              "force_upcast": false,
         | 
| 19 | 
            +
              "in_channels": 3,
         | 
| 20 | 
            +
              "latent_channels": 4,
         | 
| 21 | 
            +
              "layers_per_block": 2,
         | 
| 22 | 
            +
              "norm_num_groups": 32,
         | 
| 23 | 
            +
              "out_channels": 3,
         | 
| 24 | 
            +
              "sample_size": 512,
         | 
| 25 | 
            +
              "scaling_factor": 0.13025,
         | 
| 26 | 
            +
              "up_block_types": [
         | 
| 27 | 
            +
                "UpDecoderBlock2D",
         | 
| 28 | 
            +
                "UpDecoderBlock2D",
         | 
| 29 | 
            +
                "UpDecoderBlock2D",
         | 
| 30 | 
            +
                "UpDecoderBlock2D"
         | 
| 31 | 
            +
              ]
         | 
| 32 | 
            +
            }
         | 
    	
        vae/diffusion_pytorch_model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:98a14dc6fe8d71c83576f135a87c61a16561c9c080abba418d2cc976ee034f88
         | 
| 3 | 
            +
            size 334643268
         | 
