Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	Upload 6 files
#126
by
						
strategicpeanut1
	
							
						- opened
							
					
- README.md +4 -9
 - VQGAN_autoencoder.pth +3 -0
 - configuration.json +35 -0
 - gitattributes +34 -0
 - open_clip_pytorch_model.bin +3 -0
 - text2video_pytorch_model.pth +3 -0
 
    	
        README.md
    CHANGED
    
    | 
         @@ -1,12 +1,7 @@ 
     | 
|
| 1 | 
         
             
            ---
         
     | 
| 2 | 
         
            -
             
     | 
| 3 | 
         
            -
            emoji: π
         
     | 
| 4 | 
         
            -
            colorFrom: pink
         
     | 
| 5 | 
         
            -
            colorTo: pink
         
     | 
| 6 | 
         
            -
            sdk: gradio
         
     | 
| 7 | 
         
            -
            sdk_version: 3.23.0
         
     | 
| 8 | 
         
            -
            app_file: app.py
         
     | 
| 9 | 
         
            -
            pinned: false
         
     | 
| 10 | 
         
             
            ---
         
     | 
| 11 | 
         | 
| 12 | 
         
            -
             
     | 
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
             
            ---
         
     | 
| 2 | 
         
            +
            license: cc-by-nc-4.0
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 3 | 
         
             
            ---
         
     | 
| 4 | 
         | 
| 5 | 
         
            +
            https://huggingface.co/damo-vilab/modelscope-damo-text-to-video-synthesis, but with fp16 (half precision) weights
         
     | 
| 6 | 
         
            +
             
     | 
| 7 | 
         
            +
            Read all the info here https://huggingface.co/damo-vilab/modelscope-damo-text-to-video-synthesis/blob/main/README.md
         
     | 
    	
        VQGAN_autoencoder.pth
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:930e9865584beae2405d29bc06a05db3bb6a5b34eedd40a7db29b9156ed7d098
         
     | 
| 3 | 
         
            +
            size 2607657443
         
     | 
    	
        configuration.json
    ADDED
    
    | 
         @@ -0,0 +1,35 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {   "framework": "pytorch",
         
     | 
| 2 | 
         
            +
                "task": "text-to-video-synthesis",
         
     | 
| 3 | 
         
            +
                "model": {
         
     | 
| 4 | 
         
            +
                    "type": "latent-text-to-video-synthesis",
         
     | 
| 5 | 
         
            +
                    "model_args": {
         
     | 
| 6 | 
         
            +
                        "ckpt_clip": "open_clip_pytorch_model.bin",
         
     | 
| 7 | 
         
            +
                        "ckpt_unet": "text2video_pytorch_model.pth",
         
     | 
| 8 | 
         
            +
                        "ckpt_autoencoder": "VQGAN_autoencoder.pth",
         
     | 
| 9 | 
         
            +
                        "max_frames": 16,
         
     | 
| 10 | 
         
            +
                        "tiny_gpu": 1
         
     | 
| 11 | 
         
            +
                    },
         
     | 
| 12 | 
         
            +
                    "model_cfg": {
         
     | 
| 13 | 
         
            +
                        "unet_in_dim": 4,
         
     | 
| 14 | 
         
            +
                        "unet_dim": 320,
         
     | 
| 15 | 
         
            +
                        "unet_y_dim": 768,
         
     | 
| 16 | 
         
            +
                        "unet_context_dim": 1024,
         
     | 
| 17 | 
         
            +
                        "unet_out_dim": 4,
         
     | 
| 18 | 
         
            +
                        "unet_dim_mult": [1, 2, 4, 4],
         
     | 
| 19 | 
         
            +
                        "unet_num_heads": 8,
         
     | 
| 20 | 
         
            +
                        "unet_head_dim": 64,
         
     | 
| 21 | 
         
            +
                        "unet_res_blocks": 2,
         
     | 
| 22 | 
         
            +
                        "unet_attn_scales": [1, 0.5, 0.25],
         
     | 
| 23 | 
         
            +
                        "unet_dropout": 0.1,
         
     | 
| 24 | 
         
            +
                        "temporal_attention": "True",
         
     | 
| 25 | 
         
            +
                        "num_timesteps": 1000,
         
     | 
| 26 | 
         
            +
                        "mean_type": "eps",
         
     | 
| 27 | 
         
            +
                        "var_type": "fixed_small",
         
     | 
| 28 | 
         
            +
                        "loss_type": "mse"
         
     | 
| 29 | 
         
            +
                    }
         
     | 
| 30 | 
         
            +
                },
         
     | 
| 31 | 
         
            +
                "pipeline": {
         
     | 
| 32 | 
         
            +
                    "type": "latent-text-to-video-synthesis"
         
     | 
| 33 | 
         
            +
                }
         
     | 
| 34 | 
         
            +
            }
         
     | 
| 35 | 
         
            +
             
     | 
    	
        gitattributes
    ADDED
    
    | 
         @@ -0,0 +1,34 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            *.7z filter=lfs diff=lfs merge=lfs -text
         
     | 
| 2 | 
         
            +
            *.arrow filter=lfs diff=lfs merge=lfs -text
         
     | 
| 3 | 
         
            +
            *.bin filter=lfs diff=lfs merge=lfs -text
         
     | 
| 4 | 
         
            +
            *.bz2 filter=lfs diff=lfs merge=lfs -text
         
     | 
| 5 | 
         
            +
            *.ckpt filter=lfs diff=lfs merge=lfs -text
         
     | 
| 6 | 
         
            +
            *.ftz filter=lfs diff=lfs merge=lfs -text
         
     | 
| 7 | 
         
            +
            *.gz filter=lfs diff=lfs merge=lfs -text
         
     | 
| 8 | 
         
            +
            *.h5 filter=lfs diff=lfs merge=lfs -text
         
     | 
| 9 | 
         
            +
            *.joblib filter=lfs diff=lfs merge=lfs -text
         
     | 
| 10 | 
         
            +
            *.lfs.* filter=lfs diff=lfs merge=lfs -text
         
     | 
| 11 | 
         
            +
            *.mlmodel filter=lfs diff=lfs merge=lfs -text
         
     | 
| 12 | 
         
            +
            *.model filter=lfs diff=lfs merge=lfs -text
         
     | 
| 13 | 
         
            +
            *.msgpack filter=lfs diff=lfs merge=lfs -text
         
     | 
| 14 | 
         
            +
            *.npy filter=lfs diff=lfs merge=lfs -text
         
     | 
| 15 | 
         
            +
            *.npz filter=lfs diff=lfs merge=lfs -text
         
     | 
| 16 | 
         
            +
            *.onnx filter=lfs diff=lfs merge=lfs -text
         
     | 
| 17 | 
         
            +
            *.ot filter=lfs diff=lfs merge=lfs -text
         
     | 
| 18 | 
         
            +
            *.parquet filter=lfs diff=lfs merge=lfs -text
         
     | 
| 19 | 
         
            +
            *.pb filter=lfs diff=lfs merge=lfs -text
         
     | 
| 20 | 
         
            +
            *.pickle filter=lfs diff=lfs merge=lfs -text
         
     | 
| 21 | 
         
            +
            *.pkl filter=lfs diff=lfs merge=lfs -text
         
     | 
| 22 | 
         
            +
            *.pt filter=lfs diff=lfs merge=lfs -text
         
     | 
| 23 | 
         
            +
            *.pth filter=lfs diff=lfs merge=lfs -text
         
     | 
| 24 | 
         
            +
            *.rar filter=lfs diff=lfs merge=lfs -text
         
     | 
| 25 | 
         
            +
            *.safetensors filter=lfs diff=lfs merge=lfs -text
         
     | 
| 26 | 
         
            +
            saved_model/**/* filter=lfs diff=lfs merge=lfs -text
         
     | 
| 27 | 
         
            +
            *.tar.* filter=lfs diff=lfs merge=lfs -text
         
     | 
| 28 | 
         
            +
            *.tflite filter=lfs diff=lfs merge=lfs -text
         
     | 
| 29 | 
         
            +
            *.tgz filter=lfs diff=lfs merge=lfs -text
         
     | 
| 30 | 
         
            +
            *.wasm filter=lfs diff=lfs merge=lfs -text
         
     | 
| 31 | 
         
            +
            *.xz filter=lfs diff=lfs merge=lfs -text
         
     | 
| 32 | 
         
            +
            *.zip filter=lfs diff=lfs merge=lfs -text
         
     | 
| 33 | 
         
            +
            *.zst filter=lfs diff=lfs merge=lfs -text
         
     | 
| 34 | 
         
            +
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         
     | 
    	
        open_clip_pytorch_model.bin
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:73c32c62eebf1112b0693ff9e3ecfa0573ba02cd279420ea4da4af1cbfb39e3b
         
     | 
| 3 | 
         
            +
            size 1972451989
         
     | 
    	
        text2video_pytorch_model.pth
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:cbba5db98d5432378f9ccdb6bd572768c7ff190dd83b9b76c3218594c793fedd
         
     | 
| 3 | 
         
            +
            size 2823000429
         
     |