Spaces:
Paused
Paused
Upload requirements.txt
Browse files- requirements.txt +38 -0
requirements.txt
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# base dependencies, for Wan and Gradio
|
| 2 |
+
|
| 3 |
+
# we don't have a flash attention binary yet for Pytorch 2.7
|
| 4 |
+
# so we use pytorch 2.6
|
| 5 |
+
torch==2.6.0
|
| 6 |
+
torchvision==0.21.0
|
| 7 |
+
torchdata==0.10.1
|
| 8 |
+
torchao==0.9.0
|
| 9 |
+
|
| 10 |
+
numpy>=1.26.4
|
| 11 |
+
|
| 12 |
+
# some important dependendencies
|
| 13 |
+
ftfy
|
| 14 |
+
|
| 15 |
+
# peft==0.15.1
|
| 16 |
+
peft
|
| 17 |
+
|
| 18 |
+
# it is recommended to always use the latest version
|
| 19 |
+
diffusers @ git+https://github.com/huggingface/diffusers.git@main
|
| 20 |
+
|
| 21 |
+
flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
| 22 |
+
|
| 23 |
+
transformers>=4.45.2
|
| 24 |
+
|
| 25 |
+
accelerate
|
| 26 |
+
safetensors
|
| 27 |
+
|
| 28 |
+
imageio
|
| 29 |
+
imageio-ffmpeg
|
| 30 |
+
|
| 31 |
+
gradio==5.38.2
|
| 32 |
+
|
| 33 |
+
# more dependencies, specific to OmniAvatar
|
| 34 |
+
librosa==0.10.2.post1
|
| 35 |
+
tqdm
|
| 36 |
+
scipy==1.14.0
|
| 37 |
+
xfuser==0.4.1
|
| 38 |
+
einops
|