Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -27,7 +27,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
|
| 27 |
|
| 28 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 29 |
|
| 30 |
-
# Load
|
| 31 |
MODEL_ID_M = "Qwen/Qwen2.5-VL-7B-Instruct"
|
| 32 |
processor_m = AutoProcessor.from_pretrained(MODEL_ID_M, trust_remote_code=True)
|
| 33 |
model_m = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|
|
@@ -36,7 +36,7 @@ model_m = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|
|
| 36 |
torch_dtype=torch.float16
|
| 37 |
).to(device).eval()
|
| 38 |
|
| 39 |
-
# Load
|
| 40 |
MODEL_ID_X = "Qwen/Qwen2.5-VL-3B-Instruct"
|
| 41 |
processor_x = AutoProcessor.from_pretrained(MODEL_ID_X, trust_remote_code=True)
|
| 42 |
model_x = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|
|
|
|
| 27 |
|
| 28 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 29 |
|
| 30 |
+
# Load Qwen2.5-VL-7B-Instruct
|
| 31 |
MODEL_ID_M = "Qwen/Qwen2.5-VL-7B-Instruct"
|
| 32 |
processor_m = AutoProcessor.from_pretrained(MODEL_ID_M, trust_remote_code=True)
|
| 33 |
model_m = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|
|
|
|
| 36 |
torch_dtype=torch.float16
|
| 37 |
).to(device).eval()
|
| 38 |
|
| 39 |
+
# Load Qwen2.5-VL-3B-Instruct
|
| 40 |
MODEL_ID_X = "Qwen/Qwen2.5-VL-3B-Instruct"
|
| 41 |
processor_x = AutoProcessor.from_pretrained(MODEL_ID_X, trust_remote_code=True)
|
| 42 |
model_x = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|