{ "model": "google/paligemma-3b-pt-224", "model_type": "paligemma", "model_revision": null, "task_type": "causal_lm", "torch_dtype": "float16", "attn_impl": null, "new_special_tokens": [], "num_labels": null, "problem_type": null, "rope_scaling": null, "device_map": null, "max_memory": {}, "max_model_len": null, "local_repo_path": null, "init_strategy": null, "template": "paligemma", "system": null, "max_length": 2048, "truncation_strategy": "delete", "max_pixels": null, "agent_template": null, "norm_bbox": null, "use_chat_template": true, "padding_free": false, "padding_side": "right", "loss_scale": "default", "sequence_parallel_size": 1, "response_prefix": null, "template_backend": "swift", "dataset": [], "val_dataset": [], "split_dataset_ratio": 0.0, "data_seed": 42, "dataset_num_proc": 1, "load_from_cache_file": true, "dataset_shuffle": true, "val_dataset_shuffle": false, "streaming": false, "interleave_prob": null, "stopping_strategy": "first_exhausted", "shuffle_buffer_size": 1000, "download_mode": "reuse_dataset_if_exists", "columns": {}, "strict": false, "remove_unused_columns": true, "model_name": null, "model_author": null, "custom_dataset_info": [], "quant_method": "bnb", "quant_bits": 4, "hqq_axis": null, "bnb_4bit_compute_dtype": "float16", "bnb_4bit_quant_type": "nf4", "bnb_4bit_use_double_quant": true, "bnb_4bit_quant_storage": null, "max_new_tokens": null, "temperature": null, "top_k": null, "top_p": null, "repetition_penalty": null, "num_beams": 1, "stream": false, "stop_words": [], "logprobs": false, "top_logprobs": null, "ckpt_dir": "/kaggle/working/output_Kvasir-VQA-x1/v0-20250902-124816/checkpoint-188", "lora_modules": [], "tuner_backend": "peft", "train_type": "lora", "adapters": [ "/kaggle/working/output_Kvasir-VQA-x1/v0-20250902-124816/checkpoint-188" ], "external_plugins": [], "seed": 42, "model_kwargs": {}, "load_args": true, "load_data_args": false, "packing": false, "lazy_tokenize": true, "cached_dataset": [], "custom_register_path": [], "use_hf": true, "hub_token": null, "ddp_timeout": 18000000, "ddp_backend": null, "ignore_args_error": false, "use_swift_lora": false, "merge_lora": false, "safe_serialization": true, "max_shard_size": "5GB", "output_dir": "/kaggle/working/output_Kvasir-VQA-x1/v0-20250902-124816/checkpoint-188-bnb-int4", "quant_n_samples": 256, "quant_batch_size": 1, "group_size": 128, "to_cached_dataset": false, "to_ollama": false, "to_mcore": false, "to_hf": false, "mcore_model": null, "mcore_adapters": [], "thread_count": null, "test_convert_precision": false, "push_to_hub": true, "hub_model_id": "Kvasir-VQA-x1-lora_250902-1244", "hub_private_repo": false, "commit_message": "update files", "to_peft_format": false, "exist_ok": false, "rank": -1, "local_rank": -1, "global_world_size": 1, "local_world_size": 1, "model_suffix": "paligemma-3b-pt-224", "model_info": "ModelInfo(model_type='paligemma', model_dir='/root/.cache/huggingface/hub/models--google--paligemma-3b-pt-224/snapshots/35e4f46485b4d07967e7e9935bc3786aad50687c', torch_dtype=torch.float16, max_model_len=8192, quant_method='bnb', quant_bits=4, rope_scaling=None, is_moe_model=False, config=None, task_type='causal_lm', num_labels=None)", "model_meta": "ModelMeta(model_type='paligemma', model_groups=[ModelGroup(models=[Model(ms_model_id='AI-ModelScope/paligemma-3b-pt-224', hf_model_id='google/paligemma-3b-pt-224', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma-3b-pt-448', hf_model_id='google/paligemma-3b-pt-448', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma-3b-pt-896', hf_model_id='google/paligemma-3b-pt-896', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='AI-ModelScope/paligemma-3b-mix-224', hf_model_id='google/paligemma-3b-mix-224', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma-3b-mix-448', hf_model_id='google/paligemma-3b-mix-448', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='AI-ModelScope/paligemma2-3b-pt-224', hf_model_id='google/paligemma2-3b-pt-224', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-3b-pt-448', hf_model_id='google/paligemma2-3b-pt-448', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-3b-pt-896', hf_model_id='google/paligemma2-3b-pt-896', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-10b-pt-224', hf_model_id='google/paligemma2-10b-pt-224', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-10b-pt-448', hf_model_id='google/paligemma2-10b-pt-448', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-10b-pt-896', hf_model_id='google/paligemma2-10b-pt-896', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-28b-pt-224', hf_model_id='google/paligemma2-28b-pt-224', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-28b-pt-448', hf_model_id='google/paligemma2-28b-pt-448', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-28b-pt-896', hf_model_id='google/paligemma2-28b-pt-896', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='AI-ModelScope/paligemma2-3b-ft-docci-448', hf_model_id='google/paligemma2-3b-ft-docci-448', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='AI-ModelScope/paligemma2-10b-ft-docci-448', hf_model_id='google/paligemma2-10b-ft-docci-448', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='paligemma', get_function=, model_arch=MultiModelKeys(arch_name='llava_hf', embedding=None, module_list=None, lm_head=None, q_proj=None, k_proj=None, v_proj=None, o_proj=None, attention=None, mlp=None, down_proj=None, qkv_proj=None, qk_proj=None, qa_proj=None, qb_proj=None, kv_proj=None, kva_proj=None, kvb_proj=None, language_model=['model.language_model'], aligner=['model.multi_modal_projector'], vision_tower=['model.vision_tower'], generator=[]), architectures=['PaliGemmaForConditionalGeneration'], additional_saved_files=[], torch_dtype=None, is_multimodal=True, is_reward=False, task_type=None, ignore_patterns=None, requires=['transformers>=4.41'], tags=['vision'])", "model_dir": "/root/.cache/huggingface/hub/models--google--paligemma-3b-pt-224/snapshots/35e4f46485b4d07967e7e9935bc3786aad50687c", "hub": "" }