Spaces:
Running
on
Zero
Running
on
Zero
| # This file is modified from https://github.com/haotian-liu/LLaVA/ | |
| import os | |
| import torch | |
| from .base_projector import MultimodalProjectorConfig, MultimodalProjector | |
| from transformers import PretrainedConfig, PreTrainedModel | |
| def build_mm_projector( | |
| model_type_or_path: str, config: PretrainedConfig | |
| ) -> PreTrainedModel: | |
| if model_type_or_path is None: | |
| return None | |
| ## load from pretrained model | |
| if config.resume_path: | |
| assert os.path.exists( | |
| model_type_or_path | |
| ), f"Resume mm projector path {model_type_or_path} does not exist!" | |
| return MultimodalProjector.from_pretrained( | |
| model_type_or_path, config, torch_dtype=eval(config.model_dtype) | |
| ) | |
| ## build from scratch | |
| else: | |
| mm_projector_cfg = MultimodalProjectorConfig(model_type_or_path) | |
| mm_projector = MultimodalProjector(mm_projector_cfg, config).to( | |
| eval(config.model_dtype) | |
| ) | |
| return mm_projector | |