Model weights loading issue for 1.6 B model
#4
by
Veb-BLK
- opened
from transformers import AutoModelForImageTextToText, AutoProcessor
model_id = "LiquidAI/LFM2-VL-1.6B"
processor_id = model_id
processor = AutoProcessor.from_pretrained(
model_id,
dtype="bfloat16",
trust_remote_code=True,
device_map="auto",
)
model = AutoModelForImageTextToText.from_pretrained(
model_id,
dtype="bfloat16",
trust_remote_code=True,
device_map="auto",
)
Output:
model.safetensors:β100%
β3.17G/3.17Gβ[00:24<00:00,β601MB/s]
Some weights of the model checkpoint at LiquidAI/LFM2-VL-1.6B were not used when initializing Lfm2VlForConditionalGeneration: ['model.vision_tower.vision_model.encoder.layers.25.layer_norm1.bias', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm1.weight', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm2.bias', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm2.weight', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc1.bias', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc1.weight', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc2.bias', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc2.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.weight']
- This IS expected if you are initializing Lfm2VlForConditionalGeneration from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing Lfm2VlForConditionalGeneration from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).