Model weights loading issue for 1.6 B model

#4
by Veb-BLK - opened
from transformers import AutoModelForImageTextToText, AutoProcessor

model_id = "LiquidAI/LFM2-VL-1.6B"

processor_id = model_id


processor = AutoProcessor.from_pretrained(
    model_id,
    dtype="bfloat16",
    trust_remote_code=True,
    device_map="auto",
)

model = AutoModelForImageTextToText.from_pretrained(
    model_id,
    dtype="bfloat16",
    trust_remote_code=True,
    device_map="auto",
)

Output:

model.safetensors: 100%
 3.17G/3.17G [00:24<00:00, 601MB/s]
Some weights of the model checkpoint at LiquidAI/LFM2-VL-1.6B were not used when initializing Lfm2VlForConditionalGeneration: ['model.vision_tower.vision_model.encoder.layers.25.layer_norm1.bias', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm1.weight', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm2.bias', 'model.vision_tower.vision_model.encoder.layers.25.layer_norm2.weight', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc1.bias', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc1.weight', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc2.bias', 'model.vision_tower.vision_model.encoder.layers.25.mlp.fc2.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.weight', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.bias', 'model.vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.weight']
- This IS expected if you are initializing Lfm2VlForConditionalGeneration from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing Lfm2VlForConditionalGeneration from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).

Sign up or log in to comment