Error for fine tuning model when using FSDP: auto wrap: Could not find the transformer layer class LlavaOnevisionVisionAttention in the model.
#6 opened 3 days ago
by
liuzijing2014
Error when attempting to run either model... ValueError: embed_dim must be divisible by num_heads (got `embed_dim`: 1152 and `num_heads`: 14).
1
#4 opened 2 months ago
by
jdc4429
Download transformers for LlavaOnevisionForConditionalGeneration
2
#1 opened 3 months ago
by
mjbooo