gghfez/Llama-3.3-90B-Vision-merged

Since Meta have finished up with Llama3 and likely won't be releasing a version 3.3 for their vision model, I've swapped out all the text layers from Llama-3.2-90B-Vision-Instruct (which are identical to Llama-3.1-70B-instruct) with those from Llama-3.3-70b-instruct) so we can get the benefits of Llama3.3-70b-Instruct when doing vision tasks.

I've switched to this model now and it's working as expected. If anyone has a comprehensive vision benchmark let me know. I'd be curious to see if there's a measurable performance improvement.

Downloads last month
236
Safetensors
Model size
88.6B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for gghfez/Llama-3.3-90B-Vision-merged

Finetuned
(7)
this model