πŸ”§ Convert model binaries for Transformers.js v3

#2
by whitphx HF Staff - opened

This automated migration updates the model repository for Transformers.js v3.

πŸ”§ Model Binary Migration Results

Models processed: decoder_model.onnx, decoder_model_merged.onnx, decoder_with_past_model.onnx, encoder_model.onnx
βœ… Added quantization modes (missing files):

  • decoder_model.onnx: q4f16
  • decoder_with_past_model.onnx: q4f16
  • encoder_model.onnx: q4f16

πŸ“‹ Summary

  • Migration type: Model binaries quantization
  • Purpose: Ensure all required quantized model variants are available and valid for Transformers.js v3
  • Quantization modes: fp16, q8, int8, uint8, q4, q4f16, bnb4

βœ… What was done:

  1. Validation: Checked existing quantized model files for validity
  2. Quantization: Generated missing or invalid quantized variants using transformers.js quantization tools
  3. Post-validation: Verified all generated models load correctly with Transformers.js v3
  4. Cleanup: Removed any invalid models that couldn't be regenerated

πŸ” Please review:

  • Verify that the quantized models work correctly with your use case
  • Test the models in your application to ensure compatibility
  • Check that file sizes are reasonable for your deployment needs

πŸ€– Generated by transformers-js-v3-migrator

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment