optimum-internal-testing/tiny_random_bert_neuronx
Feature Extraction
•
Updated
•
612
None defined yet.
diffusers
🧨bistandbytes
as the official backend but using others like torchao
is already very simple. enable_model_cpu_offload()
from optimum.onnxruntime import ORTModelForSequenceClassification
# Load the model from the hub and export it to the ONNX format
model_id = "distilbert-base-uncased-finetuned-sst-2-english"
model = ORTModelForSequenceClassification.from_pretrained(model_id, export=True)
torch.compile()
them. from_single_file()
method, highlighting our efforts for making our library more amenable to community features 🤗