Example code:
# Sample text to predict
text = "I love this movie, it was fantastic!"
# Tokenize the input text
inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True)
# Get model predictions
with torch.no_grad():
outputs = model(**inputs)
# Get the logits (model's raw output)
logits = outputs.logits
# Convert logits to probabilities (if needed) and get the predicted class (0 or 1)
predictions = torch.argmax(logits, dim=-1).item()
# Map the prediction to sentiment labels
labels = {0: "NEGATIVE", 1: "POSITIVE"} # Assuming binary classification
predicted_label = labels[predictions]
print(f"Predicted Sentiment: {predicted_label}")
- Downloads last month
- 23,527
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for appleboiy/DistilBERT-Sentiment
Base model
distilbert/distilbert-base-uncased