OrcaleSeek / inference.py
prelington's picture
Create inference.py
1814385 verified
from transformers import pipeline
import torch
# Load your model
classifier = pipeline(
"text-classification",
model="your-username/your-model-name", # Replace with your model path
tokenizer="your-username/your-model-name"
)
def predict(text):
"""Simple prediction function"""
result = classifier(text)
return result
# Example usage
if __name__ == "__main__":
sample_text = "This is an amazing model!"
prediction = predict(sample_text)
print(f"Input: {sample_text}")
print(f"Prediction: {prediction}")