🤖 AI Toolkit — Model Evaluation Tool Demo

← Back to Store

Model Metrics Calculator

Enter predicted vs. actual labels to compute precision, recall, F1-score, and confusion matrix.

📥 Input Labels

🤖 Predicted Labels

Accuracy
Precision (macro)
Recall (macro)
F1 Score (macro)

📊 Confusion Matrix

Run evaluation to see matrix

📋 Per-Class Metrics

Run evaluation to see breakdown

Get the Full Model Evaluation Tool

Python + HuggingFace integration: BLEU/ROUGE scoring, side-by-side model comparison, evaluation datasets, and HTML report export.

Buy Full Version — $49.00