This skill allows Claude to evaluate machine learning models using a comprehensive suite of metrics. It should be used when the user requests model performance analysis, validation, or testing. Claude can use this skill to assess model accuracy, precision, recall, F1-score, and other relevant metrics. Trigger this skill when the user mentions "evaluate model", "model performance", "testing metrics", "validation results", or requests a comprehensive "model evaluation".
Evaluates ML model performance using a suite of metrics; supports comparison and improvement analysis.
Download the skill repository as a ZIP file
/plugin marketplace add jeremylongshore/claude-code-plugins-plus