mirror of
https://github.com/unslothai/unsloth.git
synced 2026-05-17 03:56:07 +00:00
697 B
697 B
Language Model Perplexity Evaluator
A Python module for evaluating language models using perplexity metrics with sliding window approach for long sequences. This evaluator provides efficient computation of perplexity scores across datasets with model comparison capabilities.
Basic Usage
from perplexity_evaluator import ppl_model, add_to_comparison, print_model_comparison
# Simple perplexity evaluation
dataset = {"text": ["Your text samples here...", "Another text sample..."]}
perplexity = ppl_model(model, tokenizer, dataset)
print(f"Model Perplexity: {perplexity:.4f}")
# Add to comparison tracker
add_to_comparison("My Model", perplexity)
print_model_comparison()