{ | |
"name": "Summarization Fine-tuning Dataset", | |
"description": "A dataset for fine-tuning small language models on summarization tasks", | |
"format": "alpaca", | |
"statistics": { | |
"total_examples": 2000, | |
"train_examples": 1600, | |
"val_examples": 200, | |
"test_examples": 200, | |
"dataset_distribution": { | |
"xsum": { | |
"count": 2000, | |
"percentage": 100.0 | |
} | |
} | |
}, | |
"configuration": { | |
"max_tokens": 2000, | |
"tokenizer": "gpt2", | |
"seed": 42 | |
} | |
} |