Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -13,7 +13,7 @@ import json
|
|
13 |
from datasets import concatenate_datasets, Dataset
|
14 |
from datasets import load_dataset
|
15 |
|
16 |
-
from huggingface_hub import hf_hub_url
|
17 |
|
18 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
19 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification, TrainingArguments, Trainer
|
@@ -84,5 +84,11 @@ if __name__ == "__main__":
|
|
84 |
tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
|
85 |
output = compute_model_card_evaluation_results(tokenizer, model_checkpoint, raw_datasets, metric)
|
86 |
print(json.dumps(output))
|
87 |
-
st.text_area(label="Output Data:", value=st.json(output, expanded=True), height=300)
|
|
|
|
|
|
|
|
|
|
|
|
|
88 |
|
|
|
13 |
from datasets import concatenate_datasets, Dataset
|
14 |
from datasets import load_dataset
|
15 |
|
16 |
+
from huggingface_hub import hf_hub_url, ModelCard
|
17 |
|
18 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
19 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification, TrainingArguments, Trainer
|
|
|
84 |
tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
|
85 |
output = compute_model_card_evaluation_results(tokenizer, model_checkpoint, raw_datasets, metric)
|
86 |
print(json.dumps(output))
|
87 |
+
#st.text_area(label="Output Data:", value=st.json(output, expanded=True), height=300)
|
88 |
+
st.json(output, expanded=True)
|
89 |
+
card = ModelCard.load(model_checkpoint)
|
90 |
+
|
91 |
+
st.text_area(label="Model Card Data:", value=card.data.eval_results)
|
92 |
+
|
93 |
+
|
94 |
|