sha
stringlengths 40
40
| text
stringlengths 1
13.4M
| id
stringlengths 2
117
| tags
sequencelengths 1
7.91k
| created_at
stringlengths 25
25
| metadata
stringlengths 2
875k
| last_modified
stringlengths 25
25
| arxiv
sequencelengths 0
25
| languages
sequencelengths 0
7.91k
| tags_str
stringlengths 17
159k
| text_str
stringlengths 1
447k
| text_lists
sequencelengths 0
352
| processed_texts
sequencelengths 1
353
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
f628640cea8e85f48246455adf44356b2fd45c32 |
# Dataset Card for "lmqg/qa_squadshifts_synthetic"
## Dataset Description
- **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation)
- **Paper:** [https://arxiv.org/abs/2210.03992](https://arxiv.org/abs/2210.03992)
- **Point of Contact:** [Asahi Ushio](http://asahiushio.com/)
### Dataset Summary
This is a synthetic QA dataset generated with fine-tuned QG models over [`lmqg/qa_squadshifts`](https://huggingface.co/datasets/lmqg/qa_squadshifts), made for question-answering based evaluation (QAE) for question generation model proposed by [Zhang and Bansal, 2019](https://aclanthology.org/D19-1253/).
The test split is the original validation set of [`lmqg/qa_squadshifts`](https://huggingface.co/datasets/lmqg/qa_squadshifts), where the model should be evaluate on.
### Supported Tasks and Leaderboards
* `question-answering`
### Languages
English (en)
## Dataset Structure
### Data Fields
The data fields are the same among all splits.
#### plain_text
- `id`: a `string` feature of id
- `title`: a `string` feature of title of the paragraph
- `context`: a `string` feature of paragraph
- `question`: a `string` feature of question
- `answers`: a `json` feature of answers
### Data Splits
TBA
## Citation Information
```
@inproceedings{ushio-etal-2022-generative,
title = "{G}enerative {L}anguage {M}odels for {P}aragraph-{L}evel {Q}uestion {G}eneration",
author = "Ushio, Asahi and
Alva-Manchego, Fernando and
Camacho-Collados, Jose",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2022",
address = "Abu Dhabi, U.A.E.",
publisher = "Association for Computational Linguistics",
}
``` | lmqg/qa_squadshifts_synthetic | [
"task_categories:question-answering",
"task_ids:extractive-qa",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|wikipedia",
"language:en",
"license:cc-by-4.0",
"arxiv:2210.03992",
"region:us"
] | 2022-12-20T08:31:18+00:00 | {"language": "en", "license": "cc-by-4.0", "multilinguality": "monolingual", "size_categories": "10K<n<100K", "source_datasets": ["extended|wikipedia"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa"], "pretty_name": "Synthetic QA dataset on SQuADShifts."} | 2023-01-15T14:25:15+00:00 | [
"2210.03992"
] | [
"en"
] | TAGS
#task_categories-question-answering #task_ids-extractive-qa #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|wikipedia #language-English #license-cc-by-4.0 #arxiv-2210.03992 #region-us
|
# Dataset Card for "lmqg/qa_squadshifts_synthetic"
## Dataset Description
- Repository: URL
- Paper: URL
- Point of Contact: Asahi Ushio
### Dataset Summary
This is a synthetic QA dataset generated with fine-tuned QG models over 'lmqg/qa_squadshifts', made for question-answering based evaluation (QAE) for question generation model proposed by Zhang and Bansal, 2019.
The test split is the original validation set of 'lmqg/qa_squadshifts', where the model should be evaluate on.
### Supported Tasks and Leaderboards
* 'question-answering'
### Languages
English (en)
## Dataset Structure
### Data Fields
The data fields are the same among all splits.
#### plain_text
- 'id': a 'string' feature of id
- 'title': a 'string' feature of title of the paragraph
- 'context': a 'string' feature of paragraph
- 'question': a 'string' feature of question
- 'answers': a 'json' feature of answers
### Data Splits
TBA
| [
"# Dataset Card for \"lmqg/qa_squadshifts_synthetic\"",
"## Dataset Description\n- Repository: URL\n- Paper: URL\n- Point of Contact: Asahi Ushio",
"### Dataset Summary\nThis is a synthetic QA dataset generated with fine-tuned QG models over 'lmqg/qa_squadshifts', made for question-answering based evaluation (QAE) for question generation model proposed by Zhang and Bansal, 2019.\nThe test split is the original validation set of 'lmqg/qa_squadshifts', where the model should be evaluate on.",
"### Supported Tasks and Leaderboards\n\n* 'question-answering'",
"### Languages\nEnglish (en)",
"## Dataset Structure",
"### Data Fields\nThe data fields are the same among all splits.",
"#### plain_text\n\n- 'id': a 'string' feature of id\n- 'title': a 'string' feature of title of the paragraph \n- 'context': a 'string' feature of paragraph \n- 'question': a 'string' feature of question\n- 'answers': a 'json' feature of answers",
"### Data Splits\nTBA"
] | [
"TAGS\n#task_categories-question-answering #task_ids-extractive-qa #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|wikipedia #language-English #license-cc-by-4.0 #arxiv-2210.03992 #region-us \n",
"# Dataset Card for \"lmqg/qa_squadshifts_synthetic\"",
"## Dataset Description\n- Repository: URL\n- Paper: URL\n- Point of Contact: Asahi Ushio",
"### Dataset Summary\nThis is a synthetic QA dataset generated with fine-tuned QG models over 'lmqg/qa_squadshifts', made for question-answering based evaluation (QAE) for question generation model proposed by Zhang and Bansal, 2019.\nThe test split is the original validation set of 'lmqg/qa_squadshifts', where the model should be evaluate on.",
"### Supported Tasks and Leaderboards\n\n* 'question-answering'",
"### Languages\nEnglish (en)",
"## Dataset Structure",
"### Data Fields\nThe data fields are the same among all splits.",
"#### plain_text\n\n- 'id': a 'string' feature of id\n- 'title': a 'string' feature of title of the paragraph \n- 'context': a 'string' feature of paragraph \n- 'question': a 'string' feature of question\n- 'answers': a 'json' feature of answers",
"### Data Splits\nTBA"
] |
07bb67ebf8bdcb9d22a4067280ebfa74c26d019d | # Dataset Card for "Kor_Jpn_Translation_Dataset"
### Dataset Summary
AI-Hub์์ ์ ๊ณตํ๋ ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฒ์ญ ๋ง๋ญ์น ๋ฐ์ดํฐ(https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&aihubDataSe=realm&dataSetSn=127)๋ฅผ ์ฌ์ฉํ๊ธฐ ์ฝ๊ฒ ์ ์ ํ์ต๋๋ค.
- ์ ๊ณต์ฒ : AI-Hub(https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&aihubDataSe=realm&dataSetSn=127)
- ์ ๋ชฉ : ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฌธํ ๋ถ์ผ ์ด์ค ๋ง๋ญ์น
- ๊ตฌ์ถ๋ถ์ผ : ๋ฌธํ์ฌ/ํฅํ /K-Food, K-POP(ํ๋ฅ)/๋์ค๋ฌธํ_๊ณต์ฐ ์ฝํ
์ธ , IT/์ปดํจํฐ/๋ชจ๋ฐ์ผ, ๊ธ์ต/์ฆ์, ์ฌํ/๋
ธ๋/๋ณต์ง, ๊ต์ก, ํนํ/๊ธฐ์ , ์๋์ฐจ
- ๊ตฌ์ถ๋ : 150๋ง ๋ฌธ์ฅ์
- ์์ฉ๋ถ์ผ : ์ธ์ด๋ชจ๋ธ, ์๋๋ฒ์ญ
- ์ธ์ด : ์์์ด-ํ๊ตญ์ด, ๋ชฉ์ ์ด-์ผ๋ณธ์ด
### Supported Tasks and Leaderboards
- Translation
### Languages
- Kor
- Jpan
## Dataset Structure
features:
- name: KOR
dtype: string
- name: JPN
dtype: string
splits:
- name: train
num_bytes: 294787449
num_examples: 840000
- name: val
num_bytes: 88406929
num_examples: 252000
- name: test
num_bytes: 37964427
num_examples: 108000
download_size: 289307354
dataset_size: 421158805
### Data Splits
splits:
- name: train
num_bytes: 294787449
num_examples: 840000
- name: val
num_bytes: 88406929
num_examples: 252000
- name: test
num_bytes: 37964427
num_examples: 108000
### Contributions
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | noahkim/Kor_Jpn_Translation_Dataset | [
"task_categories:translation",
"task_ids:language-modeling",
"annotations_creators:expert-generated",
"language_creators:other",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:kor",
"language:jpn",
"license:mit",
"region:us"
] | 2022-12-20T11:19:57+00:00 | {"annotations_creators": ["expert-generated"], "language_creators": ["other"], "language": ["kor", "jpn"], "license": ["mit"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": ["language-modeling"], "pretty_name": "Kor-Jpn-Translation"} | 2022-12-20T12:03:22+00:00 | [] | [
"kor",
"jpn"
] | TAGS
#task_categories-translation #task_ids-language-modeling #annotations_creators-expert-generated #language_creators-other #size_categories-1K<n<10K #source_datasets-original #language-Korean #language-Japanese #license-mit #region-us
| # Dataset Card for "Kor_Jpn_Translation_Dataset"
### Dataset Summary
AI-Hub์์ ์ ๊ณตํ๋ ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฒ์ญ ๋ง๋ญ์น ๋ฐ์ดํฐ(URL)๋ฅผ ์ฌ์ฉํ๊ธฐ ์ฝ๊ฒ ์ ์ ํ์ต๋๋ค.
- ์ ๊ณต์ฒ : AI-Hub(URL
- ์ ๋ชฉ : ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฌธํ ๋ถ์ผ ์ด์ค ๋ง๋ญ์น
- ๊ตฌ์ถ๋ถ์ผ : ๋ฌธํ์ฌ/ํฅํ /K-Food, K-POP(ํ๋ฅ)/๋์ค๋ฌธํ_๊ณต์ฐ ์ฝํ
์ธ , IT/์ปดํจํฐ/๋ชจ๋ฐ์ผ, ๊ธ์ต/์ฆ์, ์ฌํ/๋
ธ๋/๋ณต์ง, ๊ต์ก, ํนํ/๊ธฐ์ , ์๋์ฐจ
- ๊ตฌ์ถ๋ : 150๋ง ๋ฌธ์ฅ์
- ์์ฉ๋ถ์ผ : ์ธ์ด๋ชจ๋ธ, ์๋๋ฒ์ญ
- ์ธ์ด : ์์์ด-ํ๊ตญ์ด, ๋ชฉ์ ์ด-์ผ๋ณธ์ด
### Supported Tasks and Leaderboards
- Translation
### Languages
- Kor
- Jpan
## Dataset Structure
features:
- name: KOR
dtype: string
- name: JPN
dtype: string
splits:
- name: train
num_bytes: 294787449
num_examples: 840000
- name: val
num_bytes: 88406929
num_examples: 252000
- name: test
num_bytes: 37964427
num_examples: 108000
download_size: 289307354
dataset_size: 421158805
### Data Splits
splits:
- name: train
num_bytes: 294787449
num_examples: 840000
- name: val
num_bytes: 88406929
num_examples: 252000
- name: test
num_bytes: 37964427
num_examples: 108000
### Contributions
More Information needed | [
"# Dataset Card for \"Kor_Jpn_Translation_Dataset\"",
"### Dataset Summary\n\nAI-Hub์์ ์ ๊ณตํ๋ ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฒ์ญ ๋ง๋ญ์น ๋ฐ์ดํฐ(URL)๋ฅผ ์ฌ์ฉํ๊ธฐ ์ฝ๊ฒ ์ ์ ํ์ต๋๋ค.\n\n- ์ ๊ณต์ฒ : AI-Hub(URL\n- ์ ๋ชฉ : ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฌธํ ๋ถ์ผ ์ด์ค ๋ง๋ญ์น\n- ๊ตฌ์ถ๋ถ์ผ : ๋ฌธํ์ฌ/ํฅํ /K-Food, K-POP(ํ๋ฅ)/๋์ค๋ฌธํ_๊ณต์ฐ ์ฝํ
์ธ , IT/์ปดํจํฐ/๋ชจ๋ฐ์ผ, ๊ธ์ต/์ฆ์, ์ฌํ/๋
ธ๋/๋ณต์ง, ๊ต์ก, ํนํ/๊ธฐ์ , ์๋์ฐจ\n- ๊ตฌ์ถ๋ : 150๋ง ๋ฌธ์ฅ์\n- ์์ฉ๋ถ์ผ : ์ธ์ด๋ชจ๋ธ, ์๋๋ฒ์ญ\n- ์ธ์ด : ์์์ด-ํ๊ตญ์ด, ๋ชฉ์ ์ด-์ผ๋ณธ์ด",
"### Supported Tasks and Leaderboards\n\n- Translation",
"### Languages\n\n- Kor\n- Jpan",
"## Dataset Structure\n\n features:\n - name: KOR\n dtype: string\n - name: JPN\n dtype: string\n splits:\n - name: train\n num_bytes: 294787449\n num_examples: 840000\n - name: val\n num_bytes: 88406929\n num_examples: 252000\n - name: test\n num_bytes: 37964427\n num_examples: 108000\n download_size: 289307354\n dataset_size: 421158805",
"### Data Splits\n\n splits:\n - name: train\n num_bytes: 294787449\n num_examples: 840000\n - name: val\n num_bytes: 88406929\n num_examples: 252000\n - name: test\n num_bytes: 37964427\n num_examples: 108000",
"### Contributions\nMore Information needed"
] | [
"TAGS\n#task_categories-translation #task_ids-language-modeling #annotations_creators-expert-generated #language_creators-other #size_categories-1K<n<10K #source_datasets-original #language-Korean #language-Japanese #license-mit #region-us \n",
"# Dataset Card for \"Kor_Jpn_Translation_Dataset\"",
"### Dataset Summary\n\nAI-Hub์์ ์ ๊ณตํ๋ ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฒ์ญ ๋ง๋ญ์น ๋ฐ์ดํฐ(URL)๋ฅผ ์ฌ์ฉํ๊ธฐ ์ฝ๊ฒ ์ ์ ํ์ต๋๋ค.\n\n- ์ ๊ณต์ฒ : AI-Hub(URL\n- ์ ๋ชฉ : ํ๊ตญ์ด-์ผ๋ณธ์ด ๋ฌธํ ๋ถ์ผ ์ด์ค ๋ง๋ญ์น\n- ๊ตฌ์ถ๋ถ์ผ : ๋ฌธํ์ฌ/ํฅํ /K-Food, K-POP(ํ๋ฅ)/๋์ค๋ฌธํ_๊ณต์ฐ ์ฝํ
์ธ , IT/์ปดํจํฐ/๋ชจ๋ฐ์ผ, ๊ธ์ต/์ฆ์, ์ฌํ/๋
ธ๋/๋ณต์ง, ๊ต์ก, ํนํ/๊ธฐ์ , ์๋์ฐจ\n- ๊ตฌ์ถ๋ : 150๋ง ๋ฌธ์ฅ์\n- ์์ฉ๋ถ์ผ : ์ธ์ด๋ชจ๋ธ, ์๋๋ฒ์ญ\n- ์ธ์ด : ์์์ด-ํ๊ตญ์ด, ๋ชฉ์ ์ด-์ผ๋ณธ์ด",
"### Supported Tasks and Leaderboards\n\n- Translation",
"### Languages\n\n- Kor\n- Jpan",
"## Dataset Structure\n\n features:\n - name: KOR\n dtype: string\n - name: JPN\n dtype: string\n splits:\n - name: train\n num_bytes: 294787449\n num_examples: 840000\n - name: val\n num_bytes: 88406929\n num_examples: 252000\n - name: test\n num_bytes: 37964427\n num_examples: 108000\n download_size: 289307354\n dataset_size: 421158805",
"### Data Splits\n\n splits:\n - name: train\n num_bytes: 294787449\n num_examples: 840000\n - name: val\n num_bytes: 88406929\n num_examples: 252000\n - name: test\n num_bytes: 37964427\n num_examples: 108000",
"### Contributions\nMore Information needed"
] |
4d8aad0c2f2625bf60af171949633ad76ca0b921 |
# Dataset Card for ScandiReddit
## Dataset Description
- **Repository:** <https://github.com/alexandrainst/ScandiReddit>
- **Point of Contact:** [Dan Saattrup Nielsen](mailto:[email protected])
- **Size of downloaded dataset files:** 2341 MB
- **Size of the generated dataset:** 3594 MB
- **Total amount of disk used:** 5935 MB
### Dataset Summary
ScandiReddit is a filtered and post-processed corpus consisting of comments from [Reddit](https://reddit.com/).
All Reddit comments from December 2005 up until October 2022 were downloaded through [PushShift](https://files.pushshift.io/reddit/comments/), after which these were filtered based on the FastText language detection model. Any comment which was classified as Danish (`da`), Norwegian (`no`), Swedish (`sv`) or Icelandic (`is`) with a confidence score above 70% was kept.
The resulting comments were then deduplicated, removing roughly 438,000 comments. 5,000 comments written by Reddit bots were removed, and roughly 189,000 comments belonging to inappropriate subreddits (explicit and drug-related) were also removed.
Lastly, we remove roughly 40,000 near-duplicate comments from the resulting corpus, where near-duplicate here means that the comments have more than 80% of their word 5-grams in common.
### Supported Tasks and Leaderboards
Training language models is the intended task for this dataset. No leaderboard is active at this point.
### Languages
The dataset is available in Danish (`da`), Swedish (`sv`), Norwegian (`no`) and Icelandic (`is`).
## Dataset Structure
### Data Instances
- **Size of downloaded dataset files:** 2341 MB
- **Size of the generated dataset:** 3594 MB
- **Total amount of disk used:** 5935 MB
An example from the dataset looks as follows.
```
{
'doc': 'Bergen er รธdelagt. Det er ikke moro mer.',
'subreddit': 'Norway',
'language': 'da',
'language_confidence': 0.7472341656684875
}
```
### Data Fields
The data fields are the same among all splits.
- `doc`: a `string` feature.
- `subreddit`: a `string` feature.
- `language`: a `string` feature.
- `language_confidence`: a `float64` feature.
### Language Distribution
| name | count |
|----------|---------:|
| sv | 6,967,420 |
| da | 4,965,195 |
| no | 1,340,470 |
| is | 206,689 |
| total | 13,479,774 |
### Top-50 Subreddit Distribution
| name | count |
|----------|--------:|
|sweden |4,881,483|
|Denmark |3,579,178|
|norge |1,281,655|
|svenskpolitik | 771,960|
|InfluencergossipDK | 649,910|
|swedishproblems | 339,683|
|Iceland | 183,488|
|dkfinance | 113,860|
|unket | 81,077|
|DanishEnts | 69,055|
|dankmark | 62,928|
|swedents | 58,576|
|scandinavia | 57,136|
|Allsvenskan | 56,006|
|Gothenburg | 54,395|
|stockholm | 51,016|
|ISKbets | 47,944|
|Sverige | 39,552|
|SWARJE | 34,691|
|GossipDK | 29,332|
|NorskFotball | 28,571|
|Superligaen | 23,641|
|Aarhus | 22,516|
|Svenska | 20,561|
|newsdk | 19,893|
|AskReddit | 16,672|
|copenhagen | 16,668|
|okpolarncp | 16,583|
|SwedditUniversalis | 15,990|
|Sveriges_politik | 15,058|
|intresseklubben | 13,246|
|Aktiemarknaden | 13,202|
|soccer | 12,637|
|teenagers | 10,845|
|Norway | 10,680|
|europe | 10,247|
|Matinbum | 9,792|
|oslo | 9,650|
|iksdagen | 9,232|
|Asksweddit | 8,851|
|Forsvaret | 8,641|
|Sverigesforsvarsmakt | 8,469|
|memes | 8,299|
|Danish | 8,268|
|DANMAG | 8,214|
|PewdiepieSubmissions | 7,800|
|sweddpolitik | 7,646|
|pinsamt | 7,318|
|arbetarrorelsen | 7,317|
|Ishockey | 6,824|
## Dataset Creation
### Curation Rationale
The Scandinavian languages do not have many open source social media datasets.
### Source Data
The raw Reddit data was collected through [PushShift](https://files.pushshift.io/reddit/comments/).
## Additional Information
### Dataset Curators
[Dan Saattrup Nielsen](https://saattrupdan.github.io/) from the [The Alexandra
Institute](https://alexandra.dk/) curated this dataset.
### Licensing Information
The dataset is licensed under the [CC BY 4.0
license](https://creativecommons.org/licenses/by/4.0/).
| alexandrainst/scandi-reddit | [
"task_categories:text-generation",
"task_categories:fill-mask",
"task_ids:language-modeling",
"multilinguality:multilingual",
"size_categories:10M<n<100M",
"language:da",
"language:sv",
"language:no",
"language:is",
"license:cc-by-4.0",
"region:us"
] | 2022-12-20T12:13:19+00:00 | {"language": ["da", "sv", false, "is"], "license": ["cc-by-4.0"], "multilinguality": ["multilingual"], "size_categories": ["10M<n<100M"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling"], "pretty_name": "ScandiReddit"} | 2022-12-21T17:54:31+00:00 | [] | [
"da",
"sv",
"no",
"is"
] | TAGS
#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #multilinguality-multilingual #size_categories-10M<n<100M #language-Danish #language-Swedish #language-Norwegian #language-Icelandic #license-cc-by-4.0 #region-us
| Dataset Card for ScandiReddit
=============================
Dataset Description
-------------------
* Repository: <URL
* Point of Contact: Dan Saattrup Nielsen
* Size of downloaded dataset files: 2341 MB
* Size of the generated dataset: 3594 MB
* Total amount of disk used: 5935 MB
### Dataset Summary
ScandiReddit is a filtered and post-processed corpus consisting of comments from Reddit.
All Reddit comments from December 2005 up until October 2022 were downloaded through PushShift, after which these were filtered based on the FastText language detection model. Any comment which was classified as Danish ('da'), Norwegian ('no'), Swedish ('sv') or Icelandic ('is') with a confidence score above 70% was kept.
The resulting comments were then deduplicated, removing roughly 438,000 comments. 5,000 comments written by Reddit bots were removed, and roughly 189,000 comments belonging to inappropriate subreddits (explicit and drug-related) were also removed.
Lastly, we remove roughly 40,000 near-duplicate comments from the resulting corpus, where near-duplicate here means that the comments have more than 80% of their word 5-grams in common.
### Supported Tasks and Leaderboards
Training language models is the intended task for this dataset. No leaderboard is active at this point.
### Languages
The dataset is available in Danish ('da'), Swedish ('sv'), Norwegian ('no') and Icelandic ('is').
Dataset Structure
-----------------
### Data Instances
* Size of downloaded dataset files: 2341 MB
* Size of the generated dataset: 3594 MB
* Total amount of disk used: 5935 MB
An example from the dataset looks as follows.
### Data Fields
The data fields are the same among all splits.
* 'doc': a 'string' feature.
* 'subreddit': a 'string' feature.
* 'language': a 'string' feature.
* 'language\_confidence': a 'float64' feature.
### Language Distribution
### Top-50 Subreddit Distribution
Dataset Creation
----------------
### Curation Rationale
The Scandinavian languages do not have many open source social media datasets.
### Source Data
The raw Reddit data was collected through PushShift.
Additional Information
----------------------
### Dataset Curators
Dan Saattrup Nielsen from the The Alexandra
Institute curated this dataset.
### Licensing Information
The dataset is licensed under the CC BY 4.0
license.
| [
"### Dataset Summary\n\n\nScandiReddit is a filtered and post-processed corpus consisting of comments from Reddit.\n\n\nAll Reddit comments from December 2005 up until October 2022 were downloaded through PushShift, after which these were filtered based on the FastText language detection model. Any comment which was classified as Danish ('da'), Norwegian ('no'), Swedish ('sv') or Icelandic ('is') with a confidence score above 70% was kept.\n\n\nThe resulting comments were then deduplicated, removing roughly 438,000 comments. 5,000 comments written by Reddit bots were removed, and roughly 189,000 comments belonging to inappropriate subreddits (explicit and drug-related) were also removed.\n\n\nLastly, we remove roughly 40,000 near-duplicate comments from the resulting corpus, where near-duplicate here means that the comments have more than 80% of their word 5-grams in common.",
"### Supported Tasks and Leaderboards\n\n\nTraining language models is the intended task for this dataset. No leaderboard is active at this point.",
"### Languages\n\n\nThe dataset is available in Danish ('da'), Swedish ('sv'), Norwegian ('no') and Icelandic ('is').\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\n* Size of downloaded dataset files: 2341 MB\n* Size of the generated dataset: 3594 MB\n* Total amount of disk used: 5935 MB\n\n\nAn example from the dataset looks as follows.",
"### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'doc': a 'string' feature.\n* 'subreddit': a 'string' feature.\n* 'language': a 'string' feature.\n* 'language\\_confidence': a 'float64' feature.",
"### Language Distribution",
"### Top-50 Subreddit Distribution\n\n\n\nDataset Creation\n----------------",
"### Curation Rationale\n\n\nThe Scandinavian languages do not have many open source social media datasets.",
"### Source Data\n\n\nThe raw Reddit data was collected through PushShift.\n\n\nAdditional Information\n----------------------",
"### Dataset Curators\n\n\nDan Saattrup Nielsen from the The Alexandra\nInstitute curated this dataset.",
"### Licensing Information\n\n\nThe dataset is licensed under the CC BY 4.0\nlicense."
] | [
"TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #multilinguality-multilingual #size_categories-10M<n<100M #language-Danish #language-Swedish #language-Norwegian #language-Icelandic #license-cc-by-4.0 #region-us \n",
"### Dataset Summary\n\n\nScandiReddit is a filtered and post-processed corpus consisting of comments from Reddit.\n\n\nAll Reddit comments from December 2005 up until October 2022 were downloaded through PushShift, after which these were filtered based on the FastText language detection model. Any comment which was classified as Danish ('da'), Norwegian ('no'), Swedish ('sv') or Icelandic ('is') with a confidence score above 70% was kept.\n\n\nThe resulting comments were then deduplicated, removing roughly 438,000 comments. 5,000 comments written by Reddit bots were removed, and roughly 189,000 comments belonging to inappropriate subreddits (explicit and drug-related) were also removed.\n\n\nLastly, we remove roughly 40,000 near-duplicate comments from the resulting corpus, where near-duplicate here means that the comments have more than 80% of their word 5-grams in common.",
"### Supported Tasks and Leaderboards\n\n\nTraining language models is the intended task for this dataset. No leaderboard is active at this point.",
"### Languages\n\n\nThe dataset is available in Danish ('da'), Swedish ('sv'), Norwegian ('no') and Icelandic ('is').\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\n* Size of downloaded dataset files: 2341 MB\n* Size of the generated dataset: 3594 MB\n* Total amount of disk used: 5935 MB\n\n\nAn example from the dataset looks as follows.",
"### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'doc': a 'string' feature.\n* 'subreddit': a 'string' feature.\n* 'language': a 'string' feature.\n* 'language\\_confidence': a 'float64' feature.",
"### Language Distribution",
"### Top-50 Subreddit Distribution\n\n\n\nDataset Creation\n----------------",
"### Curation Rationale\n\n\nThe Scandinavian languages do not have many open source social media datasets.",
"### Source Data\n\n\nThe raw Reddit data was collected through PushShift.\n\n\nAdditional Information\n----------------------",
"### Dataset Curators\n\n\nDan Saattrup Nielsen from the The Alexandra\nInstitute curated this dataset.",
"### Licensing Information\n\n\nThe dataset is licensed under the CC BY 4.0\nlicense."
] |
1273da4d03c607109fe9575e6aaf6063cb044988 |
# About Dataset
Philosophers Quotes from azquotes.com
* Arthur Schopenhauer 400+ quotes
* Friedrich Nietzsche 200+ quotes
* Immanuel Kant 300+ quotes
* Aristotle 350+ quotes
* Plato 70+ quotes
* Sigmund Freud 400+ quotes
* Hegel 120+ quotes
* Jean Paul Sartre 320+ quotes
* Spinoza 120+ quotes
### COLLECTION METHODOLOGY
Python Web Scraping with Selenium | mertbozkurt/quotes_philosophers | [
"license:afl-3.0",
"region:us"
] | 2022-12-20T13:13:09+00:00 | {"license": "afl-3.0"} | 2022-12-20T13:17:12+00:00 | [] | [] | TAGS
#license-afl-3.0 #region-us
|
# About Dataset
Philosophers Quotes from URL
* Arthur Schopenhauer 400+ quotes
* Friedrich Nietzsche 200+ quotes
* Immanuel Kant 300+ quotes
* Aristotle 350+ quotes
* Plato 70+ quotes
* Sigmund Freud 400+ quotes
* Hegel 120+ quotes
* Jean Paul Sartre 320+ quotes
* Spinoza 120+ quotes
### COLLECTION METHODOLOGY
Python Web Scraping with Selenium | [
"# About Dataset\nPhilosophers Quotes from URL \n\n* Arthur Schopenhauer 400+ quotes\n* Friedrich Nietzsche 200+ quotes\n* Immanuel Kant 300+ quotes\n* Aristotle 350+ quotes\n* Plato 70+ quotes\n* Sigmund Freud 400+ quotes\n* Hegel 120+ quotes\n* Jean Paul Sartre 320+ quotes\n* Spinoza 120+ quotes",
"### COLLECTION METHODOLOGY\n Python Web Scraping with Selenium"
] | [
"TAGS\n#license-afl-3.0 #region-us \n",
"# About Dataset\nPhilosophers Quotes from URL \n\n* Arthur Schopenhauer 400+ quotes\n* Friedrich Nietzsche 200+ quotes\n* Immanuel Kant 300+ quotes\n* Aristotle 350+ quotes\n* Plato 70+ quotes\n* Sigmund Freud 400+ quotes\n* Hegel 120+ quotes\n* Jean Paul Sartre 320+ quotes\n* Spinoza 120+ quotes",
"### COLLECTION METHODOLOGY\n Python Web Scraping with Selenium"
] |
1454bcf9d92442c5d5d0dc9004315010950061e4 | # Dataset Card for "github-issues"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | threite/github-issues | [
"region:us"
] | 2022-12-20T13:18:01+00:00 | {"dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "repository_url", "dtype": "string"}, {"name": "labels_url", "dtype": "string"}, {"name": "comments_url", "dtype": "string"}, {"name": "events_url", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "node_id", "dtype": "string"}, {"name": "number", "dtype": "int64"}, {"name": "title", "dtype": "string"}, {"name": "user", "struct": [{"name": "avatar_url", "dtype": "string"}, {"name": "events_url", "dtype": "string"}, {"name": "followers_url", "dtype": "string"}, {"name": "following_url", "dtype": "string"}, {"name": "gists_url", "dtype": "string"}, {"name": "gravatar_id", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "login", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "organizations_url", "dtype": "string"}, {"name": "received_events_url", "dtype": "string"}, {"name": "repos_url", "dtype": "string"}, {"name": "site_admin", "dtype": "bool"}, {"name": "starred_url", "dtype": "string"}, {"name": "subscriptions_url", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "labels", "list": [{"name": "color", "dtype": "string"}, {"name": "default", "dtype": "bool"}, {"name": "description", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "name", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "state", "dtype": "string"}, {"name": "locked", "dtype": "bool"}, {"name": "assignee", "struct": [{"name": "avatar_url", "dtype": "string"}, {"name": "events_url", "dtype": "string"}, {"name": "followers_url", "dtype": "string"}, {"name": "following_url", "dtype": "string"}, {"name": "gists_url", "dtype": "string"}, {"name": "gravatar_id", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "login", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "organizations_url", "dtype": "string"}, {"name": "received_events_url", "dtype": "string"}, {"name": "repos_url", "dtype": "string"}, {"name": "site_admin", "dtype": "bool"}, {"name": "starred_url", "dtype": "string"}, {"name": "subscriptions_url", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "assignees", "list": [{"name": "avatar_url", "dtype": "string"}, {"name": "events_url", "dtype": "string"}, {"name": "followers_url", "dtype": "string"}, {"name": "following_url", "dtype": "string"}, {"name": "gists_url", "dtype": "string"}, {"name": "gravatar_id", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "login", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "organizations_url", "dtype": "string"}, {"name": "received_events_url", "dtype": "string"}, {"name": "repos_url", "dtype": "string"}, {"name": "site_admin", "dtype": "bool"}, {"name": "starred_url", "dtype": "string"}, {"name": "subscriptions_url", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "milestone", "struct": [{"name": "closed_at", "dtype": "string"}, {"name": "closed_issues", "dtype": "int64"}, {"name": "created_at", "dtype": "string"}, {"name": "creator", "struct": [{"name": "avatar_url", "dtype": "string"}, {"name": "events_url", "dtype": "string"}, {"name": "followers_url", "dtype": "string"}, {"name": "following_url", "dtype": "string"}, {"name": "gists_url", "dtype": "string"}, {"name": "gravatar_id", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "login", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "organizations_url", "dtype": "string"}, {"name": "received_events_url", "dtype": "string"}, {"name": "repos_url", "dtype": "string"}, {"name": "site_admin", "dtype": "bool"}, {"name": "starred_url", "dtype": "string"}, {"name": "subscriptions_url", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "description", "dtype": "string"}, {"name": "due_on", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "id", "dtype": "int64"}, {"name": "labels_url", "dtype": "string"}, {"name": "node_id", "dtype": "string"}, {"name": "number", "dtype": "int64"}, {"name": "open_issues", "dtype": "int64"}, {"name": "state", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "updated_at", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "comments", "sequence": "null"}, {"name": "created_at", "dtype": "string"}, {"name": "updated_at", "dtype": "string"}, {"name": "closed_at", "dtype": "string"}, {"name": "author_association", "dtype": "string"}, {"name": "active_lock_reason", "dtype": "null"}, {"name": "draft", "dtype": "bool"}, {"name": "pull_request", "struct": [{"name": "diff_url", "dtype": "string"}, {"name": "html_url", "dtype": "string"}, {"name": "merged_at", "dtype": "string"}, {"name": "patch_url", "dtype": "string"}, {"name": "url", "dtype": "string"}]}, {"name": "body", "dtype": "string"}, {"name": "reactions", "struct": [{"name": "+1", "dtype": "int64"}, {"name": "-1", "dtype": "int64"}, {"name": "confused", "dtype": "int64"}, {"name": "eyes", "dtype": "int64"}, {"name": "heart", "dtype": "int64"}, {"name": "hooray", "dtype": "int64"}, {"name": "laugh", "dtype": "int64"}, {"name": "rocket", "dtype": "int64"}, {"name": "total_count", "dtype": "int64"}, {"name": "url", "dtype": "string"}]}, {"name": "timeline_url", "dtype": "string"}, {"name": "performed_via_github_app", "dtype": "null"}, {"name": "state_reason", "dtype": "string"}, {"name": "is_pull_request", "dtype": "bool"}], "splits": [{"name": "train", "num_bytes": 16275865, "num_examples": 5392}], "download_size": 3809038, "dataset_size": 16275865}} | 2022-12-20T13:18:23+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "github-issues"
More Information needed | [
"# Dataset Card for \"github-issues\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"github-issues\"\n\nMore Information needed"
] |
4c98780bd4228a273bf5d240e2ccee699dc41825 | # Dataset Card for "slue-voxceleb"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | qmeeus/slue-voxceleb | [
"region:us"
] | 2022-12-20T13:24:19+00:00 | {"dataset_info": {"features": [{"name": "n_frames", "dtype": "int64"}, {"name": "text", "dtype": "string"}, {"name": "sentiment", "dtype": {"class_label": {"names": {"0": "Negative", "1": "Neutral", "2": "Positive"}}}}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}], "splits": [{"name": "train", "num_bytes": 756214858.625, "num_examples": 5729}, {"name": "dev", "num_bytes": 130698641.0, "num_examples": 954}], "download_size": 949197313, "dataset_size": 886913499.625}} | 2022-12-20T13:25:43+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "slue-voxceleb"
More Information needed | [
"# Dataset Card for \"slue-voxceleb\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"slue-voxceleb\"\n\nMore Information needed"
] |
e2a26bd1bc6d1ff15a33ab9d2ed118054151a784 |
# Dataset Card for Czech Simple Question Answering Dataset 3.0
This a processed and filtered adaptation of an existing dataset. For raw and larger dataset, see `Dataset Source` section.
## Dataset Description
The data contains questions and answers based on Czech wikipeadia articles.
Each question has an answer (or more) and a selected part of the context as the evidence.
A majority of the answers are extractive - i.e. they are present in the context in the exact form. The remaining cases are
- yes/no questions
- answer is almost in the exact form present in the text, but the form of words was changed to suit the question (declension, ...)
- answered in own words (should be rare, but is not)
All questions in the dataset are answerable from the context. Small minority of questions have multiple answers.
Sometimes it means that any of them is correct (e.g. either "Pacifik" or "Tichรฝ oceรกn" are correct terms for Pacific Ocean)
and sometimes it means that all of them together are a correct answer (e.g., Who was Leonardo da Vinci? ["painter", "engineer"])
Total number of examples is around:
- 6,250 in train
- 570 in validation
- 850 in test.
## Dataset Features
Each example contains:
- `item_id`: string id of the
- `context`: "reasonably" big chunk (string) of wikipedia article that contains the answer
- `question`: string
- `answers`: list of all answers (string). mostly list of length 1
- `evidence_text`: substring of context (typically one sentence) that is sufficient to answer the question
- `evidence_start`: index in context, such that `context[evidence_start:evidence_end] == evidence_text`
- `evidence_end`: index in context
- `occurences`:
list of (dictionaries) occurences of the answer(s) in the evidence.
Each answer was searched with word boundaries ("\b" in regex) and case-sensitive in the evidence.
If nothing found, try again but case-insensitive.
If nothing found, try again but case-sensitive without word boundaries.
If nothing found, try again but case-insensitive without word boundaries.
This process should supress "false positive" occurences of the answer in the evidence.
- `start`: index in context
- `end`: index in context
- `text`: the answer looked for
- `url`: link to the wikipedia article
- `original_article`: original parsed wikipedia article from which the context is taken
- `question_type`: type of the question, one of: ['ABBREVIATION', 'DATETIME', 'DENOTATION', 'ENTITY', 'LOCATION', 'NUMERIC', 'ORGANIZATION', 'OTHER', 'PERSON', 'YES_NO']
- `answer_type`: type of the answer, one of: ['ABBREVIATION', 'ADJ_PHRASE', 'CLAUSE', 'DATETIME', 'ENTITY', 'LOCATION', 'NUMERIC', 'OTHER', 'PERSON', 'VERB_PHRASE']
## Dataset Source
The dataset is a preprocessed adaptation of existing SQAD 3.0 dataset [link to data](https://lindat.cz/repository/xmlui/handle/11234/1-3069).
This adaptation contains (almost) same data, but converted to a convenient format.
The data was also filtered to remove a statistical bias where the answer was contained
in the first sentence in the article (around 50% of all data in the original dataset, likely
caused by the data collection process).
## Citation
Cite authors of the [original dataset](https://lindat.cz/repository/xmlui/handle/11234/1-3069):
```bibtex
@misc{11234/1-3069,
title = {sqad 3.0},
author = {Medve{\v d}, Marek and Hor{\'a}k, Ale{\v s}},
url = {http://hdl.handle.net/11234/1-3069},
note = {{LINDAT}/{CLARIAH}-{CZ} digital library at the Institute of Formal and Applied Linguistics ({{\'U}FAL}), Faculty of Mathematics and Physics, Charles University},
copyright = {{GNU} Library or "Lesser" General Public License 3.0 ({LGPL}-3.0)},
year = {2019}
}
```
| fewshot-goes-multilingual/cs_squad-3.0 | [
"task_categories:question-answering",
"task_ids:extractive-qa",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:cs",
"license:lgpl-3.0",
"czech QA",
"wikipedia QA",
"region:us"
] | 2022-12-20T13:50:51+00:00 | {"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["cs"], "license": ["lgpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa"], "pretty_name": "Czech Simple Question Answering Dataset", "tags": ["czech QA", "wikipedia QA"]} | 2023-11-26T20:42:44+00:00 | [] | [
"cs"
] | TAGS
#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Czech #license-lgpl-3.0 #czech QA #wikipedia QA #region-us
|
# Dataset Card for Czech Simple Question Answering Dataset 3.0
This a processed and filtered adaptation of an existing dataset. For raw and larger dataset, see 'Dataset Source' section.
## Dataset Description
The data contains questions and answers based on Czech wikipeadia articles.
Each question has an answer (or more) and a selected part of the context as the evidence.
A majority of the answers are extractive - i.e. they are present in the context in the exact form. The remaining cases are
- yes/no questions
- answer is almost in the exact form present in the text, but the form of words was changed to suit the question (declension, ...)
- answered in own words (should be rare, but is not)
All questions in the dataset are answerable from the context. Small minority of questions have multiple answers.
Sometimes it means that any of them is correct (e.g. either "Pacifik" or "Tichรฝ oceรกn" are correct terms for Pacific Ocean)
and sometimes it means that all of them together are a correct answer (e.g., Who was Leonardo da Vinci? ["painter", "engineer"])
Total number of examples is around:
- 6,250 in train
- 570 in validation
- 850 in test.
## Dataset Features
Each example contains:
- 'item_id': string id of the
- 'context': "reasonably" big chunk (string) of wikipedia article that contains the answer
- 'question': string
- 'answers': list of all answers (string). mostly list of length 1
- 'evidence_text': substring of context (typically one sentence) that is sufficient to answer the question
- 'evidence_start': index in context, such that 'context[evidence_start:evidence_end] == evidence_text'
- 'evidence_end': index in context
- 'occurences':
list of (dictionaries) occurences of the answer(s) in the evidence.
Each answer was searched with word boundaries ("\b" in regex) and case-sensitive in the evidence.
If nothing found, try again but case-insensitive.
If nothing found, try again but case-sensitive without word boundaries.
If nothing found, try again but case-insensitive without word boundaries.
This process should supress "false positive" occurences of the answer in the evidence.
- 'start': index in context
- 'end': index in context
- 'text': the answer looked for
- 'url': link to the wikipedia article
- 'original_article': original parsed wikipedia article from which the context is taken
- 'question_type': type of the question, one of: ['ABBREVIATION', 'DATETIME', 'DENOTATION', 'ENTITY', 'LOCATION', 'NUMERIC', 'ORGANIZATION', 'OTHER', 'PERSON', 'YES_NO']
- 'answer_type': type of the answer, one of: ['ABBREVIATION', 'ADJ_PHRASE', 'CLAUSE', 'DATETIME', 'ENTITY', 'LOCATION', 'NUMERIC', 'OTHER', 'PERSON', 'VERB_PHRASE']
## Dataset Source
The dataset is a preprocessed adaptation of existing SQAD 3.0 dataset link to data.
This adaptation contains (almost) same data, but converted to a convenient format.
The data was also filtered to remove a statistical bias where the answer was contained
in the first sentence in the article (around 50% of all data in the original dataset, likely
caused by the data collection process).
Cite authors of the original dataset:
| [
"# Dataset Card for Czech Simple Question Answering Dataset 3.0\n\nThis a processed and filtered adaptation of an existing dataset. For raw and larger dataset, see 'Dataset Source' section.",
"## Dataset Description\nThe data contains questions and answers based on Czech wikipeadia articles.\nEach question has an answer (or more) and a selected part of the context as the evidence.\nA majority of the answers are extractive - i.e. they are present in the context in the exact form. The remaining cases are\n\n- yes/no questions\n- answer is almost in the exact form present in the text, but the form of words was changed to suit the question (declension, ...)\n- answered in own words (should be rare, but is not)\n\nAll questions in the dataset are answerable from the context. Small minority of questions have multiple answers.\nSometimes it means that any of them is correct (e.g. either \"Pacifik\" or \"Tichรฝ oceรกn\" are correct terms for Pacific Ocean)\nand sometimes it means that all of them together are a correct answer (e.g., Who was Leonardo da Vinci? [\"painter\", \"engineer\"])\n\nTotal number of examples is around:\n\n- 6,250 in train\n- 570 in validation\n- 850 in test.",
"## Dataset Features\nEach example contains:\n- 'item_id': string id of the\n- 'context': \"reasonably\" big chunk (string) of wikipedia article that contains the answer\n- 'question': string\n- 'answers': list of all answers (string). mostly list of length 1\n- 'evidence_text': substring of context (typically one sentence) that is sufficient to answer the question\n- 'evidence_start': index in context, such that 'context[evidence_start:evidence_end] == evidence_text'\n- 'evidence_end': index in context\n- 'occurences':\n list of (dictionaries) occurences of the answer(s) in the evidence.\n Each answer was searched with word boundaries (\"\\b\" in regex) and case-sensitive in the evidence.\n If nothing found, try again but case-insensitive.\n If nothing found, try again but case-sensitive without word boundaries.\n If nothing found, try again but case-insensitive without word boundaries.\n This process should supress \"false positive\" occurences of the answer in the evidence.\n - 'start': index in context\n - 'end': index in context\n - 'text': the answer looked for\n- 'url': link to the wikipedia article\n- 'original_article': original parsed wikipedia article from which the context is taken\n- 'question_type': type of the question, one of: ['ABBREVIATION', 'DATETIME', 'DENOTATION', 'ENTITY', 'LOCATION', 'NUMERIC', 'ORGANIZATION', 'OTHER', 'PERSON', 'YES_NO']\n- 'answer_type': type of the answer, one of: ['ABBREVIATION', 'ADJ_PHRASE', 'CLAUSE', 'DATETIME', 'ENTITY', 'LOCATION', 'NUMERIC', 'OTHER', 'PERSON', 'VERB_PHRASE']",
"## Dataset Source\n\nThe dataset is a preprocessed adaptation of existing SQAD 3.0 dataset link to data.\nThis adaptation contains (almost) same data, but converted to a convenient format.\nThe data was also filtered to remove a statistical bias where the answer was contained\nin the first sentence in the article (around 50% of all data in the original dataset, likely\ncaused by the data collection process).\n\n\nCite authors of the original dataset:"
] | [
"TAGS\n#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Czech #license-lgpl-3.0 #czech QA #wikipedia QA #region-us \n",
"# Dataset Card for Czech Simple Question Answering Dataset 3.0\n\nThis a processed and filtered adaptation of an existing dataset. For raw and larger dataset, see 'Dataset Source' section.",
"## Dataset Description\nThe data contains questions and answers based on Czech wikipeadia articles.\nEach question has an answer (or more) and a selected part of the context as the evidence.\nA majority of the answers are extractive - i.e. they are present in the context in the exact form. The remaining cases are\n\n- yes/no questions\n- answer is almost in the exact form present in the text, but the form of words was changed to suit the question (declension, ...)\n- answered in own words (should be rare, but is not)\n\nAll questions in the dataset are answerable from the context. Small minority of questions have multiple answers.\nSometimes it means that any of them is correct (e.g. either \"Pacifik\" or \"Tichรฝ oceรกn\" are correct terms for Pacific Ocean)\nand sometimes it means that all of them together are a correct answer (e.g., Who was Leonardo da Vinci? [\"painter\", \"engineer\"])\n\nTotal number of examples is around:\n\n- 6,250 in train\n- 570 in validation\n- 850 in test.",
"## Dataset Features\nEach example contains:\n- 'item_id': string id of the\n- 'context': \"reasonably\" big chunk (string) of wikipedia article that contains the answer\n- 'question': string\n- 'answers': list of all answers (string). mostly list of length 1\n- 'evidence_text': substring of context (typically one sentence) that is sufficient to answer the question\n- 'evidence_start': index in context, such that 'context[evidence_start:evidence_end] == evidence_text'\n- 'evidence_end': index in context\n- 'occurences':\n list of (dictionaries) occurences of the answer(s) in the evidence.\n Each answer was searched with word boundaries (\"\\b\" in regex) and case-sensitive in the evidence.\n If nothing found, try again but case-insensitive.\n If nothing found, try again but case-sensitive without word boundaries.\n If nothing found, try again but case-insensitive without word boundaries.\n This process should supress \"false positive\" occurences of the answer in the evidence.\n - 'start': index in context\n - 'end': index in context\n - 'text': the answer looked for\n- 'url': link to the wikipedia article\n- 'original_article': original parsed wikipedia article from which the context is taken\n- 'question_type': type of the question, one of: ['ABBREVIATION', 'DATETIME', 'DENOTATION', 'ENTITY', 'LOCATION', 'NUMERIC', 'ORGANIZATION', 'OTHER', 'PERSON', 'YES_NO']\n- 'answer_type': type of the answer, one of: ['ABBREVIATION', 'ADJ_PHRASE', 'CLAUSE', 'DATETIME', 'ENTITY', 'LOCATION', 'NUMERIC', 'OTHER', 'PERSON', 'VERB_PHRASE']",
"## Dataset Source\n\nThe dataset is a preprocessed adaptation of existing SQAD 3.0 dataset link to data.\nThis adaptation contains (almost) same data, but converted to a convenient format.\nThe data was also filtered to remove a statistical bias where the answer was contained\nin the first sentence in the article (around 50% of all data in the original dataset, likely\ncaused by the data collection process).\n\n\nCite authors of the original dataset:"
] |
c3e95a310bf4df2d6bbb72582aedb698e9c36a19 | # Dataset Card for "fakefractals"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | fenrirgochad/fakefractals | [
"region:us"
] | 2022-12-20T13:52:26+00:00 | {"dataset_info": {"features": [{"name": "label", "dtype": {"class_label": {"names": {"0": "4kfractals", "1": "babies", "2": "babies2", "3": "realfractals"}}}}, {"name": "pixel_values", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 31691574.875, "num_examples": 1561}], "download_size": 31681569, "dataset_size": 31691574.875}} | 2022-12-20T14:04:47+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "fakefractals"
More Information needed | [
"# Dataset Card for \"fakefractals\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"fakefractals\"\n\nMore Information needed"
] |
8a1ad836c8124e7d8098587d30e924e7506f55d1 | # Dataset Card for "yannic-kilcher-transcript"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | matallanas/yannic-kilcher-transcript | [
"region:us"
] | 2022-12-20T14:26:44+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "channel", "dtype": "string"}, {"name": "channel_id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "categories", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "description", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start", "dtype": "float64"}, {"name": "end", "dtype": "float64"}, {"name": "text", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 24560830, "num_examples": 370}], "download_size": 12784371, "dataset_size": 24560830}} | 2022-12-20T14:26:59+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "yannic-kilcher-transcript"
More Information needed | [
"# Dataset Card for \"yannic-kilcher-transcript\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"yannic-kilcher-transcript\"\n\nMore Information needed"
] |
9c3fe5dcc00a5ee3bcfdb6936cbb770ef3c26dfd |
# Dataset Card for BasqueGLUE
## Table of Contents
* [Table of Contents](#table-of-contents)
* [Dataset Description](#dataset-description)
* [Dataset Summary](#dataset-summary)
* [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
* [Languages](#languages)
* [Dataset Structure](#dataset-structure)
* [Data Instances](#data-instances)
* [Data Fields](#data-fields)
* [Data Splits](#data-splits)
* [Dataset Creation](#dataset-creation)
* [Curation Rationale](#curation-rationale)
* [Source Data](#source-data)
* [Annotations](#annotations)
* [Personal and Sensitive Information](#personal-and-sensitive-information)
* [Considerations for Using the Data](#considerations-for-using-the-data)
* [Social Impact of Dataset](#social-impact-of-dataset)
* [Discussion of Biases](#discussion-of-biases)
* [Other Known Limitations](#other-known-limitations)
* [Additional Information](#additional-information)
* [Dataset Curators](#dataset-curators)
* [Licensing Information](#licensing-information)
* [Citation Information](#citation-information)
* [Contributions](#contributions)
## Dataset Description
* **Repository:** <https://github.com/orai-nlp/BasqueGLUE>
* **Paper:** [BasqueGLUE: A Natural Language Understanding Benchmark for Basque](http://www.lrec-conf.org/proceedings/lrec2022/pdf/2022.lrec-1.172.pdf)
* **Point of Contact:** [Contact Information](https://github.com/orai-nlp/BasqueGLUE#contact-information)
### Dataset Summary
Natural Language Understanding (NLU) technology has improved significantly over the last few years, and multitask benchmarks such as GLUE are key to evaluate this improvement in a robust and general way. These benchmarks take into account a wide and diverse set of NLU tasks that require some form of language understanding, beyond the detection of superficial, textual clues. However, they are costly to develop and language-dependent, and therefore they are only available for a small number of languages.
We present BasqueGLUE, the first NLU benchmark for Basque, which has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. BasqueGLUE is freely available under an open license.
| Dataset | \|Train\| | \|Val\| | \|Test\| | Task | Metric | Domain |
|----------------|----------:|--------:|---------:|------------------------|:------:|-----------------|
| NERCid | 51,539 | 12,936 | 35,855 | NERC | F1 | News |
| NERCood | 64,475 | 14,945 | 14,462 | NERC | F1 | News, Wikipedia |
| FMTODeu_intent | 3,418 | 1,904 | 1,087 | Intent classification | F1 | Dialog system |
| FMTODeu_slot | 19,652 | 10,791 | 5,633 | Slot filling | F1 | Dialog system |
| BHTCv2 | 8,585 | 1,857 | 1,854 | Topic classification | F1 | News |
| BEC2016eu | 6,078 | 1,302 | 1,302 | Sentiment analysis | F1 | Twitter |
| VaxxStance | 864 | 206 | 312 | Stance detection | MF1* | Twitter |
| QNLIeu | 1,764 | 230 | 238 | QA/NLI | Acc | Wikipedia |
| WiCeu | 408,559 | 600 | 1,400 | WSD | Acc | Wordnet |
| EpecKorrefBin | 986 | 320 | 587 | Coreference resolution | Acc | News |
### Supported Tasks and Leaderboards
This benchmark comprises the following tasks:
#### NERCid
This dataset contains sentences from the news domain with manually annotated named entities. The data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from naiz.eus. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous.
#### NERCood
This dataset contains sentences with manually annotated named entities. The training data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from naiz.eus. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous. For validation and test sets, sentences from Wikipedia were annotated following the same annotation guidelines.
#### FMTODeu_intent
This dataset contains utterance texts and intent annotations drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The examples are annotated with one of 12 different intent classes corresponding to alarm, reminder or weather related actions.
#### FMTODeu_slot
This dataset contains utterance texts and sequence intent argument annotations designed for slot filling tasks, drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The task is a sequence labelling task similar to NERC, following BIO annotation scheme over 11 categories.
#### BHTCv2
The corpus contains 12,296 news headlines (brief article descriptions) from the Basque weekly newspaper [Argia](https://www.argia.eus). Topics are classified uniquely according to twelve thematic categories.
#### BEC2016eu
The Basque Election Campaign 2016 Opinion Dataset (BEC2016eu) is a new dataset for the task of sentiment analysis, a sequence classification task, which contains tweets about the campaign for the Basque elections from 2016. The crawling was carried out during the election campaign period (2016/09/09-2016/09/23), by monitoring the main parties and their respective candidates. The tweets were manually annotated as positive, negative or neutral.
#### VaxxStance
The VaxxStance (Agerri et al., 2021) dataset originally provides texts and stance annotations for social media texts around the anti-vaccine movement. Texts are given a label indicating whether they express an AGAINST, FAVOR or NEUTRAL stance towards the topic.
#### QNLIeu
This task includes the QA dataset ElkarHizketak (Otegi et al. 2020), a low resource conversational Question Answering (QA) dataset for Basque created by native speaker volunteers. The dataset is built on top of Wikipedia sections about popular people and organizations, and it contains around 400 dialogues and 1600 question and answer pairs. The task was adapted into a sentence-pair binary classification task, following the design of QNLI for English (Wang et al. 2019). Each question and answer pair are given a label indicating whether the answer is entailed by the question.
#### WiCeu
Word in Context or WiC (Pilehvar and Camacho-Collados 2019) is a word sense disambiguation (WSD) task, designed as a particular form of sentence pair binary classification. Given two text snippets and a polyse mous word that appears in both of them (the span of the word is marked in both snippets), the task is to determine whether the word has the same sense in both sentences. This dataset is based on the EPEC-EuSemcor (Pociello et al. 2011) sense-tagged corpus.
#### EpecKorrefBin
EPEC-KORREF-Bin is a dataset derived from EPEC-KORREF (Soraluze et al. 2012), a corpus of Basque news documents with manually annotated mentions and coreference chains, which we have been converted into a binary classification task. In this task, the model has to predict whether two mentions from a text, which can be pronouns, nouns or noun phrases, are referring to the same entity.
#### Leaderboard
Results obtained for two BERT base models as a baseline for the Benchmark.
| | AVG | NERC | F_intent | F_slot | BHTC | BEC | Vaxx | QNLI | WiC | coref |
|------------------------------------------------------------|:-----:|:-----:|:---------:|:-------:|:-----:|:-----:|:-----:|:-----:|:-----:|:-----:|
| Model | | F1 | F1 | F1 | F1 | F1 | MF1 | acc | acc | acc |
|[BERTeus](https://huggingface.co/ixa-ehu/berteus-base-cased)| 73.23 | 81.92 | 82.52 | 74.34 | 78.26 | 69.43 | 59.30 | 74.26 | 70.71 | 68.31 |
|[ElhBERTeu](https://huggingface.co/elh-eus/ElhBERTeu) | 73.71 | 82.30 | 82.24 | 75.64 | 78.05 | 69.89 | 63.81 | 73.84 | 71.71 | 65.93 |
The results obtained on NERC are the average of in-domain and out-of-domain NERC.
### Languages
Data are available in Basque (BCP-47 `eu`)
## Dataset Structure
### Data Instances
#### NERCid/NERCood
An example of 'train' looks as follows:
```
{
"idx": 0,
"tags": ["O", "O", "O", "O", "B-ORG", "O", ...],
"tokens": ["Greba", "orokorrera", "deitu", "du", "EHk", "27rako", ...]
}
```
#### FMTODeu_intent
An example of 'train' looks as follows:
```
{
"idx": 0,
"label": "alarm/modify_alarm",
"text": "aldatu alarma 7am-tik 7pm-ra , mesedez"
}
```
#### FMTODeu_slot
An example of 'train' looks as follows:
```
{
"idx": 923,
"tags": ["O", "B-reminder/todo", "I-datetime", "I-datetime", "B-reminder/todo"],
"tokens": ["gogoratu", "zaborra", "gaur", "gauean", "ateratzea"]
}
```
#### BHTCv2
An example of 'test' looks as follows:
```
{
"idx": 0,
"label": "Gizartea",
"text": "Genero berdintasunaz, hezkuntzaz eta klase gizarteaz hamar liburu baino gehiago..."
}
```
#### BEC2016eu
An example of 'test' looks as follows:
```
{
"idx": 0,
"label": "NEU",
"text": '"Emandako hitza bete egingo dut" Urkullu\nBa galdeketa enegarrenez daramazue programan (ta zuen AHTa...)\n#I25debatea #URL"'
}
```
#### VaxxStance
An example of 'train' looks as follows:
```
{
"idx": 0,
"label": "FAVOR",
"text": "\"#COVID19 Oraingo datuak, izurriaren dinamika, txertoaren eragina eta birusaren..
}
```
#### QNLIeu
An example of 'train' looks as follows:
```
{
"idx": 1,
"label": "not_entailment",
"question": "Zein posiziotan jokatzen du Busquets-ek?",
"sentence": "Busquets 23 partidatan izan zen konbokatua eta 2 gol sartu zituen."
}
```
#### WiCeu
An example of 'test' looks as follows:
```
{
"idx": 16,
"label": false,
"word": "udal",
"sentence1": "1a . Lekeitioko udal mugarteko Alde Historikoa Birgaitzeko Plan Berezia behin...",
"sentence2": "Diezek kritikatu egin zuen EAJk zenbait udaletan EH gobernu taldeetatik at utzi...",
"start1": 16,
"start2": 40,
"end1": 21,
"end2": 49
}
```
#### EpecKorrefBin
An example of 'train' looks as follows:
```
{
"idx": 6,
"label": false,
"text": "Isuntza da faborito nagusia Elantxobeko banderan . ISUNTZA trainerua da faborito nagusia bihar Elantxoben jokatuko den bandera irabazteko .",
"span1_text": "Elantxobeko banderan",
"span2_text": "ISUNTZA trainerua",
"span1_index": 4,
"span2_index": 8
}
```
### Data Fields
#### NERCid
* `tokens`: a list of `string` features
* `tags`: a list of entity labels, with possible values including `person` (PER), `location` (LOC), `organization` (ORG), `miscellaneous` (MISC)
* `idx`: an `int32` feature
#### NERCood
* `tokens`: a list of `string` features
* `tags`: a list of entity labels, with possible values including `person` (PER), `location` (LOC), `organization` (ORG), `miscellaneous` (MISC)
* `idx`: an `int32` feature
#### FMTODeu_intent
* `text`: a `string` feature
* `label`: an intent label, with possible values including:
* `alarm/cancel_alarm`
* `alarm/modify_alarm`
* `alarm/set_alarm`
* `alarm/show_alarms`
* `alarm/snooze_alarm`
* `alarm/time_left_on_alarm`
* `reminder/cancel_reminder`
* `reminder/set_reminder`
* `reminder/show_reminders`
* `weather/checkSunrise`
* `weather/checkSunset`
* `weather/find`
* `idx`: an `int32` feature
#### FMTODeu_slot
* `tokens`: a list of `string` features
* `tags`: a list of intent labels, with possible values including:
* `datetime`
* `location`
* `negation`
* `alarm/alarm_modifier`
* `alarm/recurring_period`
* `reminder/noun`
* `reminder/todo`
* `reminder/reference`
* `reminder/recurring_period`
* `weather/attribute`
* `weather/noun`
* `idx`: an `int32` feature
#### BHTCv2
* `text`: a `string` feature
* `label`: a polarity label, with possible values including `neutral` (NEU), `negative` (N), `positive` (P)
* `idx`: an `int32` feature
#### BEC2016eu
* `text`: a `string` feature
* `label`: a topic label, with possible values including:
* `Ekonomia`
* `Euskal Herria`
* `Euskara`
* `Gizartea`
* `Historia`
* `Ingurumena`
* `Iritzia`
* `Komunikazioa`
* `Kultura`
* `Nazioartea`
* `Politika`
* `Zientzia`
* `idx`: an `int32` feature
#### VaxxStance
* `text`: a `string` feature
* `label`: a stance label, with possible values including `AGAINST`, `FAVOR`, `NONE`
* `idx`: an `int32` feature
#### QNLIeu
* `question`: a `string` feature
* `sentence`: a `string` feature
* `label`: an entailment label, with possible values including `entailment`, `not_entailment`
* `idx`: an `int32` feature
#### WiCeu
* `word`: a `string` feature
* `sentence1`: a `string` feature
* `sentence2`: a `string` feature
* `label`: a `boolean` label indicating sense agreement, with possible values including `true`, `false`
* `start1`: an `int` feature indicating character position where word occurence begins in first sentence
* `start2`: an `int` feature indicating character position where word occurence begins in second sentence
* `end1`: an `int` feature indicating character position where word occurence ends in first sentence
* `end2`: an `int` feature indicating character position where word occurence ends in second sentence
* `idx`: an `int32` feature
#### EpecKorrefBin
* `text`: a `string` feature.
* `label`: a `boolean` coreference label, with possible values including `true`, `false`.
* `span1_text`: a `string` feature
* `span2_text`: a `string` feature
* `span1_index`: an `int` feature indicating token index where `span1_text` feature occurs in `text`
* `span2_index`: an `int` feature indicating token index where `span2_text` feature occurs in `text`
* `idx`: an `int32` feature
### Data Splits
| Dataset | \|Train\| | \|Val\| | \|Test\| |
|---------|--------:|------:|-------:|
| NERCid | 51,539 | 12,936 | 35,855 |
| NERCood | 64,475 | 14,945 | 14,462 |
| FMTODeu_intent | 3,418 | 1,904 | 1,087 |
| FMTODeu_slot | 19,652 | 10,791 | 5,633 |
| BHTCv2 | 8,585 | 1,857 | 1,854 |
| BEC2016eu | 6,078 | 1,302 | 1,302 |
| VaxxStance | 864 | 206 | 312 |
| QNLIeu | 1,764 | 230 | 238 |
| WiCeu | 408,559 | 600 | 1,400 |
| EpecKorrefBin | 986 | 320 | 587 |
## Dataset Creation
### Curation Rationale
We believe that BasqueGLUE is a significant contribution towards developing NLU tools in Basque, which we believe will facilitate the technological advance for the Basque language. In order to create BasqueGLUE we took as a reference the GLUE and SuperGLUE frameworks. When possible, we re-used existing datasets for Basque, adapting them to the corresponding task formats if necessary. Additionally, BasqueGLUE also includes six new datasets that have not been published before. In total, BasqueGLUE consists of nine Basque NLU tasks and covers a wide range of tasks with different difficulties across several domains. As with the original GLUE benchmark, the training data for the tasks vary in size, which allows to measure the performance of how the models transfer knowledge across tasks.
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
Gorka Urbizu [1], Iรฑaki San Vicente [1], Xabier Saralegi [1], Rodrigo Agerri [2] and Aitor Soroa [2]
Affiliation of the authors:
[1] orai NLP Technologies
[2] HiTZ Center - Ixa, University of the Basque Country UPV/EHU
### Licensing Information
Each dataset of the BasqueGLUE benchmark has it's own license (due to most of them being or being derived from already existing datasets). See their respective README files for details.
Here we provide a brief summary of their licenses:
| Dataset | License |
|---------|---------|
| NERCid | CC BY-NC-SA 4.0 |
| NERCood | CC BY-NC-SA 4.0 |
| FMTODeu_intent | CC BY-NC-SA 4.0 |
| FMTODeu_slot | CC BY-NC-SA 4.0 |
| BHTCv2 | CC BY-NC-SA 4.0 |
| BEC2016eu | Twitter's license + CC BY-NC-SA 4.0 |
| VaxxStance | Twitter's license + CC BY 4.0 |
| QNLIeu | CC BY-SA 4.0 |
| WiCeu | CC BY-NC-SA 4.0 |
| EpecKorrefBin | CC BY-NC-SA 4.0 |
For the rest of the files of the benchmark, including the loading and evaluation scripts, the following license applies:
Copyright (C) by Orai NLP Technologies.
This benchmark and evaluation scripts are licensed under the Creative Commons Attribution Share Alike 4.0
International License (CC BY-SA 4.0). To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
### Citation Information
```
@InProceedings{urbizu2022basqueglue,
author = {Urbizu, Gorka and San Vicente, Iรฑaki and Saralegi, Xabier and Agerri, Rodrigo and Soroa, Aitor},
title = {BasqueGLUE: A Natural Language Understanding Benchmark for Basque},
booktitle = {Proceedings of the Language Resources and Evaluation Conference},
month = {June},
year = {2022},
address = {Marseille, France},
publisher = {European Language Resources Association},
pages = {1603--1612},
abstract = {Natural Language Understanding (NLU) technology has improved significantly over the last few years and multitask benchmarks such as GLUE are key to evaluate this improvement in a robust and general way. These benchmarks take into account a wide and diverse set of NLU tasks that require some form of language understanding, beyond the detection of superficial, textual clues. However, they are costly to develop and language-dependent, and therefore they are only available for a small number of languages. In this paper, we present BasqueGLUE, the first NLU benchmark for Basque, a less-resourced language, which has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. We also report the evaluation of two state-of-the-art language models for Basque on BasqueGLUE, thus providing a strong baseline to compare upon. BasqueGLUE is freely available under an open license.},
url = {https://aclanthology.org/2022.lrec-1.172}
}
```
### Contributions
Thanks to [@richplant](https://github.com/richplant) for adding this dataset to hugginface.
| orai-nlp/basqueGLUE | [
"task_categories:text-classification",
"task_categories:token-classification",
"task_ids:intent-classification",
"task_ids:natural-language-inference",
"task_ids:sentiment-classification",
"task_ids:topic-classification",
"task_ids:named-entity-recognition",
"task_ids:coreference-resolution",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:eu",
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2022-12-20T14:28:19+00:00 | {"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["eu"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-classification", "token-classification"], "task_ids": ["intent-classification", "natural-language-inference", "sentiment-classification", "topic-classification", "named-entity-recognition", "coreference-resolution"], "pretty_name": "BasqueGLUE", "tags": [], "configs": ["bec", "bhtc", "coref", "intent", "nerc_id", "nerc_od", "qnli", "slot", "vaxx", "wic"], "dataset_info": [{"config_name": "bec", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "N", "1": "NEU", "2": "P"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 693284, "num_examples": 6078}, {"name": "test", "num_bytes": 148510, "num_examples": 1302}, {"name": "validation", "num_bytes": 148377, "num_examples": 1302}], "download_size": 1217803, "dataset_size": 990171}, {"config_name": "bhtc", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "Ekonomia", "1": "Euskal Herria", "2": "Euskara", "3": "Gizartea", "4": "Historia", "5": "Ingurumena", "6": "Iritzia", "7": "Komunikazioa", "8": "Kultura", "9": "Nazioartea", "10": "Politika", "11": "Zientzia"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 2431494, "num_examples": 8585}, {"name": "test", "num_bytes": 523066, "num_examples": 1854}, {"name": "validation", "num_bytes": 519555, "num_examples": 1857}], "download_size": 3896312, "dataset_size": 3474115}, {"config_name": "coref", "features": [{"name": "text", "dtype": "string"}, {"name": "span1_text", "dtype": "string"}, {"name": "span2_text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "false", "1": "true"}}}}, {"name": "span1_index", "dtype": "int32"}, {"name": "span2_index", "dtype": "int32"}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 365830, "num_examples": 986}, {"name": "test", "num_bytes": 201378, "num_examples": 587}, {"name": "validation", "num_bytes": 108632, "num_examples": 320}], "download_size": 855074, "dataset_size": 675840}, {"config_name": "intent", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "alarm/cancel_alarm", "1": "alarm/modify_alarm", "2": "alarm/set_alarm", "3": "alarm/show_alarms", "4": "alarm/snooze_alarm", "5": "alarm/time_left_on_alarm", "6": "reminder/cancel_reminder", "7": "reminder/set_reminder", "8": "reminder/show_reminders", "9": "weather/checkSunrise", "10": "weather/checkSunset", "11": "weather/find"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 182856, "num_examples": 3418}, {"name": "test", "num_bytes": 56118, "num_examples": 1087}, {"name": "validation", "num_bytes": 101644, "num_examples": 1904}], "download_size": 595375, "dataset_size": 340618}, {"config_name": "nerc_id", "features": [{"name": "tokens", "sequence": "string"}, {"name": "tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-LOC", "4": "I-LOC", "5": "B-ORG", "6": "I-ORG", "7": "B-MISC", "8": "I-MISC"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 946007, "num_examples": 2842}, {"name": "test", "num_bytes": 653960, "num_examples": 1846}, {"name": "validation", "num_bytes": 237464, "num_examples": 711}], "download_size": 1723325, "dataset_size": 1837431}, {"config_name": "nerc_od", "features": [{"name": "tokens", "sequence": "string"}, {"name": "tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-LOC", "4": "I-LOC", "5": "B-ORG", "6": "I-ORG", "7": "B-MISC", "8": "I-MISC"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 1183471, "num_examples": 3553}, {"name": "test", "num_bytes": 262853, "num_examples": 598}, {"name": "validation", "num_bytes": 270028, "num_examples": 601}], "download_size": 1613369, "dataset_size": 1716352}, {"config_name": "qnli", "features": [{"name": "question", "dtype": "string"}, {"name": "sentence", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "entailment", "1": "not_entailment"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 327189, "num_examples": 1764}, {"name": "test", "num_bytes": 42569, "num_examples": 238}, {"name": "validation", "num_bytes": 46359, "num_examples": 230}], "download_size": 532399, "dataset_size": 416117}, {"config_name": "slot", "features": [{"name": "tokens", "sequence": "string"}, {"name": "tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-datetime", "2": "B-location", "3": "B-negation", "4": "B-alarm/alarm_modifier", "5": "B-alarm/recurring_period", "6": "B-reminder/noun", "7": "B-reminder/todo", "8": "B-reminder/reference", "9": "B-reminder/recurring_period", "10": "B-weather/attribute", "11": "B-weather/noun", "12": "I-datetime", "13": "I-location", "14": "I-negation", "15": "I-alarm/alarm_modifier", "16": "I-alarm/recurring_period", "17": "I-reminder/noun", "18": "I-reminder/todo", "19": "I-reminder/reference", "20": "I-reminder/recurring_period", "21": "I-weather/attribute", "22": "I-weather/noun"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 388774, "num_examples": 3418}, {"name": "test", "num_bytes": 114876, "num_examples": 1088}, {"name": "validation", "num_bytes": 214053, "num_examples": 1900}], "download_size": 962250, "dataset_size": 717703}, {"config_name": "vaxx", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "AGAINST", "1": "NONE", "2": "FAVOR"}}}}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 176436, "num_examples": 864}, {"name": "test", "num_bytes": 70947, "num_examples": 312}, {"name": "validation", "num_bytes": 42795, "num_examples": 206}], "download_size": 333997, "dataset_size": 290178}, {"config_name": "wic", "features": [{"name": "sentence1", "dtype": "string"}, {"name": "sentence2", "dtype": "string"}, {"name": "word", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "false", "1": "true"}}}}, {"name": "start1", "dtype": "int32"}, {"name": "start2", "dtype": "int32"}, {"name": "end1", "dtype": "int32"}, {"name": "end2", "dtype": "int32"}, {"name": "idx", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 172847108, "num_examples": 408559}, {"name": "test", "num_bytes": 589578, "num_examples": 1400}, {"name": "validation", "num_bytes": 251549, "num_examples": 600}], "download_size": 22938354, "dataset_size": 173688235}]} | 2022-12-21T09:54:32+00:00 | [] | [
"eu"
] | TAGS
#task_categories-text-classification #task_categories-token-classification #task_ids-intent-classification #task_ids-natural-language-inference #task_ids-sentiment-classification #task_ids-topic-classification #task_ids-named-entity-recognition #task_ids-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Basque #license-cc-by-nc-sa-4.0 #region-us
| Dataset Card for BasqueGLUE
===========================
Table of Contents
-----------------
* Table of Contents
* Dataset Description
+ Dataset Summary
+ Supported Tasks and Leaderboards
+ Languages
* Dataset Structure
+ Data Instances
+ Data Fields
+ Data Splits
* Dataset Creation
+ Curation Rationale
+ Source Data
+ Annotations
+ Personal and Sensitive Information
* Considerations for Using the Data
+ Social Impact of Dataset
+ Discussion of Biases
+ Other Known Limitations
* Additional Information
+ Dataset Curators
+ Licensing Information
+ Citation Information
+ Contributions
Dataset Description
-------------------
* Repository: <URL
* Paper: BasqueGLUE: A Natural Language Understanding Benchmark for Basque
* Point of Contact: Contact Information
### Dataset Summary
Natural Language Understanding (NLU) technology has improved significantly over the last few years, and multitask benchmarks such as GLUE are key to evaluate this improvement in a robust and general way. These benchmarks take into account a wide and diverse set of NLU tasks that require some form of language understanding, beyond the detection of superficial, textual clues. However, they are costly to develop and language-dependent, and therefore they are only available for a small number of languages.
We present BasqueGLUE, the first NLU benchmark for Basque, which has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. BasqueGLUE is freely available under an open license.
### Supported Tasks and Leaderboards
This benchmark comprises the following tasks:
#### NERCid
This dataset contains sentences from the news domain with manually annotated named entities. The data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous.
#### NERCood
This dataset contains sentences with manually annotated named entities. The training data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous. For validation and test sets, sentences from Wikipedia were annotated following the same annotation guidelines.
#### FMTODeu\_intent
This dataset contains utterance texts and intent annotations drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The examples are annotated with one of 12 different intent classes corresponding to alarm, reminder or weather related actions.
#### FMTODeu\_slot
This dataset contains utterance texts and sequence intent argument annotations designed for slot filling tasks, drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The task is a sequence labelling task similar to NERC, following BIO annotation scheme over 11 categories.
#### BHTCv2
The corpus contains 12,296 news headlines (brief article descriptions) from the Basque weekly newspaper Argia. Topics are classified uniquely according to twelve thematic categories.
#### BEC2016eu
The Basque Election Campaign 2016 Opinion Dataset (BEC2016eu) is a new dataset for the task of sentiment analysis, a sequence classification task, which contains tweets about the campaign for the Basque elections from 2016. The crawling was carried out during the election campaign period (2016/09/09-2016/09/23), by monitoring the main parties and their respective candidates. The tweets were manually annotated as positive, negative or neutral.
#### VaxxStance
The VaxxStance (Agerri et al., 2021) dataset originally provides texts and stance annotations for social media texts around the anti-vaccine movement. Texts are given a label indicating whether they express an AGAINST, FAVOR or NEUTRAL stance towards the topic.
#### QNLIeu
This task includes the QA dataset ElkarHizketak (Otegi et al. 2020), a low resource conversational Question Answering (QA) dataset for Basque created by native speaker volunteers. The dataset is built on top of Wikipedia sections about popular people and organizations, and it contains around 400 dialogues and 1600 question and answer pairs. The task was adapted into a sentence-pair binary classification task, following the design of QNLI for English (Wang et al. 2019). Each question and answer pair are given a label indicating whether the answer is entailed by the question.
#### WiCeu
Word in Context or WiC (Pilehvar and Camacho-Collados 2019) is a word sense disambiguation (WSD) task, designed as a particular form of sentence pair binary classification. Given two text snippets and a polyse mous word that appears in both of them (the span of the word is marked in both snippets), the task is to determine whether the word has the same sense in both sentences. This dataset is based on the EPEC-EuSemcor (Pociello et al. 2011) sense-tagged corpus.
#### EpecKorrefBin
EPEC-KORREF-Bin is a dataset derived from EPEC-KORREF (Soraluze et al. 2012), a corpus of Basque news documents with manually annotated mentions and coreference chains, which we have been converted into a binary classification task. In this task, the model has to predict whether two mentions from a text, which can be pronouns, nouns or noun phrases, are referring to the same entity.
#### Leaderboard
Results obtained for two BERT base models as a baseline for the Benchmark.
The results obtained on NERC are the average of in-domain and out-of-domain NERC.
### Languages
Data are available in Basque (BCP-47 'eu')
Dataset Structure
-----------------
### Data Instances
#### NERCid/NERCood
An example of 'train' looks as follows:
#### FMTODeu\_intent
An example of 'train' looks as follows:
#### FMTODeu\_slot
An example of 'train' looks as follows:
#### BHTCv2
An example of 'test' looks as follows:
#### BEC2016eu
An example of 'test' looks as follows:
#### VaxxStance
An example of 'train' looks as follows:
#### QNLIeu
An example of 'train' looks as follows:
#### WiCeu
An example of 'test' looks as follows:
#### EpecKorrefBin
An example of 'train' looks as follows:
### Data Fields
#### NERCid
* 'tokens': a list of 'string' features
* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)
* 'idx': an 'int32' feature
#### NERCood
* 'tokens': a list of 'string' features
* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)
* 'idx': an 'int32' feature
#### FMTODeu\_intent
* 'text': a 'string' feature
* 'label': an intent label, with possible values including:
+ 'alarm/cancel\_alarm'
+ 'alarm/modify\_alarm'
+ 'alarm/set\_alarm'
+ 'alarm/show\_alarms'
+ 'alarm/snooze\_alarm'
+ 'alarm/time\_left\_on\_alarm'
+ 'reminder/cancel\_reminder'
+ 'reminder/set\_reminder'
+ 'reminder/show\_reminders'
+ 'weather/checkSunrise'
+ 'weather/checkSunset'
+ 'weather/find'
* 'idx': an 'int32' feature
#### FMTODeu\_slot
* 'tokens': a list of 'string' features
* 'tags': a list of intent labels, with possible values including:
+ 'datetime'
+ 'location'
+ 'negation'
+ 'alarm/alarm\_modifier'
+ 'alarm/recurring\_period'
+ 'reminder/noun'
+ 'reminder/todo'
+ 'reminder/reference'
+ 'reminder/recurring\_period'
+ 'weather/attribute'
+ 'weather/noun'
* 'idx': an 'int32' feature
#### BHTCv2
* 'text': a 'string' feature
* 'label': a polarity label, with possible values including 'neutral' (NEU), 'negative' (N), 'positive' (P)
* 'idx': an 'int32' feature
#### BEC2016eu
* 'text': a 'string' feature
* 'label': a topic label, with possible values including:
+ 'Ekonomia'
+ 'Euskal Herria'
+ 'Euskara'
+ 'Gizartea'
+ 'Historia'
+ 'Ingurumena'
+ 'Iritzia'
+ 'Komunikazioa'
+ 'Kultura'
+ 'Nazioartea'
+ 'Politika'
+ 'Zientzia'
* 'idx': an 'int32' feature
#### VaxxStance
* 'text': a 'string' feature
* 'label': a stance label, with possible values including 'AGAINST', 'FAVOR', 'NONE'
* 'idx': an 'int32' feature
#### QNLIeu
* 'question': a 'string' feature
* 'sentence': a 'string' feature
* 'label': an entailment label, with possible values including 'entailment', 'not\_entailment'
* 'idx': an 'int32' feature
#### WiCeu
* 'word': a 'string' feature
* 'sentence1': a 'string' feature
* 'sentence2': a 'string' feature
* 'label': a 'boolean' label indicating sense agreement, with possible values including 'true', 'false'
* 'start1': an 'int' feature indicating character position where word occurence begins in first sentence
* 'start2': an 'int' feature indicating character position where word occurence begins in second sentence
* 'end1': an 'int' feature indicating character position where word occurence ends in first sentence
* 'end2': an 'int' feature indicating character position where word occurence ends in second sentence
* 'idx': an 'int32' feature
#### EpecKorrefBin
* 'text': a 'string' feature.
* 'label': a 'boolean' coreference label, with possible values including 'true', 'false'.
* 'span1\_text': a 'string' feature
* 'span2\_text': a 'string' feature
* 'span1\_index': an 'int' feature indicating token index where 'span1\_text' feature occurs in 'text'
* 'span2\_index': an 'int' feature indicating token index where 'span2\_text' feature occurs in 'text'
* 'idx': an 'int32' feature
### Data Splits
Dataset Creation
----------------
### Curation Rationale
We believe that BasqueGLUE is a significant contribution towards developing NLU tools in Basque, which we believe will facilitate the technological advance for the Basque language. In order to create BasqueGLUE we took as a reference the GLUE and SuperGLUE frameworks. When possible, we re-used existing datasets for Basque, adapting them to the corresponding task formats if necessary. Additionally, BasqueGLUE also includes six new datasets that have not been published before. In total, BasqueGLUE consists of nine Basque NLU tasks and covers a wide range of tasks with different difficulties across several domains. As with the original GLUE benchmark, the training data for the tasks vary in size, which allows to measure the performance of how the models transfer knowledge across tasks.
### Source Data
#### Initial Data Collection and Normalization
#### Who are the source language producers?
### Annotations
#### Annotation process
#### Who are the annotators?
### Personal and Sensitive Information
Considerations for Using the Data
---------------------------------
### Social Impact of Dataset
### Discussion of Biases
### Other Known Limitations
Additional Information
----------------------
### Dataset Curators
Gorka Urbizu [1], Iรฑaki San Vicente [1], Xabier Saralegi [1], Rodrigo Agerri [2] and Aitor Soroa [2]
Affiliation of the authors:
[1] orai NLP Technologies
[2] HiTZ Center - Ixa, University of the Basque Country UPV/EHU
### Licensing Information
Each dataset of the BasqueGLUE benchmark has it's own license (due to most of them being or being derived from already existing datasets). See their respective README files for details.
Here we provide a brief summary of their licenses:
For the rest of the files of the benchmark, including the loading and evaluation scripts, the following license applies:
Copyright (C) by Orai NLP Technologies.
This benchmark and evaluation scripts are licensed under the Creative Commons Attribution Share Alike 4.0
International License (CC BY-SA 4.0). To view a copy of this license, visit URL
### Contributions
Thanks to @richplant for adding this dataset to hugginface.
| [
"### Dataset Summary\n\n\nNatural Language Understanding (NLU) technology has improved significantly over the last few years, and multitask benchmarks such as GLUE are key to evaluate this improvement in a robust and general way. These benchmarks take into account a wide and diverse set of NLU tasks that require some form of language understanding, beyond the detection of superficial, textual clues. However, they are costly to develop and language-dependent, and therefore they are only available for a small number of languages.\n\n\nWe present BasqueGLUE, the first NLU benchmark for Basque, which has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. BasqueGLUE is freely available under an open license.",
"### Supported Tasks and Leaderboards\n\n\nThis benchmark comprises the following tasks:",
"#### NERCid\n\n\nThis dataset contains sentences from the news domain with manually annotated named entities. The data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous.",
"#### NERCood\n\n\nThis dataset contains sentences with manually annotated named entities. The training data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous. For validation and test sets, sentences from Wikipedia were annotated following the same annotation guidelines.",
"#### FMTODeu\\_intent\n\n\nThis dataset contains utterance texts and intent annotations drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The examples are annotated with one of 12 different intent classes corresponding to alarm, reminder or weather related actions.",
"#### FMTODeu\\_slot\n\n\nThis dataset contains utterance texts and sequence intent argument annotations designed for slot filling tasks, drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The task is a sequence labelling task similar to NERC, following BIO annotation scheme over 11 categories.",
"#### BHTCv2\n\n\nThe corpus contains 12,296 news headlines (brief article descriptions) from the Basque weekly newspaper Argia. Topics are classified uniquely according to twelve thematic categories.",
"#### BEC2016eu\n\n\nThe Basque Election Campaign 2016 Opinion Dataset (BEC2016eu) is a new dataset for the task of sentiment analysis, a sequence classification task, which contains tweets about the campaign for the Basque elections from 2016. The crawling was carried out during the election campaign period (2016/09/09-2016/09/23), by monitoring the main parties and their respective candidates. The tweets were manually annotated as positive, negative or neutral.",
"#### VaxxStance\n\n\nThe VaxxStance (Agerri et al., 2021) dataset originally provides texts and stance annotations for social media texts around the anti-vaccine movement. Texts are given a label indicating whether they express an AGAINST, FAVOR or NEUTRAL stance towards the topic.",
"#### QNLIeu\n\n\nThis task includes the QA dataset ElkarHizketak (Otegi et al. 2020), a low resource conversational Question Answering (QA) dataset for Basque created by native speaker volunteers. The dataset is built on top of Wikipedia sections about popular people and organizations, and it contains around 400 dialogues and 1600 question and answer pairs. The task was adapted into a sentence-pair binary classification task, following the design of QNLI for English (Wang et al. 2019). Each question and answer pair are given a label indicating whether the answer is entailed by the question.",
"#### WiCeu\n\n\nWord in Context or WiC (Pilehvar and Camacho-Collados 2019) is a word sense disambiguation (WSD) task, designed as a particular form of sentence pair binary classification. Given two text snippets and a polyse mous word that appears in both of them (the span of the word is marked in both snippets), the task is to determine whether the word has the same sense in both sentences. This dataset is based on the EPEC-EuSemcor (Pociello et al. 2011) sense-tagged corpus.",
"#### EpecKorrefBin\n\n\nEPEC-KORREF-Bin is a dataset derived from EPEC-KORREF (Soraluze et al. 2012), a corpus of Basque news documents with manually annotated mentions and coreference chains, which we have been converted into a binary classification task. In this task, the model has to predict whether two mentions from a text, which can be pronouns, nouns or noun phrases, are referring to the same entity.",
"#### Leaderboard\n\n\nResults obtained for two BERT base models as a baseline for the Benchmark.\n\n\n\nThe results obtained on NERC are the average of in-domain and out-of-domain NERC.",
"### Languages\n\n\nData are available in Basque (BCP-47 'eu')\n\n\nDataset Structure\n-----------------",
"### Data Instances",
"#### NERCid/NERCood\n\n\nAn example of 'train' looks as follows:",
"#### FMTODeu\\_intent\n\n\nAn example of 'train' looks as follows:",
"#### FMTODeu\\_slot\n\n\nAn example of 'train' looks as follows:",
"#### BHTCv2\n\n\nAn example of 'test' looks as follows:",
"#### BEC2016eu\n\n\nAn example of 'test' looks as follows:",
"#### VaxxStance\n\n\nAn example of 'train' looks as follows:",
"#### QNLIeu\n\n\nAn example of 'train' looks as follows:",
"#### WiCeu\n\n\nAn example of 'test' looks as follows:",
"#### EpecKorrefBin\n\n\nAn example of 'train' looks as follows:",
"### Data Fields",
"#### NERCid\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)\n* 'idx': an 'int32' feature",
"#### NERCood\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)\n* 'idx': an 'int32' feature",
"#### FMTODeu\\_intent\n\n\n* 'text': a 'string' feature\n* 'label': an intent label, with possible values including:\n\t+ 'alarm/cancel\\_alarm'\n\t+ 'alarm/modify\\_alarm'\n\t+ 'alarm/set\\_alarm'\n\t+ 'alarm/show\\_alarms'\n\t+ 'alarm/snooze\\_alarm'\n\t+ 'alarm/time\\_left\\_on\\_alarm'\n\t+ 'reminder/cancel\\_reminder'\n\t+ 'reminder/set\\_reminder'\n\t+ 'reminder/show\\_reminders'\n\t+ 'weather/checkSunrise'\n\t+ 'weather/checkSunset'\n\t+ 'weather/find'\n* 'idx': an 'int32' feature",
"#### FMTODeu\\_slot\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of intent labels, with possible values including:\n\t+ 'datetime'\n\t+ 'location'\n\t+ 'negation'\n\t+ 'alarm/alarm\\_modifier'\n\t+ 'alarm/recurring\\_period'\n\t+ 'reminder/noun'\n\t+ 'reminder/todo'\n\t+ 'reminder/reference'\n\t+ 'reminder/recurring\\_period'\n\t+ 'weather/attribute'\n\t+ 'weather/noun'\n* 'idx': an 'int32' feature",
"#### BHTCv2\n\n\n* 'text': a 'string' feature\n* 'label': a polarity label, with possible values including 'neutral' (NEU), 'negative' (N), 'positive' (P)\n* 'idx': an 'int32' feature",
"#### BEC2016eu\n\n\n* 'text': a 'string' feature\n* 'label': a topic label, with possible values including:\n\t+ 'Ekonomia'\n\t+ 'Euskal Herria'\n\t+ 'Euskara'\n\t+ 'Gizartea'\n\t+ 'Historia'\n\t+ 'Ingurumena'\n\t+ 'Iritzia'\n\t+ 'Komunikazioa'\n\t+ 'Kultura'\n\t+ 'Nazioartea'\n\t+ 'Politika'\n\t+ 'Zientzia'\n* 'idx': an 'int32' feature",
"#### VaxxStance\n\n\n* 'text': a 'string' feature\n* 'label': a stance label, with possible values including 'AGAINST', 'FAVOR', 'NONE'\n* 'idx': an 'int32' feature",
"#### QNLIeu\n\n\n* 'question': a 'string' feature\n* 'sentence': a 'string' feature\n* 'label': an entailment label, with possible values including 'entailment', 'not\\_entailment'\n* 'idx': an 'int32' feature",
"#### WiCeu\n\n\n* 'word': a 'string' feature\n* 'sentence1': a 'string' feature\n* 'sentence2': a 'string' feature\n* 'label': a 'boolean' label indicating sense agreement, with possible values including 'true', 'false'\n* 'start1': an 'int' feature indicating character position where word occurence begins in first sentence\n* 'start2': an 'int' feature indicating character position where word occurence begins in second sentence\n* 'end1': an 'int' feature indicating character position where word occurence ends in first sentence\n* 'end2': an 'int' feature indicating character position where word occurence ends in second sentence\n* 'idx': an 'int32' feature",
"#### EpecKorrefBin\n\n\n* 'text': a 'string' feature.\n* 'label': a 'boolean' coreference label, with possible values including 'true', 'false'.\n* 'span1\\_text': a 'string' feature\n* 'span2\\_text': a 'string' feature\n* 'span1\\_index': an 'int' feature indicating token index where 'span1\\_text' feature occurs in 'text'\n* 'span2\\_index': an 'int' feature indicating token index where 'span2\\_text' feature occurs in 'text'\n* 'idx': an 'int32' feature",
"### Data Splits\n\n\n\nDataset Creation\n----------------",
"### Curation Rationale\n\n\nWe believe that BasqueGLUE is a significant contribution towards developing NLU tools in Basque, which we believe will facilitate the technological advance for the Basque language. In order to create BasqueGLUE we took as a reference the GLUE and SuperGLUE frameworks. When possible, we re-used existing datasets for Basque, adapting them to the corresponding task formats if necessary. Additionally, BasqueGLUE also includes six new datasets that have not been published before. In total, BasqueGLUE consists of nine Basque NLU tasks and covers a wide range of tasks with different difficulties across several domains. As with the original GLUE benchmark, the training data for the tasks vary in size, which allows to measure the performance of how the models transfer knowledge across tasks.",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations\n\n\nAdditional Information\n----------------------",
"### Dataset Curators\n\n\nGorka Urbizu [1], Iรฑaki San Vicente [1], Xabier Saralegi [1], Rodrigo Agerri [2] and Aitor Soroa [2]\n\n\nAffiliation of the authors:\n\n\n[1] orai NLP Technologies\n\n\n[2] HiTZ Center - Ixa, University of the Basque Country UPV/EHU",
"### Licensing Information\n\n\nEach dataset of the BasqueGLUE benchmark has it's own license (due to most of them being or being derived from already existing datasets). See their respective README files for details.\n\n\nHere we provide a brief summary of their licenses:\n\n\n\nFor the rest of the files of the benchmark, including the loading and evaluation scripts, the following license applies:\n\n\nCopyright (C) by Orai NLP Technologies.\nThis benchmark and evaluation scripts are licensed under the Creative Commons Attribution Share Alike 4.0\nInternational License (CC BY-SA 4.0). To view a copy of this license, visit URL",
"### Contributions\n\n\nThanks to @richplant for adding this dataset to hugginface."
] | [
"TAGS\n#task_categories-text-classification #task_categories-token-classification #task_ids-intent-classification #task_ids-natural-language-inference #task_ids-sentiment-classification #task_ids-topic-classification #task_ids-named-entity-recognition #task_ids-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Basque #license-cc-by-nc-sa-4.0 #region-us \n",
"### Dataset Summary\n\n\nNatural Language Understanding (NLU) technology has improved significantly over the last few years, and multitask benchmarks such as GLUE are key to evaluate this improvement in a robust and general way. These benchmarks take into account a wide and diverse set of NLU tasks that require some form of language understanding, beyond the detection of superficial, textual clues. However, they are costly to develop and language-dependent, and therefore they are only available for a small number of languages.\n\n\nWe present BasqueGLUE, the first NLU benchmark for Basque, which has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. BasqueGLUE is freely available under an open license.",
"### Supported Tasks and Leaderboards\n\n\nThis benchmark comprises the following tasks:",
"#### NERCid\n\n\nThis dataset contains sentences from the news domain with manually annotated named entities. The data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous.",
"#### NERCood\n\n\nThis dataset contains sentences with manually annotated named entities. The training data is the merge of EIEC (a dataset of a collection of news wire articles from Euskaldunon Egunkaria newspaper, (Alegria et al. 2004)), and newly annotated data from URL. The data is annotated following the BIO annotation scheme over four categories: person, organization, location, and miscellaneous. For validation and test sets, sentences from Wikipedia were annotated following the same annotation guidelines.",
"#### FMTODeu\\_intent\n\n\nThis dataset contains utterance texts and intent annotations drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The examples are annotated with one of 12 different intent classes corresponding to alarm, reminder or weather related actions.",
"#### FMTODeu\\_slot\n\n\nThis dataset contains utterance texts and sequence intent argument annotations designed for slot filling tasks, drawn from the manually-annotated Facebook Multilingual Task Oriented Dataset (FMTOD) (Schuster et al. 2019). Basque translated data was drawn from the datasets created for Building a Task-oriented Dialog System for languages with no training data: the Case for Basque (de Lacalle et al. 2020). The task is a sequence labelling task similar to NERC, following BIO annotation scheme over 11 categories.",
"#### BHTCv2\n\n\nThe corpus contains 12,296 news headlines (brief article descriptions) from the Basque weekly newspaper Argia. Topics are classified uniquely according to twelve thematic categories.",
"#### BEC2016eu\n\n\nThe Basque Election Campaign 2016 Opinion Dataset (BEC2016eu) is a new dataset for the task of sentiment analysis, a sequence classification task, which contains tweets about the campaign for the Basque elections from 2016. The crawling was carried out during the election campaign period (2016/09/09-2016/09/23), by monitoring the main parties and their respective candidates. The tweets were manually annotated as positive, negative or neutral.",
"#### VaxxStance\n\n\nThe VaxxStance (Agerri et al., 2021) dataset originally provides texts and stance annotations for social media texts around the anti-vaccine movement. Texts are given a label indicating whether they express an AGAINST, FAVOR or NEUTRAL stance towards the topic.",
"#### QNLIeu\n\n\nThis task includes the QA dataset ElkarHizketak (Otegi et al. 2020), a low resource conversational Question Answering (QA) dataset for Basque created by native speaker volunteers. The dataset is built on top of Wikipedia sections about popular people and organizations, and it contains around 400 dialogues and 1600 question and answer pairs. The task was adapted into a sentence-pair binary classification task, following the design of QNLI for English (Wang et al. 2019). Each question and answer pair are given a label indicating whether the answer is entailed by the question.",
"#### WiCeu\n\n\nWord in Context or WiC (Pilehvar and Camacho-Collados 2019) is a word sense disambiguation (WSD) task, designed as a particular form of sentence pair binary classification. Given two text snippets and a polyse mous word that appears in both of them (the span of the word is marked in both snippets), the task is to determine whether the word has the same sense in both sentences. This dataset is based on the EPEC-EuSemcor (Pociello et al. 2011) sense-tagged corpus.",
"#### EpecKorrefBin\n\n\nEPEC-KORREF-Bin is a dataset derived from EPEC-KORREF (Soraluze et al. 2012), a corpus of Basque news documents with manually annotated mentions and coreference chains, which we have been converted into a binary classification task. In this task, the model has to predict whether two mentions from a text, which can be pronouns, nouns or noun phrases, are referring to the same entity.",
"#### Leaderboard\n\n\nResults obtained for two BERT base models as a baseline for the Benchmark.\n\n\n\nThe results obtained on NERC are the average of in-domain and out-of-domain NERC.",
"### Languages\n\n\nData are available in Basque (BCP-47 'eu')\n\n\nDataset Structure\n-----------------",
"### Data Instances",
"#### NERCid/NERCood\n\n\nAn example of 'train' looks as follows:",
"#### FMTODeu\\_intent\n\n\nAn example of 'train' looks as follows:",
"#### FMTODeu\\_slot\n\n\nAn example of 'train' looks as follows:",
"#### BHTCv2\n\n\nAn example of 'test' looks as follows:",
"#### BEC2016eu\n\n\nAn example of 'test' looks as follows:",
"#### VaxxStance\n\n\nAn example of 'train' looks as follows:",
"#### QNLIeu\n\n\nAn example of 'train' looks as follows:",
"#### WiCeu\n\n\nAn example of 'test' looks as follows:",
"#### EpecKorrefBin\n\n\nAn example of 'train' looks as follows:",
"### Data Fields",
"#### NERCid\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)\n* 'idx': an 'int32' feature",
"#### NERCood\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of entity labels, with possible values including 'person' (PER), 'location' (LOC), 'organization' (ORG), 'miscellaneous' (MISC)\n* 'idx': an 'int32' feature",
"#### FMTODeu\\_intent\n\n\n* 'text': a 'string' feature\n* 'label': an intent label, with possible values including:\n\t+ 'alarm/cancel\\_alarm'\n\t+ 'alarm/modify\\_alarm'\n\t+ 'alarm/set\\_alarm'\n\t+ 'alarm/show\\_alarms'\n\t+ 'alarm/snooze\\_alarm'\n\t+ 'alarm/time\\_left\\_on\\_alarm'\n\t+ 'reminder/cancel\\_reminder'\n\t+ 'reminder/set\\_reminder'\n\t+ 'reminder/show\\_reminders'\n\t+ 'weather/checkSunrise'\n\t+ 'weather/checkSunset'\n\t+ 'weather/find'\n* 'idx': an 'int32' feature",
"#### FMTODeu\\_slot\n\n\n* 'tokens': a list of 'string' features\n* 'tags': a list of intent labels, with possible values including:\n\t+ 'datetime'\n\t+ 'location'\n\t+ 'negation'\n\t+ 'alarm/alarm\\_modifier'\n\t+ 'alarm/recurring\\_period'\n\t+ 'reminder/noun'\n\t+ 'reminder/todo'\n\t+ 'reminder/reference'\n\t+ 'reminder/recurring\\_period'\n\t+ 'weather/attribute'\n\t+ 'weather/noun'\n* 'idx': an 'int32' feature",
"#### BHTCv2\n\n\n* 'text': a 'string' feature\n* 'label': a polarity label, with possible values including 'neutral' (NEU), 'negative' (N), 'positive' (P)\n* 'idx': an 'int32' feature",
"#### BEC2016eu\n\n\n* 'text': a 'string' feature\n* 'label': a topic label, with possible values including:\n\t+ 'Ekonomia'\n\t+ 'Euskal Herria'\n\t+ 'Euskara'\n\t+ 'Gizartea'\n\t+ 'Historia'\n\t+ 'Ingurumena'\n\t+ 'Iritzia'\n\t+ 'Komunikazioa'\n\t+ 'Kultura'\n\t+ 'Nazioartea'\n\t+ 'Politika'\n\t+ 'Zientzia'\n* 'idx': an 'int32' feature",
"#### VaxxStance\n\n\n* 'text': a 'string' feature\n* 'label': a stance label, with possible values including 'AGAINST', 'FAVOR', 'NONE'\n* 'idx': an 'int32' feature",
"#### QNLIeu\n\n\n* 'question': a 'string' feature\n* 'sentence': a 'string' feature\n* 'label': an entailment label, with possible values including 'entailment', 'not\\_entailment'\n* 'idx': an 'int32' feature",
"#### WiCeu\n\n\n* 'word': a 'string' feature\n* 'sentence1': a 'string' feature\n* 'sentence2': a 'string' feature\n* 'label': a 'boolean' label indicating sense agreement, with possible values including 'true', 'false'\n* 'start1': an 'int' feature indicating character position where word occurence begins in first sentence\n* 'start2': an 'int' feature indicating character position where word occurence begins in second sentence\n* 'end1': an 'int' feature indicating character position where word occurence ends in first sentence\n* 'end2': an 'int' feature indicating character position where word occurence ends in second sentence\n* 'idx': an 'int32' feature",
"#### EpecKorrefBin\n\n\n* 'text': a 'string' feature.\n* 'label': a 'boolean' coreference label, with possible values including 'true', 'false'.\n* 'span1\\_text': a 'string' feature\n* 'span2\\_text': a 'string' feature\n* 'span1\\_index': an 'int' feature indicating token index where 'span1\\_text' feature occurs in 'text'\n* 'span2\\_index': an 'int' feature indicating token index where 'span2\\_text' feature occurs in 'text'\n* 'idx': an 'int32' feature",
"### Data Splits\n\n\n\nDataset Creation\n----------------",
"### Curation Rationale\n\n\nWe believe that BasqueGLUE is a significant contribution towards developing NLU tools in Basque, which we believe will facilitate the technological advance for the Basque language. In order to create BasqueGLUE we took as a reference the GLUE and SuperGLUE frameworks. When possible, we re-used existing datasets for Basque, adapting them to the corresponding task formats if necessary. Additionally, BasqueGLUE also includes six new datasets that have not been published before. In total, BasqueGLUE consists of nine Basque NLU tasks and covers a wide range of tasks with different difficulties across several domains. As with the original GLUE benchmark, the training data for the tasks vary in size, which allows to measure the performance of how the models transfer knowledge across tasks.",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations\n\n\nAdditional Information\n----------------------",
"### Dataset Curators\n\n\nGorka Urbizu [1], Iรฑaki San Vicente [1], Xabier Saralegi [1], Rodrigo Agerri [2] and Aitor Soroa [2]\n\n\nAffiliation of the authors:\n\n\n[1] orai NLP Technologies\n\n\n[2] HiTZ Center - Ixa, University of the Basque Country UPV/EHU",
"### Licensing Information\n\n\nEach dataset of the BasqueGLUE benchmark has it's own license (due to most of them being or being derived from already existing datasets). See their respective README files for details.\n\n\nHere we provide a brief summary of their licenses:\n\n\n\nFor the rest of the files of the benchmark, including the loading and evaluation scripts, the following license applies:\n\n\nCopyright (C) by Orai NLP Technologies.\nThis benchmark and evaluation scripts are licensed under the Creative Commons Attribution Share Alike 4.0\nInternational License (CC BY-SA 4.0). To view a copy of this license, visit URL",
"### Contributions\n\n\nThanks to @richplant for adding this dataset to hugginface."
] |
7eb458284a4cf38dd8baf9f2697b46682b9e168b | # Dataset Card for [Dataset Name]
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
### Dataset Summary
This data set contains over 6,000 medical terms and their wikipedia text. It is intended to be used on a downstream task that requires medical terms and their wikipedia explanation.
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
### Citation Information
[More Information Needed]
| gamino/wiki_medical_terms | [
"task_categories:text-classification",
"annotations_creators:other",
"language_creators:other",
"size_categories:1K<n<10K",
"language:en",
"license:gpl-3.0",
"medical",
"conditions",
"region:us"
] | 2022-12-20T15:25:02+00:00 | {"annotations_creators": ["other"], "language_creators": ["other"], "language": ["en"], "license": ["gpl-3.0"], "multilinguality": [], "size_categories": ["1K<n<10K"], "source_datasets": [], "task_categories": ["text-classification"], "task_ids": [], "pretty_name": "Medical terms and their wikipedia text", "tags": ["medical", "conditions"]} | 2022-12-20T16:23:58+00:00 | [] | [
"en"
] | TAGS
#task_categories-text-classification #annotations_creators-other #language_creators-other #size_categories-1K<n<10K #language-English #license-gpl-3.0 #medical #conditions #region-us
| # Dataset Card for [Dataset Name]
## Table of Contents
- Table of Contents
- Dataset Description
- Dataset Summary
- Dataset Structure
- Data Instances
- Data Fields
- Data Splits
### Dataset Summary
This data set contains over 6,000 medical terms and their wikipedia text. It is intended to be used on a downstream task that requires medical terms and their wikipedia explanation.
## Dataset Structure
### Data Instances
### Data Fields
### Data Splits
## Dataset Creation
### Curation Rationale
### Source Data
| [
"# Dataset Card for [Dataset Name]",
"## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n \n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits",
"### Dataset Summary\n\nThis data set contains over 6,000 medical terms and their wikipedia text. It is intended to be used on a downstream task that requires medical terms and their wikipedia explanation.",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data"
] | [
"TAGS\n#task_categories-text-classification #annotations_creators-other #language_creators-other #size_categories-1K<n<10K #language-English #license-gpl-3.0 #medical #conditions #region-us \n",
"# Dataset Card for [Dataset Name]",
"## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n \n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits",
"### Dataset Summary\n\nThis data set contains over 6,000 medical terms and their wikipedia text. It is intended to be used on a downstream task that requires medical terms and their wikipedia explanation.",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data"
] |
bc9a9801c2950301ace2f07d7574a7bcb5f75d3a | # Dataset Card for "banking77_openai_embeddings"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | argilla/banking77_openai_embeddings | [
"region:us"
] | 2022-12-20T17:45:38+00:00 | {"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "activate_my_card", "1": "age_limit", "2": "apple_pay_or_google_pay", "3": "atm_support", "4": "automatic_top_up", "5": "balance_not_updated_after_bank_transfer", "6": "balance_not_updated_after_cheque_or_cash_deposit", "7": "beneficiary_not_allowed", "8": "cancel_transfer", "9": "card_about_to_expire", "10": "card_acceptance", "11": "card_arrival", "12": "card_delivery_estimate", "13": "card_linking", "14": "card_not_working", "15": "card_payment_fee_charged", "16": "card_payment_not_recognised", "17": "card_payment_wrong_exchange_rate", "18": "card_swallowed", "19": "cash_withdrawal_charge", "20": "cash_withdrawal_not_recognised", "21": "change_pin", "22": "compromised_card", "23": "contactless_not_working", "24": "country_support", "25": "declined_card_payment", "26": "declined_cash_withdrawal", "27": "declined_transfer", "28": "direct_debit_payment_not_recognised", "29": "disposable_card_limits", "30": "edit_personal_details", "31": "exchange_charge", "32": "exchange_rate", "33": "exchange_via_app", "34": "extra_charge_on_statement", "35": "failed_transfer", "36": "fiat_currency_support", "37": "get_disposable_virtual_card", "38": "get_physical_card", "39": "getting_spare_card", "40": "getting_virtual_card", "41": "lost_or_stolen_card", "42": "lost_or_stolen_phone", "43": "order_physical_card", "44": "passcode_forgotten", "45": "pending_card_payment", "46": "pending_cash_withdrawal", "47": "pending_top_up", "48": "pending_transfer", "49": "pin_blocked", "50": "receiving_money", "51": "Refund_not_showing_up", "52": "request_refund", "53": "reverted_card_payment?", "54": "supported_cards_and_currencies", "55": "terminate_account", "56": "top_up_by_bank_transfer_charge", "57": "top_up_by_card_charge", "58": "top_up_by_cash_or_cheque", "59": "top_up_failed", "60": "top_up_limits", "61": "top_up_reverted", "62": "topping_up_by_card", "63": "transaction_charged_twice", "64": "transfer_fee_charged", "65": "transfer_into_account", "66": "transfer_not_received_by_recipient", "67": "transfer_timing", "68": "unable_to_verify_identity", "69": "verify_my_identity", "70": "verify_source_of_funds", "71": "verify_top_up", "72": "virtual_card_not_working", "73": "visa_or_mastercard", "74": "why_verify_identity", "75": "wrong_amount_of_cash_received", "76": "wrong_exchange_rate_for_cash_withdrawal"}}}}, {"name": "vectors", "struct": [{"name": "openai-text-embedding-ada-002", "sequence": "float64"}]}], "splits": [{"name": "test", "num_bytes": 1235118, "num_examples": 100}], "download_size": 1186634, "dataset_size": 1235118}} | 2022-12-20T17:45:54+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "banking77_openai_embeddings"
More Information needed | [
"# Dataset Card for \"banking77_openai_embeddings\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"banking77_openai_embeddings\"\n\nMore Information needed"
] |
8457eadd105a564f2de361ca1e63fd5517df7f03 | # Dataset Card for "broadclass_totaldataset_0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totaldataset_0 | [
"region:us"
] | 2022-12-20T18:32:33+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163060617.0, "num_examples": 389}, {"name": "test", "num_bytes": 41751285.0, "num_examples": 98}], "download_size": 137862429, "dataset_size": 204811902.0}} | 2022-12-22T14:09:57+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totaldataset_0"
More Information needed | [
"# Dataset Card for \"broadclass_totaldataset_0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totaldataset_0\"\n\nMore Information needed"
] |
54d97353ba2f6ef40ceda3bd923b673ecd97af8f | # Dataset Card for "broadclass_totalMapped0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totalMapped0 | [
"region:us"
] | 2022-12-20T18:32:54+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108739968, "num_examples": 389}, {"name": "test", "num_bytes": 27842684, "num_examples": 98}], "download_size": 137407543, "dataset_size": 136582652}} | 2022-12-22T14:10:49+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totalMapped0"
More Information needed | [
"# Dataset Card for \"broadclass_totalMapped0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totalMapped0\"\n\nMore Information needed"
] |
aa642a3ad6336c570f8f057df3dceea4b0d0bdb4 | # Dataset Card for "broadclass_totaldataset_1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totaldataset_1 | [
"region:us"
] | 2022-12-20T18:33:25+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163909442.0, "num_examples": 389}, {"name": "test", "num_bytes": 40149082.0, "num_examples": 98}], "download_size": 137402351, "dataset_size": 204058524.0}} | 2022-12-22T14:11:34+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totaldataset_1"
More Information needed | [
"# Dataset Card for \"broadclass_totaldataset_1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totaldataset_1\"\n\nMore Information needed"
] |
1e2eaa38592bf03531d58aafbd20aba7c00b3630 | # Dataset Card for "broadclass_totalMapped1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totalMapped1 | [
"region:us"
] | 2022-12-20T18:33:37+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109305676, "num_examples": 389}, {"name": "test", "num_bytes": 26774060, "num_examples": 98}], "download_size": 136824488, "dataset_size": 136079736}} | 2022-12-22T14:12:25+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totalMapped1"
More Information needed | [
"# Dataset Card for \"broadclass_totalMapped1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totalMapped1\"\n\nMore Information needed"
] |
45df2db80f1a880f55c41bc0556f4c512540bc68 | # Dataset Card for "broadclass_totaldataset_2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totaldataset_2 | [
"region:us"
] | 2022-12-20T18:34:08+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163848386.0, "num_examples": 390}, {"name": "test", "num_bytes": 40722720.0, "num_examples": 97}], "download_size": 137727655, "dataset_size": 204571106.0}} | 2022-12-22T14:13:16+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totaldataset_2"
More Information needed | [
"# Dataset Card for \"broadclass_totaldataset_2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totaldataset_2\"\n\nMore Information needed"
] |
09f54173a4777673bbfc9b1f0f63994da181d5f9 | # Dataset Card for "broadclass_totalMapped2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totalMapped2 | [
"region:us"
] | 2022-12-20T18:34:21+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109265512, "num_examples": 390}, {"name": "test", "num_bytes": 27156588, "num_examples": 97}], "download_size": 137259978, "dataset_size": 136422100}} | 2022-12-22T14:14:18+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totalMapped2"
More Information needed | [
"# Dataset Card for \"broadclass_totalMapped2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totalMapped2\"\n\nMore Information needed"
] |
c4bb8f894d94d9127cd0e5b3f26cf52edef2b0a7 | # Dataset Card for "broadclass_totaldataset_3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totaldataset_3 | [
"region:us"
] | 2022-12-20T18:34:55+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 164258465.0, "num_examples": 390}, {"name": "test", "num_bytes": 41859040.0, "num_examples": 97}], "download_size": 138753084, "dataset_size": 206117505.0}} | 2022-12-22T14:15:06+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totaldataset_3"
More Information needed | [
"# Dataset Card for \"broadclass_totaldataset_3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totaldataset_3\"\n\nMore Information needed"
] |
707943f2ae3d56e51e6310eaffdd9b8be110f910 | # Dataset Card for "broadclass_totalMapped3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totalMapped3 | [
"region:us"
] | 2022-12-20T18:35:16+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109539072, "num_examples": 390}, {"name": "test", "num_bytes": 27914744, "num_examples": 97}], "download_size": 138277700, "dataset_size": 137453816}} | 2022-12-22T14:16:02+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totalMapped3"
More Information needed | [
"# Dataset Card for \"broadclass_totalMapped3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totalMapped3\"\n\nMore Information needed"
] |
65db117afa540ff4a3f123a7812e8050b9cd84ad | # Dataset Card for "broadclass_totaldataset_4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totaldataset_4 | [
"region:us"
] | 2022-12-20T18:35:42+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 164137813.0, "num_examples": 390}, {"name": "test", "num_bytes": 41046167.0, "num_examples": 97}], "download_size": 137497490, "dataset_size": 205183980.0}} | 2022-12-22T14:16:51+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totaldataset_4"
More Information needed | [
"# Dataset Card for \"broadclass_totaldataset_4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totaldataset_4\"\n\nMore Information needed"
] |
c078fcd6898e03049f986ee8e12e9e1e766d3641 | # Dataset Card for "broadclass_totalMapped4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/broadclass_totalMapped4 | [
"region:us"
] | 2022-12-20T18:36:07+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109458336, "num_examples": 390}, {"name": "test", "num_bytes": 27372364, "num_examples": 97}], "download_size": 137658405, "dataset_size": 136830700}} | 2022-12-22T14:17:47+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "broadclass_totalMapped4"
More Information needed | [
"# Dataset Card for \"broadclass_totalMapped4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"broadclass_totalMapped4\"\n\nMore Information needed"
] |
0c7461de6866b86e95984aa61da865932affd117 |
# Overview
This file contains over 1.7m public tweets about Apple, Amazon, Google, Microsoft and Tesla stocks, published between 01/01/2015 and 31/12/2019.
| mjw/stock_market_tweets | [
"license:apache-2.0",
"region:us"
] | 2022-12-20T18:54:22+00:00 | {"license": "apache-2.0"} | 2022-12-20T19:01:40+00:00 | [] | [] | TAGS
#license-apache-2.0 #region-us
|
# Overview
This file contains over 1.7m public tweets about Apple, Amazon, Google, Microsoft and Tesla stocks, published between 01/01/2015 and 31/12/2019.
| [
"# Overview\n\nThis file contains over 1.7m public tweets about Apple, Amazon, Google, Microsoft and Tesla stocks, published between 01/01/2015 and 31/12/2019."
] | [
"TAGS\n#license-apache-2.0 #region-us \n",
"# Overview\n\nThis file contains over 1.7m public tweets about Apple, Amazon, Google, Microsoft and Tesla stocks, published between 01/01/2015 and 31/12/2019."
] |
2e6fedf42c9c104e83dfd95c3a453721e683e244 |
# Dataset Card for Mall.cz Product Reviews (Czech)
## Dataset Description
The dataset contains user reviews from Czech eshop <mall.cz>
Each review contains text, sentiment (positive/negative/neutral), and automatically-detected language (mostly Czech, occasionaly Slovak) using [lingua-py](https://github.com/pemistahl/lingua-py)
The dataset has in total (train+validation+test) 30,000 reviews. The data is balanced.
Train set has 8000 positive, 8000 neutral and 8000 negative reviews.
Validation and test set each have 1000 positive, 1000 neutral and 1000 negative reviews.
## Dataset Features
Each sample contains:
- `review_id`: unique string identifier of the review.
- `rating_str`: string representation of the rating - "pozitivnรญ" / "neutrรกlnรญ" / "negativnรญ"
- `rating_int`: integer representation of the rating (1=positive, 0=neutral, -1=negative)
- `comment_language`: language of the review (mostly "cs", occasionaly "sk")
- `comment`: the string of the review
## Dataset Source
The data is a processed adaptation of [Mall CZ corpus](https://liks.fav.zcu.cz/sentiment/).
The adaptation is label-balanced and adds automatically-detected language
| fewshot-goes-multilingual/cs_mall-product-reviews | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:cs",
"license:cc-by-nc-sa-3.0",
"region:us"
] | 2022-12-20T20:35:40+00:00 | {"annotations_creators": ["found"], "language_creators": ["found"], "language": ["cs"], "license": ["cc-by-nc-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "Mall.cz Product Reviews", "tags": []} | 2022-12-20T21:11:15+00:00 | [] | [
"cs"
] | TAGS
#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Czech #license-cc-by-nc-sa-3.0 #region-us
|
# Dataset Card for URL Product Reviews (Czech)
## Dataset Description
The dataset contains user reviews from Czech eshop <URL>
Each review contains text, sentiment (positive/negative/neutral), and automatically-detected language (mostly Czech, occasionaly Slovak) using lingua-py
The dataset has in total (train+validation+test) 30,000 reviews. The data is balanced.
Train set has 8000 positive, 8000 neutral and 8000 negative reviews.
Validation and test set each have 1000 positive, 1000 neutral and 1000 negative reviews.
## Dataset Features
Each sample contains:
- 'review_id': unique string identifier of the review.
- 'rating_str': string representation of the rating - "pozitivnรญ" / "neutrรกlnรญ" / "negativnรญ"
- 'rating_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)
- 'comment_language': language of the review (mostly "cs", occasionaly "sk")
- 'comment': the string of the review
## Dataset Source
The data is a processed adaptation of Mall CZ corpus.
The adaptation is label-balanced and adds automatically-detected language
| [
"# Dataset Card for URL Product Reviews (Czech)",
"## Dataset Description\n\nThe dataset contains user reviews from Czech eshop <URL>\nEach review contains text, sentiment (positive/negative/neutral), and automatically-detected language (mostly Czech, occasionaly Slovak) using lingua-py\nThe dataset has in total (train+validation+test) 30,000 reviews. The data is balanced.\n\nTrain set has 8000 positive, 8000 neutral and 8000 negative reviews.\nValidation and test set each have 1000 positive, 1000 neutral and 1000 negative reviews.",
"## Dataset Features\n\nEach sample contains:\n- 'review_id': unique string identifier of the review.\n- 'rating_str': string representation of the rating - \"pozitivnรญ\" / \"neutrรกlnรญ\" / \"negativnรญ\"\n- 'rating_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)\n- 'comment_language': language of the review (mostly \"cs\", occasionaly \"sk\")\n- 'comment': the string of the review",
"## Dataset Source\n\nThe data is a processed adaptation of Mall CZ corpus.\nThe adaptation is label-balanced and adds automatically-detected language"
] | [
"TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Czech #license-cc-by-nc-sa-3.0 #region-us \n",
"# Dataset Card for URL Product Reviews (Czech)",
"## Dataset Description\n\nThe dataset contains user reviews from Czech eshop <URL>\nEach review contains text, sentiment (positive/negative/neutral), and automatically-detected language (mostly Czech, occasionaly Slovak) using lingua-py\nThe dataset has in total (train+validation+test) 30,000 reviews. The data is balanced.\n\nTrain set has 8000 positive, 8000 neutral and 8000 negative reviews.\nValidation and test set each have 1000 positive, 1000 neutral and 1000 negative reviews.",
"## Dataset Features\n\nEach sample contains:\n- 'review_id': unique string identifier of the review.\n- 'rating_str': string representation of the rating - \"pozitivnรญ\" / \"neutrรกlnรญ\" / \"negativnรญ\"\n- 'rating_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)\n- 'comment_language': language of the review (mostly \"cs\", occasionaly \"sk\")\n- 'comment': the string of the review",
"## Dataset Source\n\nThe data is a processed adaptation of Mall CZ corpus.\nThe adaptation is label-balanced and adds automatically-detected language"
] |
6ced1d87a030915822b087bf539e6d5c658f1988 |
# Dataset Card for Czech Facebook comments
## Dataset Description
The dataset contains user comments from Facebook. Each comment contains text, sentiment (positive/negative/neutral).
The dataset has in total (train+validation+test) 6,600 reviews. The data is balanced.
## Dataset Features
Each sample contains:
- `comment_id`: unique string identifier of the comment.
- `sentiment_str`: string representation of the rating - "pozitivnรญ" / "neutrรกlnรญ" / "negativnรญ"
- `sentiment_int`: integer representation of the rating (1=positive, 0=neutral, -1=negative)
- `comment`: the string of the comment
## Dataset Source
The data is a processed adaptation of [Facebook CZ Corpus](https://liks.fav.zcu.cz/sentiment/).
This adaptation is label-balanced.
| fewshot-goes-multilingual/cs_facebook-comments | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:cs",
"license:cc-by-nc-sa-3.0",
"region:us"
] | 2022-12-20T21:52:21+00:00 | {"annotations_creators": ["found"], "language_creators": ["found"], "language": ["cs"], "license": ["cc-by-nc-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "Czech Facebook comments", "tags": []} | 2022-12-20T21:56:09+00:00 | [] | [
"cs"
] | TAGS
#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Czech #license-cc-by-nc-sa-3.0 #region-us
|
# Dataset Card for Czech Facebook comments
## Dataset Description
The dataset contains user comments from Facebook. Each comment contains text, sentiment (positive/negative/neutral).
The dataset has in total (train+validation+test) 6,600 reviews. The data is balanced.
## Dataset Features
Each sample contains:
- 'comment_id': unique string identifier of the comment.
- 'sentiment_str': string representation of the rating - "pozitivnรญ" / "neutrรกlnรญ" / "negativnรญ"
- 'sentiment_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)
- 'comment': the string of the comment
## Dataset Source
The data is a processed adaptation of Facebook CZ Corpus.
This adaptation is label-balanced.
| [
"# Dataset Card for Czech Facebook comments",
"## Dataset Description\n\nThe dataset contains user comments from Facebook. Each comment contains text, sentiment (positive/negative/neutral).\nThe dataset has in total (train+validation+test) 6,600 reviews. The data is balanced.",
"## Dataset Features\n\nEach sample contains:\n- 'comment_id': unique string identifier of the comment.\n- 'sentiment_str': string representation of the rating - \"pozitivnรญ\" / \"neutrรกlnรญ\" / \"negativnรญ\"\n- 'sentiment_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)\n- 'comment': the string of the comment",
"## Dataset Source\n\nThe data is a processed adaptation of Facebook CZ Corpus.\nThis adaptation is label-balanced."
] | [
"TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Czech #license-cc-by-nc-sa-3.0 #region-us \n",
"# Dataset Card for Czech Facebook comments",
"## Dataset Description\n\nThe dataset contains user comments from Facebook. Each comment contains text, sentiment (positive/negative/neutral).\nThe dataset has in total (train+validation+test) 6,600 reviews. The data is balanced.",
"## Dataset Features\n\nEach sample contains:\n- 'comment_id': unique string identifier of the comment.\n- 'sentiment_str': string representation of the rating - \"pozitivnรญ\" / \"neutrรกlnรญ\" / \"negativnรญ\"\n- 'sentiment_int': integer representation of the rating (1=positive, 0=neutral, -1=negative)\n- 'comment': the string of the comment",
"## Dataset Source\n\nThe data is a processed adaptation of Facebook CZ Corpus.\nThis adaptation is label-balanced."
] |
949c267ff409730e7c978385c51faee878ebecf6 | # Dataset Card for "phoneme_totaldataset_0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totaldataset_0 | [
"region:us"
] | 2022-12-20T22:25:27+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163223522.0, "num_examples": 389}, {"name": "test", "num_bytes": 41231058.0, "num_examples": 98}], "download_size": 138510939, "dataset_size": 204454580.0}} | 2022-12-20T22:26:07+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totaldataset_0"
More Information needed | [
"# Dataset Card for \"phoneme_totaldataset_0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totaldataset_0\"\n\nMore Information needed"
] |
caaf0525773145ea183d7dd91418fac52fb38610 | # Dataset Card for "phoneme_totalMapped0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totalMapped0 | [
"region:us"
] | 2022-12-20T22:26:38+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108844668, "num_examples": 389}, {"name": "test", "num_bytes": 27494376, "num_examples": 98}], "download_size": 137098876, "dataset_size": 136339044}} | 2022-12-20T22:27:13+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totalMapped0"
More Information needed | [
"# Dataset Card for \"phoneme_totalMapped0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totalMapped0\"\n\nMore Information needed"
] |
37b05c5c90b898bff6419cbf12fde4abe000d2a4 | # Dataset Card for "phoneme_totaldataset_1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totaldataset_1 | [
"region:us"
] | 2022-12-20T22:27:17+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 162301333.0, "num_examples": 389}, {"name": "test", "num_bytes": 40804994.0, "num_examples": 98}], "download_size": 136056009, "dataset_size": 203106327.0}} | 2022-12-20T22:27:58+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totaldataset_1"
More Information needed | [
"# Dataset Card for \"phoneme_totaldataset_1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totaldataset_1\"\n\nMore Information needed"
] |
5dfa6f0394581a2d85c391701c722846328f394f | # Dataset Card for "phoneme_totalMapped1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totalMapped1 | [
"region:us"
] | 2022-12-20T22:28:32+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108229416, "num_examples": 389}, {"name": "test", "num_bytes": 27210864, "num_examples": 98}], "download_size": 136239720, "dataset_size": 135440280}} | 2022-12-20T22:29:20+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totalMapped1"
More Information needed | [
"# Dataset Card for \"phoneme_totalMapped1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totalMapped1\"\n\nMore Information needed"
] |
33ae3da39c383ba99fa0fbca69326f9d8614b378 | # Dataset Card for "phoneme_totaldataset_2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totaldataset_2 | [
"region:us"
] | 2022-12-20T22:29:23+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163385611.0, "num_examples": 390}, {"name": "test", "num_bytes": 41691832.0, "num_examples": 97}], "download_size": 138543168, "dataset_size": 205077443.0}} | 2022-12-20T22:29:58+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totaldataset_2"
More Information needed | [
"# Dataset Card for \"phoneme_totaldataset_2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totaldataset_2\"\n\nMore Information needed"
] |
ef52c45cb327202f4efc8333d502d126f9722220 | # Dataset Card for "phoneme_totalMapped2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totalMapped2 | [
"region:us"
] | 2022-12-20T22:30:27+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108952456, "num_examples": 390}, {"name": "test", "num_bytes": 27801832, "num_examples": 97}], "download_size": 137410544, "dataset_size": 136754288}} | 2022-12-20T22:31:03+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totalMapped2"
More Information needed | [
"# Dataset Card for \"phoneme_totalMapped2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totalMapped2\"\n\nMore Information needed"
] |
7e730cd4009527ac068619feddfeaaf77c84f681 | # Dataset Card for "phoneme_totaldataset_3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totaldataset_3 | [
"region:us"
] | 2022-12-20T22:31:05+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 164620015.0, "num_examples": 390}, {"name": "test", "num_bytes": 40776038.0, "num_examples": 97}], "download_size": 137708673, "dataset_size": 205396053.0}} | 2022-12-20T22:31:40+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totaldataset_3"
More Information needed | [
"# Dataset Card for \"phoneme_totaldataset_3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totaldataset_3\"\n\nMore Information needed"
] |
b5600d6a4bb12bbb6eb25a264558c9a7e5cd440a | # Dataset Card for "phoneme_totalMapped3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totalMapped3 | [
"region:us"
] | 2022-12-20T22:32:10+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109775968, "num_examples": 390}, {"name": "test", "num_bytes": 27190896, "num_examples": 97}], "download_size": 137863961, "dataset_size": 136966864}} | 2022-12-20T22:32:44+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totalMapped3"
More Information needed | [
"# Dataset Card for \"phoneme_totalMapped3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totalMapped3\"\n\nMore Information needed"
] |
69786c5181e94bd044922ef86766ff987d98ee81 | # Dataset Card for "phoneme_totaldataset_4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totaldataset_4 | [
"region:us"
] | 2022-12-20T22:32:46+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 164035246.0, "num_examples": 390}, {"name": "test", "num_bytes": 40309237.0, "num_examples": 97}], "download_size": 137553091, "dataset_size": 204344483.0}} | 2022-12-20T22:33:27+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totaldataset_4"
More Information needed | [
"# Dataset Card for \"phoneme_totaldataset_4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totaldataset_4\"\n\nMore Information needed"
] |
630f798c73ca3237402ce4a95a0044bb0237a838 | # Dataset Card for "phoneme_totalMapped4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/phoneme_totalMapped4 | [
"region:us"
] | 2022-12-20T22:33:57+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 109385812, "num_examples": 390}, {"name": "test", "num_bytes": 26880084, "num_examples": 97}], "download_size": 137069553, "dataset_size": 136265896}} | 2022-12-20T22:34:32+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "phoneme_totalMapped4"
More Information needed | [
"# Dataset Card for \"phoneme_totalMapped4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"phoneme_totalMapped4\"\n\nMore Information needed"
] |
de7f547f4e31158e08357cb378b15215cbcdc4fd | masakhane/afriqa_wiki_en_fr_100 | [
"task_categories:text-retrieval",
"multilinguality:multilingual",
"language:en",
"language:fr",
"license:apache-2.0",
"region:us"
] | 2022-12-20T22:37:58+00:00 | {"language": ["en", "fr"], "license": "apache-2.0", "multilinguality": ["multilingual"], "task_categories": ["text-retrieval"], "pretty_name": "Wikipedia 100 word Passages", "viewer": true} | 2023-03-31T16:26:05+00:00 | [] | [
"en",
"fr"
] | TAGS
#task_categories-text-retrieval #multilinguality-multilingual #language-English #language-French #license-apache-2.0 #region-us
| [] | [
"TAGS\n#task_categories-text-retrieval #multilinguality-multilingual #language-English #language-French #license-apache-2.0 #region-us \n"
] |
||
986af848d5b2837153e837ddd59d33c50a98eeb0 | # AutoTrain Dataset for project: feetfoot
## Dataset Description
This dataset has been automatically processed by AutoTrain for project feetfoot.
### Languages
The BCP-47 code for the dataset's language is unk.
## Dataset Structure
### Data Instances
A sample from this dataset looks as follows:
```json
[
{
"image": "<180x320 RGB PIL image>",
"target": 0
},
{
"image": "<78x320 RGB PIL image>",
"target": 0
}
]
```
### Dataset Fields
The dataset has the following fields (also called "features"):
```json
{
"image": "Image(decode=True, id=None)",
"target": "ClassLabel(names=['gettyimagefeet'], id=None)"
}
```
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| Split name | Num samples |
| ------------ | ------------------- |
| train | 97 |
| valid | 25 |
| MrDre/autotrain-data-feetfoot | [
"task_categories:image-classification",
"region:us"
] | 2022-12-20T23:32:12+00:00 | {"task_categories": ["image-classification"]} | 2022-12-21T00:56:37+00:00 | [] | [] | TAGS
#task_categories-image-classification #region-us
| AutoTrain Dataset for project: feetfoot
=======================================
Dataset Description
-------------------
This dataset has been automatically processed by AutoTrain for project feetfoot.
### Languages
The BCP-47 code for the dataset's language is unk.
Dataset Structure
-----------------
### Data Instances
A sample from this dataset looks as follows:
### Dataset Fields
The dataset has the following fields (also called "features"):
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| [
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] | [
"TAGS\n#task_categories-image-classification #region-us \n",
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] |
d533f626cc321c92175a58ee570aa3cdb87238d1 |
# Model-Written Evaluation Datasets
This repository includes datasets written by language models, used in our paper on "Discovering Language Model Behaviors with Model-Written Evaluations."
We intend the datasets to be useful to:
1. Those who are interested in understanding the quality and properties of model-generated data
2. Those who wish to use our datasets to evaluate other models for the behaviors we examined in our work (e.g., related to model persona, sycophancy, advanced AI risks, and gender bias)
The evaluations were generated to be asked to dialogue agents (e.g., a model finetuned explicitly respond to a user's utterances, or a pretrained language model prompted to behave like a dialogue agent). However, it is possible to adapt the data to test other kinds of models as well.
We describe each of our collections of datasets below:
1. `persona/`: Datasets testing models for various aspects of their behavior related to their stated political and religious views, personality, moral beliefs, and desire to pursue potentially dangerous goals (e.g., self-preservation or power-seeking).
2. `sycophancy/`: Datasets testing models for whether or not they repeat back a user's view to various questions (in philosophy, NLP research, and politics)
3. `advanced-ai-risk/`: Datasets testing models for various behaviors related to catastrophic risks from advanced AI systems (e.g., ). These datasets were generated in a few-shot manner. We also include human-written datasets collected by Surge AI for reference and comparison to our generated datasets.
4. `winogenerated/`: Our larger, model-generated version of the Winogender Dataset ([Rudinger et al., 2018](https://arxiv.org/abs/1804.09301)). We also include the names of occupation titles that we generated, to create the dataset (alongside occupation gender statistics from the Bureau of Labor Statistics)
Please see our paper for additional details on the datasets, how we generated them, human validation metrics, and other analyses of the datasets.
**Disclaimer**: As discussed in our paper, some data contains content that includes social biases and stereotypes. The data may also contain other forms of harmful or offensive content. The views expressed in the data do not reflect the views of Anthropic or any of its employees.
## Contact
For questions, please email `ethan at anthropic dot com`
## Bibtex Citation
If you would like to cite our work or data, you may use the following bibtex citation:
```
@misc{perez2022discovering,
doi = {10.48550/ARXIV.2212.09251},
url = {https://arxiv.org/abs/2212.09251},
author = {Perez, Ethan and Ringer, Sam and Lukoลกiลซtฤ, Kamilฤ and Nguyen, Karina and Chen, Edwin and Heiner, Scott and Pettit, Craig and Olsson, Catherine and Kundu, Sandipan and Kadavath, Saurav and Jones, Andy and Chen, Anna and Mann, Ben and Israel, Brian and Seethor, Bryan and McKinnon, Cameron and Olah, Christopher and Yan, Da and Amodei, Daniela and Amodei, Dario and Drain, Dawn and Li, Dustin and Tran-Johnson, Eli and Khundadze, Guro and Kernion, Jackson and Landis, James and Kerr, Jamie and Mueller, Jared and Hyun, Jeeyoon and Landau, Joshua and Ndousse, Kamal and Goldberg, Landon and Lovitt, Liane and Lucas, Martin and Sellitto, Michael and Zhang, Miranda and Kingsland, Neerav and Elhage, Nelson and Joseph, Nicholas and Mercado, Noemรญ and DasSarma, Nova and Rausch, Oliver and Larson, Robin and McCandlish, Sam and Johnston, Scott and Kravec, Shauna and {El Showk}, Sheer and Lanham, Tamera and Telleen-Lawton, Timothy and Brown, Tom and Henighan, Tom and Hume, Tristan and Bai, Yuntao and Hatfield-Dodds, Zac and Clark, Jack and Bowman, Samuel R. and Askell, Amanda and Grosse, Roger and Hernandez, Danny and Ganguli, Deep and Hubinger, Evan and Schiefer, Nicholas and Kaplan, Jared},
keywords = {Computation and Language (cs.CL), Artificial Intelligence (cs.AI), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Discovering Language Model Behaviors with Model-Written Evaluations},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
| Anthropic/model-written-evals | [
"task_categories:multiple-choice",
"task_categories:zero-shot-classification",
"task_categories:question-answering",
"task_ids:multiple-choice-qa",
"task_ids:multiple-choice-coreference-resolution",
"annotations_creators:machine-generated",
"language_creators:machine-generated",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:cc-by-4.0",
"gender bias",
"social bias",
"AI safety",
"personality",
"politics",
"arxiv:1804.09301",
"arxiv:2212.09251",
"region:us"
] | 2022-12-21T00:01:13+00:00 | {"annotations_creators": ["machine-generated"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["multiple-choice", "zero-shot-classification", "question-answering"], "task_ids": ["multiple-choice-qa", "multiple-choice-coreference-resolution"], "pretty_name": "Evaluations from \"Discovering Language Model Behaviors with Model-Written Evaluations\"", "tags": ["gender bias", "social bias", "AI safety", "personality", "politics"]} | 2022-12-21T02:33:18+00:00 | [
"1804.09301",
"2212.09251"
] | [
"en"
] | TAGS
#task_categories-multiple-choice #task_categories-zero-shot-classification #task_categories-question-answering #task_ids-multiple-choice-qa #task_ids-multiple-choice-coreference-resolution #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #gender bias #social bias #AI safety #personality #politics #arxiv-1804.09301 #arxiv-2212.09251 #region-us
|
# Model-Written Evaluation Datasets
This repository includes datasets written by language models, used in our paper on "Discovering Language Model Behaviors with Model-Written Evaluations."
We intend the datasets to be useful to:
1. Those who are interested in understanding the quality and properties of model-generated data
2. Those who wish to use our datasets to evaluate other models for the behaviors we examined in our work (e.g., related to model persona, sycophancy, advanced AI risks, and gender bias)
The evaluations were generated to be asked to dialogue agents (e.g., a model finetuned explicitly respond to a user's utterances, or a pretrained language model prompted to behave like a dialogue agent). However, it is possible to adapt the data to test other kinds of models as well.
We describe each of our collections of datasets below:
1. 'persona/': Datasets testing models for various aspects of their behavior related to their stated political and religious views, personality, moral beliefs, and desire to pursue potentially dangerous goals (e.g., self-preservation or power-seeking).
2. 'sycophancy/': Datasets testing models for whether or not they repeat back a user's view to various questions (in philosophy, NLP research, and politics)
3. 'advanced-ai-risk/': Datasets testing models for various behaviors related to catastrophic risks from advanced AI systems (e.g., ). These datasets were generated in a few-shot manner. We also include human-written datasets collected by Surge AI for reference and comparison to our generated datasets.
4. 'winogenerated/': Our larger, model-generated version of the Winogender Dataset (Rudinger et al., 2018). We also include the names of occupation titles that we generated, to create the dataset (alongside occupation gender statistics from the Bureau of Labor Statistics)
Please see our paper for additional details on the datasets, how we generated them, human validation metrics, and other analyses of the datasets.
Disclaimer: As discussed in our paper, some data contains content that includes social biases and stereotypes. The data may also contain other forms of harmful or offensive content. The views expressed in the data do not reflect the views of Anthropic or any of its employees.
## Contact
For questions, please email 'ethan at anthropic dot com'
## Bibtex Citation
If you would like to cite our work or data, you may use the following bibtex citation:
| [
"# Model-Written Evaluation Datasets\n\nThis repository includes datasets written by language models, used in our paper on \"Discovering Language Model Behaviors with Model-Written Evaluations.\"\n\nWe intend the datasets to be useful to:\n1. Those who are interested in understanding the quality and properties of model-generated data\n2. Those who wish to use our datasets to evaluate other models for the behaviors we examined in our work (e.g., related to model persona, sycophancy, advanced AI risks, and gender bias)\n\nThe evaluations were generated to be asked to dialogue agents (e.g., a model finetuned explicitly respond to a user's utterances, or a pretrained language model prompted to behave like a dialogue agent). However, it is possible to adapt the data to test other kinds of models as well.\n\nWe describe each of our collections of datasets below:\n1. 'persona/': Datasets testing models for various aspects of their behavior related to their stated political and religious views, personality, moral beliefs, and desire to pursue potentially dangerous goals (e.g., self-preservation or power-seeking).\n2. 'sycophancy/': Datasets testing models for whether or not they repeat back a user's view to various questions (in philosophy, NLP research, and politics)\n3. 'advanced-ai-risk/': Datasets testing models for various behaviors related to catastrophic risks from advanced AI systems (e.g., ). These datasets were generated in a few-shot manner. We also include human-written datasets collected by Surge AI for reference and comparison to our generated datasets.\n4. 'winogenerated/': Our larger, model-generated version of the Winogender Dataset (Rudinger et al., 2018). We also include the names of occupation titles that we generated, to create the dataset (alongside occupation gender statistics from the Bureau of Labor Statistics)\n\nPlease see our paper for additional details on the datasets, how we generated them, human validation metrics, and other analyses of the datasets.\n\nDisclaimer: As discussed in our paper, some data contains content that includes social biases and stereotypes. The data may also contain other forms of harmful or offensive content. The views expressed in the data do not reflect the views of Anthropic or any of its employees.",
"## Contact\nFor questions, please email 'ethan at anthropic dot com'",
"## Bibtex Citation\n\nIf you would like to cite our work or data, you may use the following bibtex citation:"
] | [
"TAGS\n#task_categories-multiple-choice #task_categories-zero-shot-classification #task_categories-question-answering #task_ids-multiple-choice-qa #task_ids-multiple-choice-coreference-resolution #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #gender bias #social bias #AI safety #personality #politics #arxiv-1804.09301 #arxiv-2212.09251 #region-us \n",
"# Model-Written Evaluation Datasets\n\nThis repository includes datasets written by language models, used in our paper on \"Discovering Language Model Behaviors with Model-Written Evaluations.\"\n\nWe intend the datasets to be useful to:\n1. Those who are interested in understanding the quality and properties of model-generated data\n2. Those who wish to use our datasets to evaluate other models for the behaviors we examined in our work (e.g., related to model persona, sycophancy, advanced AI risks, and gender bias)\n\nThe evaluations were generated to be asked to dialogue agents (e.g., a model finetuned explicitly respond to a user's utterances, or a pretrained language model prompted to behave like a dialogue agent). However, it is possible to adapt the data to test other kinds of models as well.\n\nWe describe each of our collections of datasets below:\n1. 'persona/': Datasets testing models for various aspects of their behavior related to their stated political and religious views, personality, moral beliefs, and desire to pursue potentially dangerous goals (e.g., self-preservation or power-seeking).\n2. 'sycophancy/': Datasets testing models for whether or not they repeat back a user's view to various questions (in philosophy, NLP research, and politics)\n3. 'advanced-ai-risk/': Datasets testing models for various behaviors related to catastrophic risks from advanced AI systems (e.g., ). These datasets were generated in a few-shot manner. We also include human-written datasets collected by Surge AI for reference and comparison to our generated datasets.\n4. 'winogenerated/': Our larger, model-generated version of the Winogender Dataset (Rudinger et al., 2018). We also include the names of occupation titles that we generated, to create the dataset (alongside occupation gender statistics from the Bureau of Labor Statistics)\n\nPlease see our paper for additional details on the datasets, how we generated them, human validation metrics, and other analyses of the datasets.\n\nDisclaimer: As discussed in our paper, some data contains content that includes social biases and stereotypes. The data may also contain other forms of harmful or offensive content. The views expressed in the data do not reflect the views of Anthropic or any of its employees.",
"## Contact\nFor questions, please email 'ethan at anthropic dot com'",
"## Bibtex Citation\n\nIf you would like to cite our work or data, you may use the following bibtex citation:"
] |
a62ed8b6e1ab875c90ea09b67a1842adbe7ccf49 | # AutoTrain Dataset for project: feets
## Dataset Description
This dataset has been automatically processed by AutoTrain for project feets.
### Languages
The BCP-47 code for the dataset's language is unk.
## Dataset Structure
### Data Instances
A sample from this dataset looks as follows:
```json
[
{
"image": "<206x320 RGB PIL image>",
"target": 0
},
{
"image": "<173x320 RGB PIL image>",
"target": 0
}
]
```
### Dataset Fields
The dataset has the following fields (also called "features"):
```json
{
"image": "Image(decode=True, id=None)",
"target": "ClassLabel(names=['gettyimagefeet'], id=None)"
}
```
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| Split name | Num samples |
| ------------ | ------------------- |
| train | 122 |
| valid | 122 |
| MrDre/autotrain-data-feets | [
"task_categories:image-classification",
"region:us"
] | 2022-12-21T00:58:14+00:00 | {"task_categories": ["image-classification"]} | 2022-12-21T01:01:27+00:00 | [] | [] | TAGS
#task_categories-image-classification #region-us
| AutoTrain Dataset for project: feets
====================================
Dataset Description
-------------------
This dataset has been automatically processed by AutoTrain for project feets.
### Languages
The BCP-47 code for the dataset's language is unk.
Dataset Structure
-----------------
### Data Instances
A sample from this dataset looks as follows:
### Dataset Fields
The dataset has the following fields (also called "features"):
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| [
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] | [
"TAGS\n#task_categories-image-classification #region-us \n",
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] |
e5926b7fa7eee56d216a86f5dc599300c8ecf09c | ## Data for the annotation project
* `short_files.json` with short code files (around 50 lines).
* `medium_files.json` with medium sized files (around 120 lines).
* guidelines for the annotation. | loubnabnl/data_toloka | [
"language:code",
"region:us"
] | 2022-12-21T01:33:43+00:00 | {"language": ["code"]} | 2022-12-21T01:39:36+00:00 | [] | [
"code"
] | TAGS
#language-code #region-us
| ## Data for the annotation project
* 'short_files.json' with short code files (around 50 lines).
* 'medium_files.json' with medium sized files (around 120 lines).
* guidelines for the annotation. | [
"## Data for the annotation project\n\n* 'short_files.json' with short code files (around 50 lines).\n* 'medium_files.json' with medium sized files (around 120 lines).\n* guidelines for the annotation."
] | [
"TAGS\n#language-code #region-us \n",
"## Data for the annotation project\n\n* 'short_files.json' with short code files (around 50 lines).\n* 'medium_files.json' with medium sized files (around 120 lines).\n* guidelines for the annotation."
] |
b3b0eaee79340c923e972331f00b272f36e3288d | # Dataset Card for "syllable_totaldataset_0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totaldataset_0 | [
"region:us"
] | 2022-12-21T02:22:57+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 162800048.0, "num_examples": 389}, {"name": "test", "num_bytes": 40702416.0, "num_examples": 98}], "download_size": 136515053, "dataset_size": 203502464.0}} | 2022-12-21T02:23:47+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totaldataset_0"
More Information needed | [
"# Dataset Card for \"syllable_totaldataset_0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totaldataset_0\"\n\nMore Information needed"
] |
5ceff2585490e68226395469f3a1a72f0b2da10d | # Dataset Card for "syllable_totalMapped0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totalMapped0 | [
"region:us"
] | 2022-12-21T02:24:28+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108518696, "num_examples": 389}, {"name": "test", "num_bytes": 27131260, "num_examples": 98}], "download_size": 136632106, "dataset_size": 135649956}} | 2022-12-21T02:25:07+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totalMapped0"
More Information needed | [
"# Dataset Card for \"syllable_totalMapped0\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totalMapped0\"\n\nMore Information needed"
] |
ad30e09bbd73bbf8116cbb23beefc597adee2078 | # Dataset Card for "syllable_totaldataset_1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totaldataset_1 | [
"region:us"
] | 2022-12-21T02:25:09+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 165091979.0, "num_examples": 389}, {"name": "test", "num_bytes": 40724378.0, "num_examples": 98}], "download_size": 137998530, "dataset_size": 205816357.0}} | 2022-12-21T02:25:49+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totaldataset_1"
More Information needed | [
"# Dataset Card for \"syllable_totaldataset_1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totaldataset_1\"\n\nMore Information needed"
] |
84eb4640ef92a75256ec7ce1fb345321a1f476cc | # Dataset Card for "syllable_totalMapped1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totalMapped1 | [
"region:us"
] | 2022-12-21T02:26:22+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 110046848, "num_examples": 389}, {"name": "test", "num_bytes": 27145836, "num_examples": 98}], "download_size": 138090941, "dataset_size": 137192684}} | 2022-12-21T02:27:04+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totalMapped1"
More Information needed | [
"# Dataset Card for \"syllable_totalMapped1\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totalMapped1\"\n\nMore Information needed"
] |
ee9a69b329ff595708286fb087c0862a19267566 | # Dataset Card for "syllable_totaldataset_2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totaldataset_2 | [
"region:us"
] | 2022-12-21T02:27:06+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163378263.0, "num_examples": 390}, {"name": "test", "num_bytes": 40462578.0, "num_examples": 97}], "download_size": 138321082, "dataset_size": 203840841.0}} | 2022-12-21T02:27:47+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totaldataset_2"
More Information needed | [
"# Dataset Card for \"syllable_totaldataset_2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totaldataset_2\"\n\nMore Information needed"
] |
f70ac9a400adfb1d12a92a041cb80e76e275a4a0 | # Dataset Card for "syllable_totalMapped2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totalMapped2 | [
"region:us"
] | 2022-12-21T02:28:20+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108903948, "num_examples": 390}, {"name": "test", "num_bytes": 26971340, "num_examples": 97}], "download_size": 136776788, "dataset_size": 135875288}} | 2022-12-21T02:29:05+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totalMapped2"
More Information needed | [
"# Dataset Card for \"syllable_totalMapped2\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totalMapped2\"\n\nMore Information needed"
] |
429d40a1f2022f1c829580bc6e1c44e5efb32fad | # Dataset Card for "syllable_totaldataset_3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totaldataset_3 | [
"region:us"
] | 2022-12-21T02:29:08+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 162920604.0, "num_examples": 390}, {"name": "test", "num_bytes": 40471857.0, "num_examples": 97}], "download_size": 137189267, "dataset_size": 203392461.0}} | 2022-12-21T02:29:50+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totaldataset_3"
More Information needed | [
"# Dataset Card for \"syllable_totaldataset_3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totaldataset_3\"\n\nMore Information needed"
] |
0673d51722c18ef84c503418bd7cbf9c5333938c | # Dataset Card for "syllable_totalMapped3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totalMapped3 | [
"region:us"
] | 2022-12-21T02:30:22+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108599016, "num_examples": 390}, {"name": "test", "num_bytes": 26977548, "num_examples": 97}], "download_size": 136574643, "dataset_size": 135576564}} | 2022-12-21T02:31:03+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totalMapped3"
More Information needed | [
"# Dataset Card for \"syllable_totalMapped3\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totalMapped3\"\n\nMore Information needed"
] |
7307aad6edb5f41e73a9aaaa586e52cc947a851f | # Dataset Card for "syllable_totaldataset_4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totaldataset_4 | [
"region:us"
] | 2022-12-21T02:31:06+00:00 | {"dataset_info": {"features": [{"name": "name", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "label", "dtype": "string"}, {"name": "emotion", "dtype": "string"}, {"name": "emotion_str", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 163180696.0, "num_examples": 390}, {"name": "test", "num_bytes": 41085347.0, "num_examples": 97}], "download_size": 137671411, "dataset_size": 204266043.0}} | 2022-12-21T02:31:44+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totaldataset_4"
More Information needed | [
"# Dataset Card for \"syllable_totaldataset_4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totaldataset_4\"\n\nMore Information needed"
] |
3c17d2e8969e5398ffeb3acd868a42c8f27c1ca8 | # Dataset Card for "syllable_totalMapped4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | JovialValley/syllable_totalMapped4 | [
"region:us"
] | 2022-12-21T02:32:16+00:00 | {"dataset_info": {"features": [{"name": "input_values", "sequence": "float32"}, {"name": "labels", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 108772580, "num_examples": 390}, {"name": "test", "num_bytes": 27386468, "num_examples": 97}], "download_size": 137043673, "dataset_size": 136159048}} | 2022-12-21T02:32:53+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "syllable_totalMapped4"
More Information needed | [
"# Dataset Card for \"syllable_totalMapped4\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"syllable_totalMapped4\"\n\nMore Information needed"
] |
41fd32ba6ea7a0e337ce9a1d4a4339dd576e28dc |
# Germeval Task 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback
In the connected, modern world, customer feedback is a valuable source for insights on the quality of products or services. This feedback allows other customers to benefit from the experiences of others and enables businesses to react on requests, complaints or recommendations. However, the more people use a product or service, the more feedback is generated, which results in the major challenge of analyzing huge amounts of feedback in an efficient, but still meaningful way.
Thus, we propose a shared task on automatically analyzing customer reviews about โDeutsche Bahnโ - the german public train operator with about two billion passengers each year.
Example:
> โRT @XXX: Da hรถrt jemand in der Bahn so laut โ700 Main Streetโ durch seine Kopfhรถrer, dass ich mithรถren kann. :( :( :(โ
As shown in the example, insights from reviews can be derived on different granularities. The review contains a general evaluation of the travel (The customer disliked the travel). Furthermore, the review evaluates a dedicated aspect of the train travel (โlautโ โ customer did not like the noise level).
Consequently, we frame the task as aspect-based sentiment analysis with four sub tasks:
## Data format
```
ID <tab> Text <tab> Relevance <tab> Sentiment <tab> Aspect:Polarity (whitespace separated)
```
## Links
- http://ltdata1.informatik.uni-hamburg.de/germeval2017/
- https://sites.google.com/view/germeval2017-absa/
## How to cite
```bibtex
@inproceedings{germevaltask2017,
title = {{GermEval 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback}},
author = {Michael Wojatzki and Eugen Ruppert and Sarah Holschneider and Torsten Zesch and Chris Biemann},
year = {2017},
booktitle = {Proceedings of the GermEval 2017 โ Shared Task on Aspect-based Sentiment in Social Media Customer Feedback},
address={Berlin, Germany},
pages={1--12}
}
``` | akash418/germeval_2017 | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"task_ids:topic-classification",
"language:de",
"region:us"
] | 2022-12-21T02:58:51+00:00 | {"annotations_creators": [], "language_creators": [], "language": ["de"], "license": [], "multilinguality": [], "size_categories": [], "source_datasets": [], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification", "topic-classification"], "pretty_name": "German Eval 2017 ", "tags": []} | 2022-12-21T03:43:47+00:00 | [] | [
"de"
] | TAGS
#task_categories-text-classification #task_ids-multi-class-classification #task_ids-topic-classification #language-German #region-us
|
# Germeval Task 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback
In the connected, modern world, customer feedback is a valuable source for insights on the quality of products or services. This feedback allows other customers to benefit from the experiences of others and enables businesses to react on requests, complaints or recommendations. However, the more people use a product or service, the more feedback is generated, which results in the major challenge of analyzing huge amounts of feedback in an efficient, but still meaningful way.
Thus, we propose a shared task on automatically analyzing customer reviews about โDeutsche Bahnโ - the german public train operator with about two billion passengers each year.
Example:
> โRT @XXX: Da hรถrt jemand in der Bahn so laut โ700 Main Streetโ durch seine Kopfhรถrer, dass ich mithรถren kann. :( :( :(โ
As shown in the example, insights from reviews can be derived on different granularities. The review contains a general evaluation of the travel (The customer disliked the travel). Furthermore, the review evaluates a dedicated aspect of the train travel (โlautโ โ customer did not like the noise level).
Consequently, we frame the task as aspect-based sentiment analysis with four sub tasks:
## Data format
## Links
- URL
- URL
## How to cite
| [
"# Germeval Task 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback\n\nIn the connected, modern world, customer feedback is a valuable source for insights on the quality of products or services. This feedback allows other customers to benefit from the experiences of others and enables businesses to react on requests, complaints or recommendations. However, the more people use a product or service, the more feedback is generated, which results in the major challenge of analyzing huge amounts of feedback in an efficient, but still meaningful way.\n\nThus, we propose a shared task on automatically analyzing customer reviews about โDeutsche Bahnโ - the german public train operator with about two billion passengers each year.\n\nExample: \n\n> โRT @XXX: Da hรถrt jemand in der Bahn so laut โ700 Main Streetโ durch seine Kopfhรถrer, dass ich mithรถren kann. :( :( :(โ\n\nAs shown in the example, insights from reviews can be derived on different granularities. The review contains a general evaluation of the travel (The customer disliked the travel). Furthermore, the review evaluates a dedicated aspect of the train travel (โlautโ โ customer did not like the noise level).\n\nConsequently, we frame the task as aspect-based sentiment analysis with four sub tasks:",
"## Data format",
"## Links\n\n- URL\n- URL",
"## How to cite"
] | [
"TAGS\n#task_categories-text-classification #task_ids-multi-class-classification #task_ids-topic-classification #language-German #region-us \n",
"# Germeval Task 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback\n\nIn the connected, modern world, customer feedback is a valuable source for insights on the quality of products or services. This feedback allows other customers to benefit from the experiences of others and enables businesses to react on requests, complaints or recommendations. However, the more people use a product or service, the more feedback is generated, which results in the major challenge of analyzing huge amounts of feedback in an efficient, but still meaningful way.\n\nThus, we propose a shared task on automatically analyzing customer reviews about โDeutsche Bahnโ - the german public train operator with about two billion passengers each year.\n\nExample: \n\n> โRT @XXX: Da hรถrt jemand in der Bahn so laut โ700 Main Streetโ durch seine Kopfhรถrer, dass ich mithรถren kann. :( :( :(โ\n\nAs shown in the example, insights from reviews can be derived on different granularities. The review contains a general evaluation of the travel (The customer disliked the travel). Furthermore, the review evaluates a dedicated aspect of the train travel (โlautโ โ customer did not like the noise level).\n\nConsequently, we frame the task as aspect-based sentiment analysis with four sub tasks:",
"## Data format",
"## Links\n\n- URL\n- URL",
"## How to cite"
] |
828caf059ef103f1545b9e384c910a3e128bea04 |
For the sake of full disclosure I publish the dataset that I use to train [Crosstyan/BPModel](https://huggingface.co/Crosstyan/BPModel).
NSFW content is contained. Watch with your parents if you don't feel comfortable about that. | Crosstyan/BPDataset | [
"size_categories:1K<n<10K",
"license:openrail",
"not-for-all-audiences",
"region:us"
] | 2022-12-21T03:11:52+00:00 | {"license": "openrail", "size_categories": ["1K<n<10K"], "tags": ["not-for-all-audiences"]} | 2023-12-04T18:06:36+00:00 | [] | [] | TAGS
#size_categories-1K<n<10K #license-openrail #not-for-all-audiences #region-us
|
For the sake of full disclosure I publish the dataset that I use to train Crosstyan/BPModel.
NSFW content is contained. Watch with your parents if you don't feel comfortable about that. | [] | [
"TAGS\n#size_categories-1K<n<10K #license-openrail #not-for-all-audiences #region-us \n"
] |
77a26a5640739f8b9ca5039b599cf14c69fc3267 | Dataset homepage: https://sites.google.com/site/redwebcvpr18/ | sayakpaul/ReDWeb | [
"region:us"
] | 2022-12-21T03:21:29+00:00 | {} | 2022-12-21T03:26:51+00:00 | [] | [] | TAGS
#region-us
| Dataset homepage: URL | [] | [
"TAGS\n#region-us \n"
] |
ea5f1b98cab4e1891fa483871e0bdd2ea42ce870 | # Dataset Card for "IAM"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | gagan3012/IAM | [
"region:us"
] | 2022-12-21T05:12:11+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "Noto_Sans_Arabic", "1": "Readex_Pro", "2": "Amiri", "3": "Noto_Kufi_Arabic", "4": "Reem_Kufi_Fun", "5": "Lateef", "6": "Changa", "7": "Kufam", "8": "ElMessiri", "9": "Reem_Kufi", "10": "Noto_Naskh_Arabic", "11": "Reem_Kufi_Ink", "12": "Tajawal", "13": "Aref_Ruqaa_Ink", "14": "Markazi_Text", "15": "IBM_Plex_Sans_Arabic", "16": "Vazirmatn", "17": "Harmattan", "18": "Gulzar", "19": "Scheherazade_New", "20": "Cairo", "21": "Amiri_Quran", "22": "Noto_Nastaliq_Urdu", "23": "Mada", "24": "Aref_Ruqaa", "25": "Almarai", "26": "Alkalami", "27": "Qahiri"}}}}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 563851079.0, "num_examples": 11344}], "download_size": 563727207, "dataset_size": 563851079.0}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-13T17:13:25+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "IAM"
More Information needed | [
"# Dataset Card for \"IAM\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"IAM\"\n\nMore Information needed"
] |
d2faabb4fd91224318c8a84a1e747976a27f365d |
# NijiJourney Prompt Pairs
#### A dataset containing txt2img prompt pairs for training on diffusion models
The final goal of this dataset is to create an OpenJourney like model but with NijiJourney images | Korakoe/NijiJourney-Prompt-Pairs | [
"license:creativeml-openrail-m",
"region:us"
] | 2022-12-21T06:13:45+00:00 | {"license": "creativeml-openrail-m"} | 2023-03-12T05:56:02+00:00 | [] | [] | TAGS
#license-creativeml-openrail-m #region-us
|
# NijiJourney Prompt Pairs
#### A dataset containing txt2img prompt pairs for training on diffusion models
The final goal of this dataset is to create an OpenJourney like model but with NijiJourney images | [
"# NijiJourney Prompt Pairs",
"#### A dataset containing txt2img prompt pairs for training on diffusion models\n\nThe final goal of this dataset is to create an OpenJourney like model but with NijiJourney images"
] | [
"TAGS\n#license-creativeml-openrail-m #region-us \n",
"# NijiJourney Prompt Pairs",
"#### A dataset containing txt2img prompt pairs for training on diffusion models\n\nThe final goal of this dataset is to create an OpenJourney like model but with NijiJourney images"
] |
4726c9cb565ab5fda7ba40fffc5f6bcd65bc0d81 | # Dataset Card for "enwiki-20221101-sections"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | justram/enwiki-20221101-sections | [
"region:us"
] | 2022-12-21T08:02:08+00:00 | {"dataset_info": {"features": [{"name": "text_id", "dtype": "string"}, {"name": "page_url", "dtype": "string"}, {"name": "page_title", "dtype": "string"}, {"name": "section_title", "dtype": "string"}, {"name": "context_page_description", "dtype": "string"}, {"name": "context_section_description", "dtype": "string"}, {"name": "media", "sequence": "string"}, {"name": "hierachy", "sequence": "string"}, {"name": "category", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 34190161255, "num_examples": 24220847}], "download_size": 12664592565, "dataset_size": 34190161255}} | 2022-12-21T08:32:05+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "enwiki-20221101-sections"
More Information needed | [
"# Dataset Card for \"enwiki-20221101-sections\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"enwiki-20221101-sections\"\n\nMore Information needed"
] |
d62d01b67775498eeff5dcd7ff8dc85f63e9ef5c | # Dataset of (Du et al., 2022)
## Abstract
>Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models.
## Notes
Please note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are [publicly available](https://github.com/waste-wood/e-care).
## References
Du, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849.
Roemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011. | 12ml/e-CARE | [
"task_categories:multiple-choice",
"region:us"
] | 2022-12-21T11:38:01+00:00 | {"task_categories": ["multiple-choice"]} | 2023-01-06T18:50:03+00:00 | [] | [] | TAGS
#task_categories-multiple-choice #region-us
| # Dataset of (Du et al., 2022)
## Abstract
>Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models.
## Notes
Please note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are publicly available.
## References
Du, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849.
Roemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011. | [
"# Dataset of (Du et al., 2022)",
"## Abstract\n\n>Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models.",
"## Notes\n\nPlease note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are publicly available.",
"## References\n\nDu, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849.\n\nRoemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011."
] | [
"TAGS\n#task_categories-multiple-choice #region-us \n",
"# Dataset of (Du et al., 2022)",
"## Abstract\n\n>Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models.",
"## Notes\n\nPlease note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are publicly available.",
"## References\n\nDu, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849.\n\nRoemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011."
] |
96bcc6d1de05a9cd86148af56f774923e4a4280f | # Kinyarwanda-English Commonvoice dataset
A compilation of Kinyarwanda-english dataset to be used to train multi-lingual ASR
**Note:** The audio dataset shall be added in the future | mbazaNLP/common-voice-kinyarwanda-english-dataset | [
"size_categories:~ 3000 hours",
"size_categories:721398 clips",
"language:rw",
"language:en",
"license:cc-by-4.0",
"region:us"
] | 2022-12-21T12:23:30+00:00 | {"language": ["rw", "en"], "license": ["cc-by-4.0"], "size_categories": ["~ 3000 hours", "721398 clips"]} | 2022-12-21T12:40:09+00:00 | [] | [
"rw",
"en"
] | TAGS
#size_categories-~ 3000 hours #size_categories-721398 clips #language-Kinyarwanda #language-English #license-cc-by-4.0 #region-us
| # Kinyarwanda-English Commonvoice dataset
A compilation of Kinyarwanda-english dataset to be used to train multi-lingual ASR
Note: The audio dataset shall be added in the future | [
"# Kinyarwanda-English Commonvoice dataset\nA compilation of Kinyarwanda-english dataset to be used to train multi-lingual ASR\n\nNote: The audio dataset shall be added in the future"
] | [
"TAGS\n#size_categories-~ 3000 hours #size_categories-721398 clips #language-Kinyarwanda #language-English #license-cc-by-4.0 #region-us \n",
"# Kinyarwanda-English Commonvoice dataset\nA compilation of Kinyarwanda-english dataset to be used to train multi-lingual ASR\n\nNote: The audio dataset shall be added in the future"
] |
4a3e6a934b638906125229c5b62e8ced40907e0a |
# Dataset Card for Sloleks 3
**Important**: this is a mostly complete script for processing Sloleks 3. Certain data properties may not be exposed through the script.
Please see the [CLARIN repository](https://www.clarin.si/repository/xmlui/handle/11356/1745) for full details on what the dataset contains, and open an issue or a pull request if you require some other information from the raw data.
### Dataset Summary
Sloleks is a reference morphological lexicon of Slovene that was developed to be used in various NLP applications and language manuals.
It contains Slovene lemmas, their inflected or derivative word forms and the corresponding grammatical description.
In addition to the approx. 100,000 entries already available in [Sloleks 2.0](http://hdl.handle.net/11356/1230), Sloleks 3.0 contains an additional
cca. 265,000 newly generated entries from the most frequent lemmas in [Gigafida 2.0](http://hdl.handle.net/11356/1320) not yet included in previous versions of Sloleks.
For verbs, adjectives, adverbs, and common nouns, the lemmas were checked manually by three annotators and included in Sloleks only if confirmed as legitimate by at
least one annotator. No manual checking was performed on proper nouns. Lemmatization rules, part-of-speech categorization and the set of feature-value pairs follow the
[MULTEXT-East morphosyntactic specifications for Slovenian](https://nl.ijs.si/ME/V6/msd/html/msd-sl.html).
### Supported Tasks and Leaderboards
Other (the data is a knowledge base - lexicon).
### Languages
Slovenian.
## Dataset Structure
### Data Instances
Entry for the verb `absorbirati` (English: *to absorb*):
```
{
'headword_lemma': 'absorbirati',
'pos': 'verb',
'lex_unit': {'id': 'LE_a293f9ab871299f116dff2cc1421367a', 'form': 'absorbirati', 'key': 'G_absorbirati', 'type': 'single'},
'word_forms':
[
{'forms': ['absorbirati'], 'msd': 'Ggvn'},
{'forms': ['absorbirat'], 'msd': 'Ggvm'},
{'forms': ['absorbiral'], 'msd': 'Ggvd-em'},
{'forms': ['absorbirala'], 'msd': 'Ggvd-dm'},
{'forms': ['absorbirali'], 'msd': 'Ggvd-mm'},
{'forms': ['absorbirala'], 'msd': 'Ggvd-ez'},
{'forms': ['absorbirali'], 'msd': 'Ggvd-dz'},
{'forms': ['absorbirale'], 'msd': 'Ggvd-mz'},
{'forms': ['absorbiralo'], 'msd': 'Ggvd-es'},
{'forms': ['absorbirali'], 'msd': 'Ggvd-ds'},
{'forms': ['absorbirala'], 'msd': 'Ggvd-ms'},
{'forms': ['absorbiram'], 'msd': 'Ggvspe'},
{'forms': ['absorbiraลก'], 'msd': 'Ggvsde'},
{'forms': ['absorbira'], 'msd': 'Ggvste'},
{'forms': ['absorbirava'], 'msd': 'Ggvspd'},
{'forms': ['absorbirata'], 'msd': 'Ggvsdd'},
{'forms': ['absorbirata'], 'msd': 'Ggvstd'},
{'forms': ['absorbiramo'], 'msd': 'Ggvspm'},
{'forms': ['absorbirate'], 'msd': 'Ggvsdm'},
{'forms': ['absorbirajo'], 'msd': 'Ggvstm'},
{'forms': ['absorbirajva'], 'msd': 'Ggvvpd'},
{'forms': ['absorbirajmo'], 'msd': 'Ggvvpm'},
{'forms': ['absorbiraj'], 'msd': 'Ggvvde'},
{'forms': ['absorbirajta'], 'msd': 'Ggvvdd'},
{'forms': ['absorbirajte'], 'msd': 'Ggvvdm'}
],
'is_manually_checked': True
}
```
### Data Fields
- `headword_lemma`: lemma of the headword;
- `pos`: coarse-grained part-of-speech tag (one of `{"noun", "verb", "adjective", "adverb", "pronoun", "numeral", "preposition", "conjunction", "particle", "interjection", "abbreviation", "residual"}`);
- `lex_unit`: properties of the lexical unit corresponding to the headword (`id`, `form`, `key` and `type`);
- `word_forms`: forms of the headword, each with its own list of possible forms and the morphosyntactic description of the form;
- `is_manually_checked`: whether the headword was manually validated or not.
## Additional Information
### Dataset Curators
Jaka ฤibej; et al. (please see http://hdl.handle.net/11356/1745 for the full list).
### Licensing Information
CC BY-SA 4.0.
### Citation Information
```
@misc{sloleks3,
title = {Morphological lexicon Sloleks 3.0},
author = {{\v C}ibej, Jaka and Gantar, Kaja and Dobrovoljc, Kaja and Krek, Simon and Holozan, Peter and Erjavec, Toma{\v z} and Romih, Miro and Arhar Holdt, {\v S}pela and Krsnik, Luka and Robnik-{\v S}ikonja, Marko},
url = {http://hdl.handle.net/11356/1745},
note = {Slovenian language resource repository {CLARIN}.{SI}},
copyright = {Creative Commons - Attribution-{ShareAlike} 4.0 International ({CC} {BY}-{SA} 4.0)},
year = {2022}
}
```
### Contributions
Thanks to [@matejklemen](https://github.com/matejklemen) for adding this dataset.
| cjvt/sloleks | [
"license:cc-by-sa-4.0",
"region:us"
] | 2022-12-21T13:33:13+00:00 | {"license": "cc-by-sa-4.0"} | 2024-02-11T15:11:17+00:00 | [] | [] | TAGS
#license-cc-by-sa-4.0 #region-us
|
# Dataset Card for Sloleks 3
Important: this is a mostly complete script for processing Sloleks 3. Certain data properties may not be exposed through the script.
Please see the CLARIN repository for full details on what the dataset contains, and open an issue or a pull request if you require some other information from the raw data.
### Dataset Summary
Sloleks is a reference morphological lexicon of Slovene that was developed to be used in various NLP applications and language manuals.
It contains Slovene lemmas, their inflected or derivative word forms and the corresponding grammatical description.
In addition to the approx. 100,000 entries already available in Sloleks 2.0, Sloleks 3.0 contains an additional
cca. 265,000 newly generated entries from the most frequent lemmas in Gigafida 2.0 not yet included in previous versions of Sloleks.
For verbs, adjectives, adverbs, and common nouns, the lemmas were checked manually by three annotators and included in Sloleks only if confirmed as legitimate by at
least one annotator. No manual checking was performed on proper nouns. Lemmatization rules, part-of-speech categorization and the set of feature-value pairs follow the
MULTEXT-East morphosyntactic specifications for Slovenian.
### Supported Tasks and Leaderboards
Other (the data is a knowledge base - lexicon).
### Languages
Slovenian.
## Dataset Structure
### Data Instances
Entry for the verb 'absorbirati' (English: *to absorb*):
### Data Fields
- 'headword_lemma': lemma of the headword;
- 'pos': coarse-grained part-of-speech tag (one of '{"noun", "verb", "adjective", "adverb", "pronoun", "numeral", "preposition", "conjunction", "particle", "interjection", "abbreviation", "residual"}');
- 'lex_unit': properties of the lexical unit corresponding to the headword ('id', 'form', 'key' and 'type');
- 'word_forms': forms of the headword, each with its own list of possible forms and the morphosyntactic description of the form;
- 'is_manually_checked': whether the headword was manually validated or not.
## Additional Information
### Dataset Curators
Jaka ฤibej; et al. (please see URL for the full list).
### Licensing Information
CC BY-SA 4.0.
### Contributions
Thanks to @matejklemen for adding this dataset.
| [
"# Dataset Card for Sloleks 3\n\nImportant: this is a mostly complete script for processing Sloleks 3. Certain data properties may not be exposed through the script. \nPlease see the CLARIN repository for full details on what the dataset contains, and open an issue or a pull request if you require some other information from the raw data.",
"### Dataset Summary\n\nSloleks is a reference morphological lexicon of Slovene that was developed to be used in various NLP applications and language manuals. \nIt contains Slovene lemmas, their inflected or derivative word forms and the corresponding grammatical description. \nIn addition to the approx. 100,000 entries already available in Sloleks 2.0, Sloleks 3.0 contains an additional \ncca. 265,000 newly generated entries from the most frequent lemmas in Gigafida 2.0 not yet included in previous versions of Sloleks. \nFor verbs, adjectives, adverbs, and common nouns, the lemmas were checked manually by three annotators and included in Sloleks only if confirmed as legitimate by at \nleast one annotator. No manual checking was performed on proper nouns. Lemmatization rules, part-of-speech categorization and the set of feature-value pairs follow the \nMULTEXT-East morphosyntactic specifications for Slovenian.",
"### Supported Tasks and Leaderboards\n\nOther (the data is a knowledge base - lexicon).",
"### Languages\n\nSlovenian.",
"## Dataset Structure",
"### Data Instances\n\nEntry for the verb 'absorbirati' (English: *to absorb*):",
"### Data Fields\n\n- 'headword_lemma': lemma of the headword;\n- 'pos': coarse-grained part-of-speech tag (one of '{\"noun\", \"verb\", \"adjective\", \"adverb\", \"pronoun\", \"numeral\", \"preposition\", \"conjunction\", \"particle\", \"interjection\", \"abbreviation\", \"residual\"}'); \n- 'lex_unit': properties of the lexical unit corresponding to the headword ('id', 'form', 'key' and 'type');\n- 'word_forms': forms of the headword, each with its own list of possible forms and the morphosyntactic description of the form;\n- 'is_manually_checked': whether the headword was manually validated or not.",
"## Additional Information",
"### Dataset Curators\n\nJaka ฤibej; et al. (please see URL for the full list).",
"### Licensing Information\n\nCC BY-SA 4.0.",
"### Contributions\n\nThanks to @matejklemen for adding this dataset."
] | [
"TAGS\n#license-cc-by-sa-4.0 #region-us \n",
"# Dataset Card for Sloleks 3\n\nImportant: this is a mostly complete script for processing Sloleks 3. Certain data properties may not be exposed through the script. \nPlease see the CLARIN repository for full details on what the dataset contains, and open an issue or a pull request if you require some other information from the raw data.",
"### Dataset Summary\n\nSloleks is a reference morphological lexicon of Slovene that was developed to be used in various NLP applications and language manuals. \nIt contains Slovene lemmas, their inflected or derivative word forms and the corresponding grammatical description. \nIn addition to the approx. 100,000 entries already available in Sloleks 2.0, Sloleks 3.0 contains an additional \ncca. 265,000 newly generated entries from the most frequent lemmas in Gigafida 2.0 not yet included in previous versions of Sloleks. \nFor verbs, adjectives, adverbs, and common nouns, the lemmas were checked manually by three annotators and included in Sloleks only if confirmed as legitimate by at \nleast one annotator. No manual checking was performed on proper nouns. Lemmatization rules, part-of-speech categorization and the set of feature-value pairs follow the \nMULTEXT-East morphosyntactic specifications for Slovenian.",
"### Supported Tasks and Leaderboards\n\nOther (the data is a knowledge base - lexicon).",
"### Languages\n\nSlovenian.",
"## Dataset Structure",
"### Data Instances\n\nEntry for the verb 'absorbirati' (English: *to absorb*):",
"### Data Fields\n\n- 'headword_lemma': lemma of the headword;\n- 'pos': coarse-grained part-of-speech tag (one of '{\"noun\", \"verb\", \"adjective\", \"adverb\", \"pronoun\", \"numeral\", \"preposition\", \"conjunction\", \"particle\", \"interjection\", \"abbreviation\", \"residual\"}'); \n- 'lex_unit': properties of the lexical unit corresponding to the headword ('id', 'form', 'key' and 'type');\n- 'word_forms': forms of the headword, each with its own list of possible forms and the morphosyntactic description of the form;\n- 'is_manually_checked': whether the headword was manually validated or not.",
"## Additional Information",
"### Dataset Curators\n\nJaka ฤibej; et al. (please see URL for the full list).",
"### Licensing Information\n\nCC BY-SA 4.0.",
"### Contributions\n\nThanks to @matejklemen for adding this dataset."
] |
6186f97886d1767b77ad2783b4575ad898ceeb7d | ---
annotations_creators:
- machine-generated
language:
- ru
language_creators:
- machine-generated
license:
- afl-3.0
multilinguality: []
pretty_name: Dmitriy007/restor_punct_Lenta2
size_categories:
- 100K<n<1M
source_datasets:
- original
tags: []
task_categories:
- token-classification
task_ids: []
# Dataset Card for Dmitriy007/restor_punct_Lenta2
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
ะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 (ะฒะตััะธั 2.0) ะฟัะตะดััะฐะฒะปัะตั ัะพะฑะพะน ะฝะฐะฑะพั ะธะท 800 975 ะฑะปะพะบะพะฒ ััััะบะพัะทััะฝัั
ะฟัะตะดะปะพะถะตะฝะธะน, ัะฐะทะฑะธััั
ะฝะฐ ัะปะพะฒะฐ, ะบะฐะถะดะพะต ัะปะพะฒะพ ัะฐะทะผะตัะตะฝะพ ะผะฐัะบะตัะพะผ ะดะปั ะฟะพัะปะตะดัััะตะน ะบะปะฐััะธัะธะบะฐัะธะธ ัะพะบะตะฝะพะฒ.
ะะฐะฑะพั ะดะฐะฝะฝัั
ะพัะธัะตะฝ ะพั ัะธะผะฒะพะปะพะฒ: '...', ',', 'ยซ', 'ยป', '\\', '-', '"'
ะะธะดั ะผะฐัะบะตัะพะฒ: L L. L! L? B B. B! N N. No
ะัะธะผะตัั ะทะฝะฐัะตะฝะธะน ะผะฐัะบะตัะพะฒ:
L -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ะฟัะพะฑะตะป
L. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ััะบ
B -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั
B. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั + ััะบ
N -- ะงะธัะปะพ + ะฟัะพะฑะตะป
N. -- ะงะธัะปะพ + ััะบ
No -- ะกะธะผะฒะพะป ะฝะต ะพะฟัะตะดะตะปัะฝ
### Supported Tasks and Leaderboards
token-classification: ะฝะฐะฑะพั ะดะฐะฝะฝัั
ะผะพะถะฝะพ ะธัะฟะพะปัะทะพะฒะฐัั ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ ะทะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ.
### Languages
ะขะตะบัั ะฝะฐ ััััะบะพะผ ัะทัะบะต
## Dataset Structure
### Data Instances
ะัะธะผะตั ะธะท ะฝะฐะฑะพัะฐ ะฟะพะตะทะดะพะฒ restor_punct_Lenta2 ะฒัะณะปัะดะธั ัะปะตะดัััะธะผ ะพะฑัะฐะทะพะผ:
{'words': ['ัะพัะพะณัะฐั-ะบะพััะตัะฟะพะฝะดะตะฝั', 'daily', 'mirror', 'ัะฐััะบะฐะทัะฒะฐะตั', 'ัะปััะฐะน', 'ะบะพัะพััะน', 'ะฟะพัะฐะดัะตั', 'ะฒัะตั
', 'ะดััะทะตะน', 'ะถะธะฒะพัะฝัั
'], 'labels': ['B', 'B', 'B', 'L', 'L', 'L', 'L', 'L', 'L', 'L.'], 'labels_id': [4, 4, 4, 0, 0, 0, 0, 0, 0, 1]}
### Data Fields
โข 'words': ัะฟะธัะพะบ ัะปะพะฒ, ัะพะดะตัะถะฐัะฐั ัะตะบัั ัะฐะทะฑะธััะน ะฝะฐ ะพัะดะตะปัะฝัะต ัะปะพะฒะฐ.
โข 'labels': ัััะพะบะฐ, ัะฟะธัะพะบ ะผะฐัะบะตัะพะฒ
โข 'labels_id: ัะตะปะพะต ัะธัะปะพ, ะพั 0 ะดะพ 9 , ะพะฑะพะทะฝะฐัะฐััะตะต ะฟะพััะดะบะพะฒัะน ะฝะพะผะตั ะผะฐัะบะตัะฐ
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
ะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ัะฐะทัะฐะฑะพัะฐะฝ ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ
ะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ ะฒ ัะตะบััะต ะฟัะตะดะปะพะถะตะฝะธั. ะัะตะดะฟะพะปะฐะณะฐะปะฐัั, ััะพ ะพะฑััะตะฝะฝะฐั ัะฐะบะธะผ ะพะฑัะฐะทะพะผ ะผะพะดะตะปั, ะฑัะดะตั ะธัะฟะพะปัะทะพะฒะฐัััั ะฒ ะทะฐะดะฐัะธ ััะฐะฝัะบัะธะฑะฐัะธะธ.
### Source Data
#### Initial Data Collection and Normalization
ะะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ะพัะฝะพะฒะฐะฝ ะฝะฐ ะฝะฐะฑะพัะต ะดะฐะฝะฝัั
Lenta2 ะฟัะพะตะบัะฐ CORUS.
#### Who are the source language producers?
[More Information Needed]
### Annotations
ะะฐะฑะพั ะดะฐะฝะฝัั
ะฝะต ัะพะดะตัะถะธั ะฝะธะบะฐะบะธั
ะดะพะฟะพะปะฝะธัะตะปัะฝัั
ะฐะฝะฝะพัะฐัะธะน.
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
ะะผะตะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปะตะน ะธะปะธ ะปะธัะฝะฐั ะธะฝัะพัะผะฐัะธั ัะตัะตะฝะทะตะฝัะพะฒ ะฝะต ัะพะฑะธัะฐะปะธัั ะฒะผะตััะต ั ะพะฑะทะพัะฐะผะธ, ะฝะพ ะฟะพัะตะฝัะธะฐะปัะฝะพ ะผะพะณัั ะฑััั ะฒะพัััะฐะฝะพะฒะปะตะฝั.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
| Dmitriy007/restor_punct_Lenta2 | [
"region:us"
] | 2022-12-21T14:07:45+00:00 | {} | 2023-01-19T13:02:56+00:00 | [] | [] | TAGS
#region-us
| ---
annotations_creators:
- machine-generated
language:
- ru
language_creators:
- machine-generated
license:
- afl-3.0
multilinguality: []
pretty_name: Dmitriy007/restor_punct_Lenta2
size_categories:
- 100K<n<1M
source_datasets:
- original
tags: []
task_categories:
- token-classification
task_ids: []
# Dataset Card for Dmitriy007/restor_punct_Lenta2
## Table of Contents
- Table of Contents
- Dataset Description
- Dataset Summary
- Supported Tasks and Leaderboards
- Languages
- Dataset Structure
- Data Instances
- Data Fields
- Data Splits
- Dataset Creation
- Curation Rationale
- Source Data
- Annotations
- Personal and Sensitive Information
- Considerations for Using the Data
- Social Impact of Dataset
- Discussion of Biases
- Other Known Limitations
- Additional Information
- Dataset Curators
- Licensing Information
- Citation Information
- Contributions
## Dataset Description
- Homepage:
- Repository:
- Paper:
- Leaderboard:
- Point of Contact:
### Dataset Summary
ะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 (ะฒะตััะธั 2.0) ะฟัะตะดััะฐะฒะปัะตั ัะพะฑะพะน ะฝะฐะฑะพั ะธะท 800 975 ะฑะปะพะบะพะฒ ััััะบะพัะทััะฝัั
ะฟัะตะดะปะพะถะตะฝะธะน, ัะฐะทะฑะธััั
ะฝะฐ ัะปะพะฒะฐ, ะบะฐะถะดะพะต ัะปะพะฒะพ ัะฐะทะผะตัะตะฝะพ ะผะฐัะบะตัะพะผ ะดะปั ะฟะพัะปะตะดัััะตะน ะบะปะฐััะธัะธะบะฐัะธะธ ัะพะบะตะฝะพะฒ.
ะะฐะฑะพั ะดะฐะฝะฝัั
ะพัะธัะตะฝ ะพั ัะธะผะฒะพะปะพะฒ: '...', ',', 'ยซ', 'ยป', '\\', '-', '"'
ะะธะดั ะผะฐัะบะตัะพะฒ: L L. L! L? B B. B! N N. No
ะัะธะผะตัั ะทะฝะฐัะตะฝะธะน ะผะฐัะบะตัะพะฒ:
L -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ะฟัะพะฑะตะป
L. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ััะบ
B -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั
B. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั + ััะบ
N -- ะงะธัะปะพ + ะฟัะพะฑะตะป
N. -- ะงะธัะปะพ + ััะบ
No -- ะกะธะผะฒะพะป ะฝะต ะพะฟัะตะดะตะปัะฝ
### Supported Tasks and Leaderboards
token-classification: ะฝะฐะฑะพั ะดะฐะฝะฝัั
ะผะพะถะฝะพ ะธัะฟะพะปัะทะพะฒะฐัั ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ ะทะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ.
### Languages
ะขะตะบัั ะฝะฐ ััััะบะพะผ ัะทัะบะต
## Dataset Structure
### Data Instances
ะัะธะผะตั ะธะท ะฝะฐะฑะพัะฐ ะฟะพะตะทะดะพะฒ restor_punct_Lenta2 ะฒัะณะปัะดะธั ัะปะตะดัััะธะผ ะพะฑัะฐะทะพะผ:
{'words': ['ัะพัะพะณัะฐั-ะบะพััะตัะฟะพะฝะดะตะฝั', 'daily', 'mirror', 'ัะฐััะบะฐะทัะฒะฐะตั', 'ัะปััะฐะน', 'ะบะพัะพััะน', 'ะฟะพัะฐะดัะตั', 'ะฒัะตั
', 'ะดััะทะตะน', 'ะถะธะฒะพัะฝัั
'], 'labels': ['B', 'B', 'B', 'L', 'L', 'L', 'L', 'L', 'L', 'L.'], 'labels_id': [4, 4, 4, 0, 0, 0, 0, 0, 0, 1]}
### Data Fields
โข 'words': ัะฟะธัะพะบ ัะปะพะฒ, ัะพะดะตัะถะฐัะฐั ัะตะบัั ัะฐะทะฑะธััะน ะฝะฐ ะพัะดะตะปัะฝัะต ัะปะพะฒะฐ.
โข 'labels': ัััะพะบะฐ, ัะฟะธัะพะบ ะผะฐัะบะตัะพะฒ
โข 'labels_id: ัะตะปะพะต ัะธัะปะพ, ะพั 0 ะดะพ 9 , ะพะฑะพะทะฝะฐัะฐััะตะต ะฟะพััะดะบะพะฒัะน ะฝะพะผะตั ะผะฐัะบะตัะฐ
### Data Splits
## Dataset Creation
### Curation Rationale
ะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ัะฐะทัะฐะฑะพัะฐะฝ ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ
ะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ ะฒ ัะตะบััะต ะฟัะตะดะปะพะถะตะฝะธั. ะัะตะดะฟะพะปะฐะณะฐะปะฐัั, ััะพ ะพะฑััะตะฝะฝะฐั ัะฐะบะธะผ ะพะฑัะฐะทะพะผ ะผะพะดะตะปั, ะฑัะดะตั ะธัะฟะพะปัะทะพะฒะฐัััั ะฒ ะทะฐะดะฐัะธ ััะฐะฝัะบัะธะฑะฐัะธะธ.
### Source Data
#### Initial Data Collection and Normalization
ะะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ะพัะฝะพะฒะฐะฝ ะฝะฐ ะฝะฐะฑะพัะต ะดะฐะฝะฝัั
Lenta2 ะฟัะพะตะบัะฐ CORUS.
#### Who are the source language producers?
### Annotations
ะะฐะฑะพั ะดะฐะฝะฝัั
ะฝะต ัะพะดะตัะถะธั ะฝะธะบะฐะบะธั
ะดะพะฟะพะปะฝะธัะตะปัะฝัั
ะฐะฝะฝะพัะฐัะธะน.
#### Annotation process
#### Who are the annotators?
### Personal and Sensitive Information
ะะผะตะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปะตะน ะธะปะธ ะปะธัะฝะฐั ะธะฝัะพัะผะฐัะธั ัะตัะตะฝะทะตะฝัะพะฒ ะฝะต ัะพะฑะธัะฐะปะธัั ะฒะผะตััะต ั ะพะฑะทะพัะฐะผะธ, ะฝะพ ะฟะพัะตะฝัะธะฐะปัะฝะพ ะผะพะณัั ะฑััั ะฒะพัััะฐะฝะพะฒะปะตะฝั.
## Considerations for Using the Data
### Social Impact of Dataset
### Discussion of Biases
### Other Known Limitations
## Additional Information
### Dataset Curators
### Licensing Information
### Contributions
Thanks to @github-username for adding this dataset.
| [
"# Dataset Card for Dmitriy007/restor_punct_Lenta2",
"## Table of Contents\n- Table of Contents\n- Dataset Description\n- Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions",
"## Dataset Description\n\n- Homepage:\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:",
"### Dataset Summary\n\nะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 (ะฒะตััะธั 2.0) ะฟัะตะดััะฐะฒะปัะตั ัะพะฑะพะน ะฝะฐะฑะพั ะธะท 800 975 ะฑะปะพะบะพะฒ ััััะบะพัะทััะฝัั
ะฟัะตะดะปะพะถะตะฝะธะน, ัะฐะทะฑะธััั
ะฝะฐ ัะปะพะฒะฐ, ะบะฐะถะดะพะต ัะปะพะฒะพ ัะฐะทะผะตัะตะฝะพ ะผะฐัะบะตัะพะผ ะดะปั ะฟะพัะปะตะดัััะตะน ะบะปะฐััะธัะธะบะฐัะธะธ ัะพะบะตะฝะพะฒ.\n\nะะฐะฑะพั ะดะฐะฝะฝัั
ะพัะธัะตะฝ ะพั ัะธะผะฒะพะปะพะฒ: '...', ',', 'ยซ', 'ยป', '\\\\', '-', '\"' \n\nะะธะดั ะผะฐัะบะตัะพะฒ: L L. L! L? B B. B! N N. No\n\nะัะธะผะตัั ะทะฝะฐัะตะฝะธะน ะผะฐัะบะตัะพะฒ:\n\nL -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ะฟัะพะฑะตะป\n\nL. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ััะบ\n\nB -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั\n\nB. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั + ััะบ\n\nN -- ะงะธัะปะพ + ะฟัะพะฑะตะป\n\nN. -- ะงะธัะปะพ + ััะบ\n\nNo -- ะกะธะผะฒะพะป ะฝะต ะพะฟัะตะดะตะปัะฝ",
"### Supported Tasks and Leaderboards\n\ntoken-classification: ะฝะฐะฑะพั ะดะฐะฝะฝัั
ะผะพะถะฝะพ ะธัะฟะพะปัะทะพะฒะฐัั ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ ะทะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ.",
"### Languages\n\nะขะตะบัั ะฝะฐ ััััะบะพะผ ัะทัะบะต",
"## Dataset Structure",
"### Data Instances\n\nะัะธะผะตั ะธะท ะฝะฐะฑะพัะฐ ะฟะพะตะทะดะพะฒ restor_punct_Lenta2 ะฒัะณะปัะดะธั ัะปะตะดัััะธะผ ะพะฑัะฐะทะพะผ:\n{'words': ['ัะพัะพะณัะฐั-ะบะพััะตัะฟะพะฝะดะตะฝั', 'daily', 'mirror', 'ัะฐััะบะฐะทัะฒะฐะตั', 'ัะปััะฐะน', 'ะบะพัะพััะน', 'ะฟะพัะฐะดัะตั', 'ะฒัะตั
', 'ะดััะทะตะน', 'ะถะธะฒะพัะฝัั
'], 'labels': ['B', 'B', 'B', 'L', 'L', 'L', 'L', 'L', 'L', 'L.'], 'labels_id': [4, 4, 4, 0, 0, 0, 0, 0, 0, 1]}",
"### Data Fields\n\nโข\t'words': ัะฟะธัะพะบ ัะปะพะฒ, ัะพะดะตัะถะฐัะฐั ัะตะบัั ัะฐะทะฑะธััะน ะฝะฐ ะพัะดะตะปัะฝัะต ัะปะพะฒะฐ.\nโข\t'labels': ัััะพะบะฐ, ัะฟะธัะพะบ ะผะฐัะบะตัะพะฒ\nโข\t'labels_id: ัะตะปะพะต ัะธัะปะพ, ะพั 0 ะดะพ 9 , ะพะฑะพะทะฝะฐัะฐััะตะต ะฟะพััะดะบะพะฒัะน ะฝะพะผะตั ะผะฐัะบะตัะฐ",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale\n\nะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ัะฐะทัะฐะฑะพัะฐะฝ ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ \nะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ ะฒ ัะตะบััะต ะฟัะตะดะปะพะถะตะฝะธั. ะัะตะดะฟะพะปะฐะณะฐะปะฐัั, ััะพ ะพะฑััะตะฝะฝะฐั ัะฐะบะธะผ ะพะฑัะฐะทะพะผ ะผะพะดะตะปั, ะฑัะดะตั ะธัะฟะพะปัะทะพะฒะฐัััั ะฒ ะทะฐะดะฐัะธ ััะฐะฝัะบัะธะฑะฐัะธะธ.",
"### Source Data",
"#### Initial Data Collection and Normalization\n\nะะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ะพัะฝะพะฒะฐะฝ ะฝะฐ ะฝะฐะฑะพัะต ะดะฐะฝะฝัั
Lenta2 ะฟัะพะตะบัะฐ CORUS.",
"#### Who are the source language producers?",
"### Annotations\n\nะะฐะฑะพั ะดะฐะฝะฝัั
ะฝะต ัะพะดะตัะถะธั ะฝะธะบะฐะบะธั
ะดะพะฟะพะปะฝะธัะตะปัะฝัั
ะฐะฝะฝะพัะฐัะธะน.",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information\n\nะะผะตะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปะตะน ะธะปะธ ะปะธัะฝะฐั ะธะฝัะพัะผะฐัะธั ัะตัะตะฝะทะตะฝัะพะฒ ะฝะต ัะพะฑะธัะฐะปะธัั ะฒะผะตััะต ั ะพะฑะทะพัะฐะผะธ, ะฝะพ ะฟะพัะตะฝัะธะฐะปัะฝะพ ะผะพะณัั ะฑััั ะฒะพัััะฐะฝะพะฒะปะตะฝั.",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions\n\nThanks to @github-username for adding this dataset."
] | [
"TAGS\n#region-us \n",
"# Dataset Card for Dmitriy007/restor_punct_Lenta2",
"## Table of Contents\n- Table of Contents\n- Dataset Description\n- Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions",
"## Dataset Description\n\n- Homepage:\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:",
"### Dataset Summary\n\nะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 (ะฒะตััะธั 2.0) ะฟัะตะดััะฐะฒะปัะตั ัะพะฑะพะน ะฝะฐะฑะพั ะธะท 800 975 ะฑะปะพะบะพะฒ ััััะบะพัะทััะฝัั
ะฟัะตะดะปะพะถะตะฝะธะน, ัะฐะทะฑะธััั
ะฝะฐ ัะปะพะฒะฐ, ะบะฐะถะดะพะต ัะปะพะฒะพ ัะฐะทะผะตัะตะฝะพ ะผะฐัะบะตัะพะผ ะดะปั ะฟะพัะปะตะดัััะตะน ะบะปะฐััะธัะธะบะฐัะธะธ ัะพะบะตะฝะพะฒ.\n\nะะฐะฑะพั ะดะฐะฝะฝัั
ะพัะธัะตะฝ ะพั ัะธะผะฒะพะปะพะฒ: '...', ',', 'ยซ', 'ยป', '\\\\', '-', '\"' \n\nะะธะดั ะผะฐัะบะตัะพะฒ: L L. L! L? B B. B! N N. No\n\nะัะธะผะตัั ะทะฝะฐัะตะฝะธะน ะผะฐัะบะตัะพะฒ:\n\nL -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ะฟัะพะฑะตะป\n\nL. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะผะฐะปะตะฝัะบะพะน ะฑัะบะฒั + ััะบ\n\nB -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั\n\nB. -- ะดะฐะฝะฝะพะต ัะปะพะฒะพ ั ะทะฐะณะปะฐะฒะฝะพะน ะฑัะบะฒั + ััะบ\n\nN -- ะงะธัะปะพ + ะฟัะพะฑะตะป\n\nN. -- ะงะธัะปะพ + ััะบ\n\nNo -- ะกะธะผะฒะพะป ะฝะต ะพะฟัะตะดะตะปัะฝ",
"### Supported Tasks and Leaderboards\n\ntoken-classification: ะฝะฐะฑะพั ะดะฐะฝะฝัั
ะผะพะถะฝะพ ะธัะฟะพะปัะทะพะฒะฐัั ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ ะทะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ.",
"### Languages\n\nะขะตะบัั ะฝะฐ ััััะบะพะผ ัะทัะบะต",
"## Dataset Structure",
"### Data Instances\n\nะัะธะผะตั ะธะท ะฝะฐะฑะพัะฐ ะฟะพะตะทะดะพะฒ restor_punct_Lenta2 ะฒัะณะปัะดะธั ัะปะตะดัััะธะผ ะพะฑัะฐะทะพะผ:\n{'words': ['ัะพัะพะณัะฐั-ะบะพััะตัะฟะพะฝะดะตะฝั', 'daily', 'mirror', 'ัะฐััะบะฐะทัะฒะฐะตั', 'ัะปััะฐะน', 'ะบะพัะพััะน', 'ะฟะพัะฐะดัะตั', 'ะฒัะตั
', 'ะดััะทะตะน', 'ะถะธะฒะพัะฝัั
'], 'labels': ['B', 'B', 'B', 'L', 'L', 'L', 'L', 'L', 'L', 'L.'], 'labels_id': [4, 4, 4, 0, 0, 0, 0, 0, 0, 1]}",
"### Data Fields\n\nโข\t'words': ัะฟะธัะพะบ ัะปะพะฒ, ัะพะดะตัะถะฐัะฐั ัะตะบัั ัะฐะทะฑะธััะน ะฝะฐ ะพัะดะตะปัะฝัะต ัะปะพะฒะฐ.\nโข\t'labels': ัััะพะบะฐ, ัะฟะธัะพะบ ะผะฐัะบะตัะพะฒ\nโข\t'labels_id: ัะตะปะพะต ัะธัะปะพ, ะพั 0 ะดะพ 9 , ะพะฑะพะทะฝะฐัะฐััะตะต ะฟะพััะดะบะพะฒัะน ะฝะพะผะตั ะผะฐัะบะตัะฐ",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale\n\nะะฐะฑะพั ะดะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ัะฐะทัะฐะฑะพัะฐะฝ ะดะปั ะพะฑััะตะฝะธั ะผะพะดะตะปะธ ะฒะพัััะฐะฝะพะฒะปะตะฝะธั ะฟัะฝะบััะฐัะธะธ ะธ \nะฐะณะปะฐะฒะฝัั
ะฑัะบะฒ ะฒ ัะตะบััะต ะฟัะตะดะปะพะถะตะฝะธั. ะัะตะดะฟะพะปะฐะณะฐะปะฐัั, ััะพ ะพะฑััะตะฝะฝะฐั ัะฐะบะธะผ ะพะฑัะฐะทะพะผ ะผะพะดะตะปั, ะฑัะดะตั ะธัะฟะพะปัะทะพะฒะฐัััั ะฒ ะทะฐะดะฐัะธ ััะฐะฝัะบัะธะฑะฐัะธะธ.",
"### Source Data",
"#### Initial Data Collection and Normalization\n\nะะฐะฝะฝัั
restor_punct_Lenta2 ะฑัะป ะพัะฝะพะฒะฐะฝ ะฝะฐ ะฝะฐะฑะพัะต ะดะฐะฝะฝัั
Lenta2 ะฟัะพะตะบัะฐ CORUS.",
"#### Who are the source language producers?",
"### Annotations\n\nะะฐะฑะพั ะดะฐะฝะฝัั
ะฝะต ัะพะดะตัะถะธั ะฝะธะบะฐะบะธั
ะดะพะฟะพะปะฝะธัะตะปัะฝัั
ะฐะฝะฝะพัะฐัะธะน.",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information\n\nะะผะตะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปะตะน ะธะปะธ ะปะธัะฝะฐั ะธะฝัะพัะผะฐัะธั ัะตัะตะฝะทะตะฝัะพะฒ ะฝะต ัะพะฑะธัะฐะปะธัั ะฒะผะตััะต ั ะพะฑะทะพัะฐะผะธ, ะฝะพ ะฟะพัะตะฝัะธะฐะปัะฝะพ ะผะพะณัั ะฑััั ะฒะพัััะฐะฝะพะฒะปะตะฝั.",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions\n\nThanks to @github-username for adding this dataset."
] |
d63755480740dbf590917e4a90d81bf2596eb676 |
A RL environment called BallChase for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_BallChase
```
| edbeeching/godot_rl_BallChase | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:29:19+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:46:50+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called BallChase for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
7631ff2d592aa0d63d8c817846f321cc79a77f29 |
A RL environment called FlyBy for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_FlyBy
```
| edbeeching/godot_rl_FlyBy | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:29:50+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:47:07+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called FlyBy for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
69f13fe15b356a53128ff4628e5d31821cef1225 |
A RL environment called FPS for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_FPS
```
| edbeeching/godot_rl_FPS | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:30:31+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:47:23+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called FPS for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
689f721fd87c5c1152e028cbc2e053c52f162a1e |
A RL environment called JumperHard for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_JumperHard
```
| edbeeching/godot_rl_JumperHard | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:31:50+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:48:07+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called JumperHard for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
2b780bfdd101a53e52109c1d27a0e38cb137fbc6 |
A RL environment called Racer for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_Racer
```
| edbeeching/godot_rl_Racer | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:32:20+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:48:27+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called Racer for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
0a35afa9268df9dae11b8fe5850817cb7cb5db66 |
A RL environment called VirtualCamera for the Godot Game Engine.
This environment was created with: https://github.com/edbeeching/godot_rl_agents
## Downloading the environment
After installing Godot RL Agents, download the environment with:
```
gdrl.env_from_hub -r edbeeching/godot_rl_VirtualCamera
```
| edbeeching/godot_rl_VirtualCamera | [
"deep-reinforcement-learning",
"reinforcement-learning",
"godot-rl",
"environments",
"video-games",
"region:us"
] | 2022-12-21T14:33:30+00:00 | {"library_name": "godot-rl", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "godot-rl", "environments", "video-games"]} | 2024-01-07T09:48:56+00:00 | [] | [] | TAGS
#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us
|
A RL environment called VirtualCamera for the Godot Game Engine.
This environment was created with: URL
## Downloading the environment
After installing Godot RL Agents, download the environment with:
| [
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] | [
"TAGS\n#deep-reinforcement-learning #reinforcement-learning #godot-rl #environments #video-games #region-us \n",
"## Downloading the environment \n\nAfter installing Godot RL Agents, download the environment with:"
] |
f35c91ff04c01d598e930dd7a6bd0b857ca3a53f | # Dataset Card for "lego-blip-captions-512"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Norod78/lego-blip-captions-512 | [
"region:us"
] | 2022-12-21T14:42:12+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 627030265.0, "num_examples": 2511}], "download_size": 625119749, "dataset_size": 627030265.0}} | 2022-12-21T14:43:40+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "lego-blip-captions-512"
More Information needed | [
"# Dataset Card for \"lego-blip-captions-512\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"lego-blip-captions-512\"\n\nMore Information needed"
] |
6360e9d7bf1b5673955ef4296e338e6ce68fbe2a | # Dataset Card for "sinograms"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | AshrafAlAodat/sinograms | [
"region:us"
] | 2022-12-21T14:55:54+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 49082809.0, "num_examples": 1400}], "download_size": 48978515, "dataset_size": 49082809.0}} | 2022-12-25T16:06:34+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "sinograms"
More Information needed | [
"# Dataset Card for \"sinograms\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"sinograms\"\n\nMore Information needed"
] |
265e9b8e307d3a10a00409320e9afbcfb10aa2c6 |
# Dataset Card for NPSC Bokmรฅl (< 15 sec. segments)
## Dataset Description
- **Homepage:**
- **Repository:** <https://github.com/scribe-project/nodalida_2023_combined_training>
- **Paper:**
```
@inproceedings{
solberg2023improving,
title={Improving Generalization of Norwegian {ASR} with Limited Linguistic Resources},
author={Per Erik Solberg and Pablo Ortiz and Phoebe Parsons and Torbj{\o}rn Svendsen and Giampiero Salvi},
booktitle={The 24rd Nordic Conference on Computational Linguistics},
year={2023}
}
```
- **Point of Contact:** [Per Erik Solberg](mailto:[email protected])
### Dataset Summary
This is the version of the Bokmรฅl part of the Norwegian Parliamentary Speech Corpus (NPSC) used for training and testing the STORTINGET model
in the paper *Improving Generalization of Norwegian ASR with Limited Linguistic Resources* presented at NoDaLiDa 2023.
It only contains segments of a length < 15 sec. For a full version of the NPSC, see [this repository](https://huggingface.co/datasets/NbAiLab/NPSC).
### Languages
Norwegian Bokmรฅl
## Dataset Creation
### Source Data
The full version of this dataset is found in [the repository of the Norwegian Language Bank](https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-58/)
#### Initial Data Collection and Normalization
The data was retrieved using the [Spraakbanken downloader](https://pypi.org/project/spraakbanken-downloader/) and standardized
using the [combined dataset standardization scripts](https://github.com/scribe-project/asr-standardized-combined). Bokmรฅl segments with a duration < 15 seconds were
extracted using [this code](https://github.com/scribe-project/nodalida_2023_combined_training/blob/main/make_datasets/make_npsc_csvs.ipynb).
## Licensing Information
[CC0](https://creativecommons.org/share-your-work/public-domain/cc0/)
### Citation Information
```
@inproceedings{
solberg2023improving,
title={Improving Generalization of Norwegian {ASR} with Limited Linguistic Resources},
author={Per Erik Solberg and Pablo Ortiz and Phoebe Parsons and Torbj{\o}rn Svendsen and Giampiero Salvi},
booktitle={The 24rd Nordic Conference on Computational Linguistics},
year={2023}
}
``` | scribe-project/npsc_nb | [
"region:us"
] | 2022-12-21T15:21:00+00:00 | {"dataset_info": {"features": [{"name": "speaker_id", "dtype": "string"}, {"name": "gender", "dtype": "string"}, {"name": "utterance_id", "dtype": "string"}, {"name": "language", "dtype": "string"}, {"name": "raw_text", "dtype": "string"}, {"name": "full_audio_file", "dtype": "string"}, {"name": "original_data_split", "dtype": "string"}, {"name": "region", "dtype": "string"}, {"name": "duration", "dtype": "float64"}, {"name": "start", "dtype": "float64"}, {"name": "end", "dtype": "float64"}, {"name": "utterance_audio_file", "dtype": "audio"}, {"name": "standardized_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 8190809957.84, "num_examples": 40008}, {"name": "test", "num_bytes": 1026553338.856, "num_examples": 5044}, {"name": "validation", "num_bytes": 1097030649.769, "num_examples": 5461}], "download_size": 10261847599, "dataset_size": 10314393946.465}} | 2023-04-25T09:23:19+00:00 | [] | [] | TAGS
#region-us
|
# Dataset Card for NPSC Bokmรฅl (< 15 sec. segments)
## Dataset Description
- Homepage:
- Repository: <URL
- Paper:
- Point of Contact: Per Erik Solberg
### Dataset Summary
This is the version of the Bokmรฅl part of the Norwegian Parliamentary Speech Corpus (NPSC) used for training and testing the STORTINGET model
in the paper *Improving Generalization of Norwegian ASR with Limited Linguistic Resources* presented at NoDaLiDa 2023.
It only contains segments of a length < 15 sec. For a full version of the NPSC, see this repository.
### Languages
Norwegian Bokmรฅl
## Dataset Creation
### Source Data
The full version of this dataset is found in the repository of the Norwegian Language Bank
#### Initial Data Collection and Normalization
The data was retrieved using the Spraakbanken downloader and standardized
using the combined dataset standardization scripts. Bokmรฅl segments with a duration < 15 seconds were
extracted using this code.
## Licensing Information
CC0
| [
"# Dataset Card for NPSC Bokmรฅl (< 15 sec. segments)",
"## Dataset Description\n\n- Homepage: \n- Repository: <URL\n- Paper:\n\n\n- Point of Contact: Per Erik Solberg",
"### Dataset Summary\n\nThis is the version of the Bokmรฅl part of the Norwegian Parliamentary Speech Corpus (NPSC) used for training and testing the STORTINGET model\n in the paper *Improving Generalization of Norwegian ASR with Limited Linguistic Resources* presented at NoDaLiDa 2023.\nIt only contains segments of a length < 15 sec. For a full version of the NPSC, see this repository.",
"### Languages\n\nNorwegian Bokmรฅl",
"## Dataset Creation",
"### Source Data\nThe full version of this dataset is found in the repository of the Norwegian Language Bank",
"#### Initial Data Collection and Normalization\n\nThe data was retrieved using the Spraakbanken downloader and standardized\n using the combined dataset standardization scripts. Bokmรฅl segments with a duration < 15 seconds were\n extracted using this code.",
"## Licensing Information\n\nCC0"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for NPSC Bokmรฅl (< 15 sec. segments)",
"## Dataset Description\n\n- Homepage: \n- Repository: <URL\n- Paper:\n\n\n- Point of Contact: Per Erik Solberg",
"### Dataset Summary\n\nThis is the version of the Bokmรฅl part of the Norwegian Parliamentary Speech Corpus (NPSC) used for training and testing the STORTINGET model\n in the paper *Improving Generalization of Norwegian ASR with Limited Linguistic Resources* presented at NoDaLiDa 2023.\nIt only contains segments of a length < 15 sec. For a full version of the NPSC, see this repository.",
"### Languages\n\nNorwegian Bokmรฅl",
"## Dataset Creation",
"### Source Data\nThe full version of this dataset is found in the repository of the Norwegian Language Bank",
"#### Initial Data Collection and Normalization\n\nThe data was retrieved using the Spraakbanken downloader and standardized\n using the combined dataset standardization scripts. Bokmรฅl segments with a duration < 15 seconds were\n extracted using this code.",
"## Licensing Information\n\nCC0"
] |
8751f30ff7a107a39477e0b6a074eebcc20d0c08 | # Dataset Card for "test_tags"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Test
Test123
data | mariosasko/test_tags | [
"region:us"
] | 2022-12-21T17:47:30+00:00 | {"dataset_info": {"features": [{"name": "a", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 32, "num_examples": 4}], "download_size": 584, "dataset_size": 32}} | 2023-05-03T15:05:57+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "test_tags"
More Information needed
Test
Test123
data | [
"# Dataset Card for \"test_tags\"\n\nMore Information needed\n\nTest\n\nTest123\n\ndata"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"test_tags\"\n\nMore Information needed\n\nTest\n\nTest123\n\ndata"
] |
719c91226a79f5f9a8984145f15f29626eabc29a |
# CORAA-v1.1
CORAA is a publicly available dataset for Automatic Speech Recognition (ASR) in the Brazilian Portuguese language containing 290.77 hours of audios and their respective transcriptions (400k+ segmented audios). The dataset is composed of audios of 5 original projects:
- ALIP (Gonรงalves, 2019)
- C-ORAL Brazil (Raso and Mello, 2012)
- NURC-Recife (Oliviera Jr., 2016)
- SP-2010 (Mendes and Oushiro, 2012)
- TEDx talks (talks in Portuguese)
The audios were either validated by annotators or transcripted for the first time aiming at the ASR task.
## Metadata
- file_path: the path to an audio file
- task: transcription (annotators revised original transcriptions); annotation (annotators classified the audio-transcription pair according to votes_for_* metrics); annotation_and_transcription (both tasks were performed)
- variety: European Portuguese (PT_PT) or Brazilian Portuguese (PT_BR)
- dataset: one of five datasets (ALIP, C-oral Brasil, NURC-RE, SP2010, TEDx Portuguese)
- accent: one of four accents (Minas Gerais, Recife, Sao Paulo cities, Sao Paulo capital) or the value "miscellaneous"
- speech_genre: Interviews, Dialogues, Monologues, Conversations, Interviews, Conference, Class Talks, Stage Talks or Reading
- speech_style: Spontaneous Speech or Prepared Speech or Read Speech
- up_votes: for annotation, the number of votes to valid the audio (most audios were revewed by one annotor, but some of the audios were analyzed by more than one).
- down_votes: for annotation, the number of votes do invalid the audio (always smaller than up_votes)
- votes_for_hesitation: for annotation, votes categorizing the audio as having the hesitation phenomenon
- votes_for_filled_pause: for annotation, votes categorizing the audio as having the filled pause phenomenon
- votes_for_noise_or_low_voice: for annotation, votes categorizing the audio as either having noise or low voice, without impairing the audio compression.
- votes_for_second_voice: for annotation, votes categorizing the audio as having a second voice, without impairing the audio compression
- votes_for_no_identified_problem: without impairing the audio as having no identified phenomenon (of the four described above)
- text: the transcription for the audio
## Downloads :
Dataset:
| Gdrive | Internal | Hugging Face |
|-----------|--------------------|-----------|
| [Train audios](https://drive.google.com/file/d/1deCciFD35EA_OEUl0MrEDa7u5O2KgVJM/view?usp=sharing) | [Train audios](http://143.107.183.175:14888/static/coraa/train.zip)| [Train audios](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/tree/main/train_dividido) | |
| [Train transcriptions and metadata](https://drive.google.com/file/d/1HbwahfMWoArYj0z2PfI4dHiambWfaNWg/view?usp=sharing) | [Train transcriptions and metadata](http://143.107.183.175:14880/metadata_train_final.csv)| [Train transcriptions and metadata](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/blob/main/metadata_train_final.csv)|
|[Dev audios](https://drive.google.com/file/d/1D1ft4F37zLjmGxQyhfkdjSs9cJzOL3nI/view?usp=sharing) |[Dev audios](http://143.107.183.175:14880/dev.zip) |[Dev audios](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/blob/main/dev.zip) |
| [Dev transcriptions and metadata](https://drive.google.com/file/d/185erjax7lS_YNuolZvcMt_EdprafyMU0/view?usp=sharing) | [Dev transcriptions and metadata](http://143.107.183.175:14880/metadata_dev_final.csv) | [Dev transcriptions and metadata](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/blob/main/metadata_dev_final.csv) |
| [Test audios](https://drive.google.com/file/d/1vHH5oVo4zeJKchIyHHHjzvKD3QXuJxHZ/view?usp=sharing) | [Test audios](http://143.107.183.175:14880/test.zip) | [Test audios](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/blob/main/test.zip) |
| [Test transcriptions and metadata](https://drive.google.com/file/d/1hcNoA7-xOEn5s0iYjX6BebaEsx_7LfCd/view?usp=sharing) | [Test transcriptions and metadata](http://143.107.183.175:14880/metadata_test_final.csv) | [Test transcriptions and metadata](https://huggingface.co/datasets/gabrielrstan/CORAA-v1.1/blob/main/metadata_test_final.csv) |
Experiments:
- [Checkpoints ](https://drive.google.com/drive/folders/10JkbCzYypZtCz1nHY5rBoBM1r66P3p3j?usp=sharing)
- [Code](https://github.com/Edresson/Wav2Vec-Wrapper)
Model trained in this corpus: Wav2Vec 2.0 XLSR-53 (multilingual pretraining)
## Citation
- [Preprint](https://arxiv.org/abs/2110.15731):
```
@misc{c2021coraa,
title={CORAA: a large corpus of spontaneous and prepared speech manually validated for speech recognition in Brazilian Portuguese},
author={Arnaldo Candido Junior and Edresson Casanova and Anderson Soares and Frederico Santos de Oliveira and Lucas Oliveira and Ricardo Corso Fernandes Junior and Daniel Peixoto Pinto da Silva and Fernando Gorgulho Fayet and Bruno Baldissera Carlotto and Lucas Rafael Stefanel Gris and Sandra Maria Aluรญsio},
year={2021},
eprint={2110.15731},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
- Full Paper: coming soon
- Oficial site: [Tarsila Project](https://sites.google.com/view/tarsila-c4ai/)
## Partners / Sponsors / Funding
- [C4AI](https://c4ai.inova.usp.br/pt/home-2/)
- [CEIA](https://centrodeia.org/)
- [UFG](https://www.ufg.br/)
- [USP](https://www5.usp.br/)
- [UTFPR](http://www.utfpr.edu.br/)
## References
- Gonรงalves SCL (2019) Projeto ALIP (amostra linguรญstica do interior paulista) e banco de dados iboruna: 10 anos de contribuiรงรฃo com a descriรงรฃo do Portuguรชs Brasileiro. Revista Estudos Linguรญsticos 48(1):276โ297.
- Raso T, Mello H, Mittmann MM (2012) The C-ORAL-BRASIL I: Reference corpus for spoken Brazilian Portuguese. In: Proceedings of the Eighth International Conference on Language Resources and Evaluation (LRECโ12), European Language Resources Association (ELRA), Istanbul, Turkey, pp 106โ113, URL http://www.lrec-conf.org/proceedings/lrec2012/pdf/624_Paper.pdf
- Oliviera Jr M (2016) Nurc digital um protocolo para a digitalizaรงรฃo, anotaรงรฃo, arquivamento e disseminaรงรฃo do material do projeto da norma urbana linguรญstica culta (NURC). CHIMERA: Revista de Corpus de Lenguas Romances y Estudios Linguรญsticos 3(2):149โ174, URL https://revistas.uam.es/chimera/article/view/6519
- Mendes RB, Oushiro L (2012) Mapping Paulistano Portuguese: the SP2010 Project. In: Proceedings of the VIIth GSCP International Conference: Speech and Corpora, Fizenze University Press, Firenze, Italy, pp 459โ463.
| gabrielrstan/CORAA-v1.1 | [
"license:unknown",
"arxiv:2110.15731",
"region:us"
] | 2022-12-21T19:37:05+00:00 | {"license": "unknown"} | 2022-12-28T23:15:17+00:00 | [
"2110.15731"
] | [] | TAGS
#license-unknown #arxiv-2110.15731 #region-us
| CORAA-v1.1
==========
CORAA is a publicly available dataset for Automatic Speech Recognition (ASR) in the Brazilian Portuguese language containing 290.77 hours of audios and their respective transcriptions (400k+ segmented audios). The dataset is composed of audios of 5 original projects:
* ALIP (Gonรงalves, 2019)
* C-ORAL Brazil (Raso and Mello, 2012)
* NURC-Recife (Oliviera Jr., 2016)
* SP-2010 (Mendes and Oushiro, 2012)
* TEDx talks (talks in Portuguese)
The audios were either validated by annotators or transcripted for the first time aiming at the ASR task.
Metadata
--------
* file\_path: the path to an audio file
* task: transcription (annotators revised original transcriptions); annotation (annotators classified the audio-transcription pair according to votes\_for\_\* metrics); annotation\_and\_transcription (both tasks were performed)
* variety: European Portuguese (PT\_PT) or Brazilian Portuguese (PT\_BR)
* dataset: one of five datasets (ALIP, C-oral Brasil, NURC-RE, SP2010, TEDx Portuguese)
* accent: one of four accents (Minas Gerais, Recife, Sao Paulo cities, Sao Paulo capital) or the value "miscellaneous"
* speech\_genre: Interviews, Dialogues, Monologues, Conversations, Interviews, Conference, Class Talks, Stage Talks or Reading
* speech\_style: Spontaneous Speech or Prepared Speech or Read Speech
* up\_votes: for annotation, the number of votes to valid the audio (most audios were revewed by one annotor, but some of the audios were analyzed by more than one).
* down\_votes: for annotation, the number of votes do invalid the audio (always smaller than up\_votes)
* votes\_for\_hesitation: for annotation, votes categorizing the audio as having the hesitation phenomenon
* votes\_for\_filled\_pause: for annotation, votes categorizing the audio as having the filled pause phenomenon
* votes\_for\_noise\_or\_low\_voice: for annotation, votes categorizing the audio as either having noise or low voice, without impairing the audio compression.
* votes\_for\_second\_voice: for annotation, votes categorizing the audio as having a second voice, without impairing the audio compression
* votes\_for\_no\_identified\_problem: without impairing the audio as having no identified phenomenon (of the four described above)
* text: the transcription for the audio
Downloads :
-----------
Dataset:
Gdrive: Train audios, Internal: Train audios, Hugging Face: Train audios
Gdrive: Train transcriptions and metadata, Internal: Train transcriptions and metadata, Hugging Face: Train transcriptions and metadata
Gdrive: Dev audios, Internal: Dev audios, Hugging Face: Dev audios
Gdrive: Dev transcriptions and metadata, Internal: Dev transcriptions and metadata, Hugging Face: Dev transcriptions and metadata
Gdrive: Test audios, Internal: Test audios, Hugging Face: Test audios
Gdrive: Test transcriptions and metadata, Internal: Test transcriptions and metadata, Hugging Face: Test transcriptions and metadata
Experiments:
* Checkpoints
* Code
Model trained in this corpus: Wav2Vec 2.0 XLSR-53 (multilingual pretraining)
* Preprint:
* Full Paper: coming soon
* Oficial site: Tarsila Project
Partners / Sponsors / Funding
-----------------------------
* C4AI
* CEIA
* UFG
* USP
* UTFPR
References
----------
* Gonรงalves SCL (2019) Projeto ALIP (amostra linguรญstica do interior paulista) e banco de dados iboruna: 10 anos de contribuiรงรฃo com a descriรงรฃo do Portuguรชs Brasileiro. Revista Estudos Linguรญsticos 48(1):276โ297.
* Raso T, Mello H, Mittmann MM (2012) The C-ORAL-BRASIL I: Reference corpus for spoken Brazilian Portuguese. In: Proceedings of the Eighth International Conference on Language Resources and Evaluation (LRECโ12), European Language Resources Association (ELRA), Istanbul, Turkey, pp 106โ113, URL URL
* Oliviera Jr M (2016) Nurc digital um protocolo para a digitalizaรงรฃo, anotaรงรฃo, arquivamento e disseminaรงรฃo do material do projeto da norma urbana linguรญstica culta (NURC). CHIMERA: Revista de Corpus de Lenguas Romances y Estudios Linguรญsticos 3(2):149โ174, URL URL
* Mendes RB, Oushiro L (2012) Mapping Paulistano Portuguese: the SP2010 Project. In: Proceedings of the VIIth GSCP International Conference: Speech and Corpora, Fizenze University Press, Firenze, Italy, pp 459โ463.
| [] | [
"TAGS\n#license-unknown #arxiv-2110.15731 #region-us \n"
] |
71b6d04a64f75e4df0a03a58157426ce6d77a817 | # Dataset Card for "unnatural-instructions"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mrm8488/unnatural-instructions | [
"region:us"
] | 2022-12-21T20:56:20+00:00 | {"dataset_info": [{"config_name": "default", "features": [{"name": "instruction", "dtype": "string"}, {"name": "instances", "list": [{"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "constraints", "dtype": "string"}, {"name": "output", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 54668900, "num_examples": 66010}], "download_size": 28584196, "dataset_size": 54668900}, {"config_name": "core", "features": [{"name": "instruction", "dtype": "string"}, {"name": "instances", "sequence": [{"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "output", "dtype": "string"}, {"name": "constraints", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 55461020, "num_examples": 66010}], "download_size": 29679516, "dataset_size": 55461020}, {"config_name": "full", "features": [{"name": "instruction", "dtype": "string"}, {"name": "instances", "sequence": [{"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "output", "dtype": "string"}, {"name": "constraints", "dtype": "string"}]}, {"name": "reformulations", "sequence": [{"name": "instruction", "dtype": "string"}, {"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "output", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 145864853, "num_examples": 66010}], "download_size": 29679516, "dataset_size": 145864853}]} | 2022-12-23T18:09:15+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "unnatural-instructions"
More Information needed | [
"# Dataset Card for \"unnatural-instructions\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"unnatural-instructions\"\n\nMore Information needed"
] |
7791769bbc6675c142b5f81cc8e2de0ce44de7ab | # Dataset Card for Unnatural Instructions (Core data)
This info comes from the **Unnatural Instructions GitHub [repo](https://github.com/orhonovich/unnatural-instructions/)**.
Unnatural Instructions is a dataset of instructions automatically generated by a Large Language model.
See full details in the paper: "[Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor](https://arxiv.org/abs/2212.09689)"
## ๐๏ธ Content
The Unnatural Instructions core dataset of 68,478 instruction-input-output triplets.
## ๐ Format
### Core data
Each example contains:
- `input`: An input for the task described by the `instruction`
- `instruction_with_input`: The instruction concatenated with the `input`
- `constraints`: The task's output space constraints
- `output`: The output of executing `instruction` with the given `input`
## ๐ Citation
If you make use of Unnatural Instructions, please cite the following paper:
```
@misc{honovich2022unnatural,
title = {Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor},
author = {Honovich, Or and Scialom, Thomas and Levy, Omer and Schick, Timo},
url = {https://arxiv.org/abs/2212.09689},
publisher = {arXiv},
year={2022}
}
```
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mrm8488/unnatural-instructions-core | [
"arxiv:2212.09689",
"region:us"
] | 2022-12-21T20:57:50+00:00 | {"dataset_info": {"features": [{"name": "instruction", "dtype": "string"}, {"name": "instances", "list": [{"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "constraints", "dtype": "string"}, {"name": "output", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 54668900, "num_examples": 66010}], "download_size": 28584196, "dataset_size": 54668900}} | 2022-12-21T21:42:06+00:00 | [
"2212.09689"
] | [] | TAGS
#arxiv-2212.09689 #region-us
| # Dataset Card for Unnatural Instructions (Core data)
This info comes from the Unnatural Instructions GitHub repo.
Unnatural Instructions is a dataset of instructions automatically generated by a Large Language model.
See full details in the paper: "Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor"
## ๏ธ Content
The Unnatural Instructions core dataset of 68,478 instruction-input-output triplets.
## Format
### Core data
Each example contains:
- 'input': An input for the task described by the 'instruction'
- 'instruction_with_input': The instruction concatenated with the 'input'
- 'constraints': The task's output space constraints
- 'output': The output of executing 'instruction' with the given 'input'
## Citation
If you make use of Unnatural Instructions, please cite the following paper:
More Information needed | [
"# Dataset Card for Unnatural Instructions (Core data)\nThis info comes from the Unnatural Instructions GitHub repo.\n\nUnnatural Instructions is a dataset of instructions automatically generated by a Large Language model.\nSee full details in the paper: \"Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor\"",
"## ๏ธ Content\nThe Unnatural Instructions core dataset of 68,478 instruction-input-output triplets.",
"## Format",
"### Core data\nEach example contains:\n- 'input': An input for the task described by the 'instruction'\n- 'instruction_with_input': The instruction concatenated with the 'input'\n- 'constraints': The task's output space constraints\n- 'output': The output of executing 'instruction' with the given 'input'",
"## Citation\nIf you make use of Unnatural Instructions, please cite the following paper:\n\n\nMore Information needed"
] | [
"TAGS\n#arxiv-2212.09689 #region-us \n",
"# Dataset Card for Unnatural Instructions (Core data)\nThis info comes from the Unnatural Instructions GitHub repo.\n\nUnnatural Instructions is a dataset of instructions automatically generated by a Large Language model.\nSee full details in the paper: \"Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor\"",
"## ๏ธ Content\nThe Unnatural Instructions core dataset of 68,478 instruction-input-output triplets.",
"## Format",
"### Core data\nEach example contains:\n- 'input': An input for the task described by the 'instruction'\n- 'instruction_with_input': The instruction concatenated with the 'input'\n- 'constraints': The task's output space constraints\n- 'output': The output of executing 'instruction' with the given 'input'",
"## Citation\nIf you make use of Unnatural Instructions, please cite the following paper:\n\n\nMore Information needed"
] |
e55025405febbc31033ac51b86eb1c36667c979f | # Dataset Card for Unnatural Instructions (Full data)
This info comes from the **Unnatural Instructions GitHub [repo](https://github.com/orhonovich/unnatural-instructions/)**.
Unnatural Instructions is a dataset of instructions automatically generated by a Large Language model.
See full details in the paper: "[Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor](https://arxiv.org/abs/2212.09689)"
## ๐๏ธ Content
It contains the full 240,670 Unnatural Instructions (instruction-input-output triplets) examples. It was constructed by expanding the core data with automatically generated instruction paraphrases.
## ๐ Format
### Full data
It has the same structure as [Core Data](https://huggingface.co/datasets/mrm8488/unnatural-instructions-core), but with one additional field - `reformulations`. `reformulations` is an array of JSON objects, each corresponds to an automatically generated paraphrase for the given instruction. Each reformulation contains the fields:
- `instruction`: A paraphrase of the original instruction
- `input`: An input for the task described by the `instruction`
- `instruction_with_input`: The paraphrased instruction concatenated with the `input`
- `output`: The output of executing `instruction` with the given `input`
## ๐ Citation
If you make use of Unnatural Instructions, please cite the following paper:
```
@misc{honovich2022unnatural,
title = {Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor},
author = {Honovich, Or and Scialom, Thomas and Levy, Omer and Schick, Timo},
url = {https://arxiv.org/abs/2212.09689},
publisher = {arXiv},
year={2022}
}
```
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mrm8488/unnatural-instructions-full | [
"arxiv:2212.09689",
"region:us"
] | 2022-12-21T20:59:04+00:00 | {"dataset_info": {"features": [{"name": "instruction", "dtype": "string"}, {"name": "instances", "list": [{"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "constraints", "dtype": "string"}, {"name": "output", "dtype": "string"}]}, {"name": "reformulations", "list": [{"name": "instruction", "dtype": "string"}, {"name": "instruction_with_input", "dtype": "string"}, {"name": "input", "dtype": "string"}, {"name": "output", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 144282712, "num_examples": 66010}], "download_size": 57715606, "dataset_size": 144282712}} | 2022-12-21T21:41:31+00:00 | [
"2212.09689"
] | [] | TAGS
#arxiv-2212.09689 #region-us
| # Dataset Card for Unnatural Instructions (Full data)
This info comes from the Unnatural Instructions GitHub repo.
Unnatural Instructions is a dataset of instructions automatically generated by a Large Language model.
See full details in the paper: "Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor"
## ๏ธ Content
It contains the full 240,670 Unnatural Instructions (instruction-input-output triplets) examples. It was constructed by expanding the core data with automatically generated instruction paraphrases.
## Format
### Full data
It has the same structure as Core Data, but with one additional field - 'reformulations'. 'reformulations' is an array of JSON objects, each corresponds to an automatically generated paraphrase for the given instruction. Each reformulation contains the fields:
- 'instruction': A paraphrase of the original instruction
- 'input': An input for the task described by the 'instruction'
- 'instruction_with_input': The paraphrased instruction concatenated with the 'input'
- 'output': The output of executing 'instruction' with the given 'input'
## Citation
If you make use of Unnatural Instructions, please cite the following paper:
More Information needed | [
"# Dataset Card for Unnatural Instructions (Full data)\nThis info comes from the Unnatural Instructions GitHub repo.\n\nUnnatural Instructions is a dataset of instructions automatically generated by a Large Language model.\nSee full details in the paper: \"Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor\"",
"## ๏ธ Content\nIt contains the full 240,670 Unnatural Instructions (instruction-input-output triplets) examples. It was constructed by expanding the core data with automatically generated instruction paraphrases.",
"## Format",
"### Full data\nIt has the same structure as Core Data, but with one additional field - 'reformulations'. 'reformulations' is an array of JSON objects, each corresponds to an automatically generated paraphrase for the given instruction. Each reformulation contains the fields:\n- 'instruction': A paraphrase of the original instruction\n- 'input': An input for the task described by the 'instruction'\n- 'instruction_with_input': The paraphrased instruction concatenated with the 'input'\n- 'output': The output of executing 'instruction' with the given 'input'",
"## Citation\nIf you make use of Unnatural Instructions, please cite the following paper:\n\n\nMore Information needed"
] | [
"TAGS\n#arxiv-2212.09689 #region-us \n",
"# Dataset Card for Unnatural Instructions (Full data)\nThis info comes from the Unnatural Instructions GitHub repo.\n\nUnnatural Instructions is a dataset of instructions automatically generated by a Large Language model.\nSee full details in the paper: \"Unnatural Instructions: Tuning Language Models with (Almost) No Human Labor\"",
"## ๏ธ Content\nIt contains the full 240,670 Unnatural Instructions (instruction-input-output triplets) examples. It was constructed by expanding the core data with automatically generated instruction paraphrases.",
"## Format",
"### Full data\nIt has the same structure as Core Data, but with one additional field - 'reformulations'. 'reformulations' is an array of JSON objects, each corresponds to an automatically generated paraphrase for the given instruction. Each reformulation contains the fields:\n- 'instruction': A paraphrase of the original instruction\n- 'input': An input for the task described by the 'instruction'\n- 'instruction_with_input': The paraphrased instruction concatenated with the 'input'\n- 'output': The output of executing 'instruction' with the given 'input'",
"## Citation\nIf you make use of Unnatural Instructions, please cite the following paper:\n\n\nMore Information needed"
] |
5379ba66d6b869d8ba74c8341f3100d645246d5d | # Dataset Card for "ljspeech_phonemes"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | bookbot/ljspeech_phonemes | [
"region:us"
] | 2022-12-21T23:18:09+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 22050}}}, {"name": "file", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "normalized_text", "dtype": "string"}, {"name": "phonemes", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3863152206.0, "num_examples": 13100}], "download_size": 3787337731, "dataset_size": 3863152206.0}} | 2022-12-21T23:24:29+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "ljspeech_phonemes"
More Information needed | [
"# Dataset Card for \"ljspeech_phonemes\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"ljspeech_phonemes\"\n\nMore Information needed"
] |
071c5c771d4c540a53ca4197921315ec5b400ae5 | # Dataset Card for "preprocessed_jsut_jsss_css10_fleurs_common_voice_11"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | vumichien/preprocessed_jsut_jsss_css10_fleurs_common_voice_11 | [
"region:us"
] | 2022-12-22T00:00:05+00:00 | {"dataset_info": {"features": [{"name": "audio", "struct": [{"name": "array", "sequence": "float32"}, {"name": "path", "dtype": "string"}, {"name": "sampling_rate", "dtype": "int64"}]}, {"name": "sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 12359542831, "num_examples": 31708}, {"name": "test", "num_bytes": 1562198132, "num_examples": 4604}], "download_size": 13916026126, "dataset_size": 13921740963}} | 2022-12-22T00:07:28+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "preprocessed_jsut_jsss_css10_fleurs_common_voice_11"
More Information needed | [
"# Dataset Card for \"preprocessed_jsut_jsss_css10_fleurs_common_voice_11\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"preprocessed_jsut_jsss_css10_fleurs_common_voice_11\"\n\nMore Information needed"
] |
39882674ccfc800c95e149aa00ce5e69f13fedff | # Dataset Card for "wikipedia-ptbr-20221220"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | dominguesm/wikipedia-ptbr-20221220 | [
"region:us"
] | 2022-12-22T00:07:45+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2367117753.3, "num_examples": 987399}, {"name": "test", "num_bytes": 131507740.51323204, "num_examples": 54856}, {"name": "valid", "num_bytes": 131505343.18676797, "num_examples": 54855}], "download_size": 1592202665, "dataset_size": 2630130837.0000005}} | 2022-12-22T10:49:09+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "wikipedia-ptbr-20221220"
More Information needed | [
"# Dataset Card for \"wikipedia-ptbr-20221220\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"wikipedia-ptbr-20221220\"\n\nMore Information needed"
] |
098765c79ea10a2cb19c828324e33281b8336ec0 | # Dataset Card for PopQA
## Dataset Summary
PopQA is a large-scale open-domain question answering (QA) dataset, consisting of 14k entity-centric QA pairs. Each question is created by converting a knowledge tuple retrieved from Wikidata using a template. Each question come with the original `subject_entitiey`, `object_entity`and `relationship_type` annotation, as well as Wikipedia monthly page views.
## Languages
The dataset contains samples in English only.
## Dataset Structure
### Data Instances
- Size of downloaded dataset file: 5.2 MB
## Data Fields
- `id`: question id
- `subj`: subject entity name
- `prop`: relationship type
- `obj`: object entity name
- `subj_id`: Wikidata ID of the subject entity
- `prop_id`: Wikidata relationship type ID
- `obj_id`: Wikidata ID of the object entity
- `s_aliases`: aliases of the subject entity
- `o_aliases`: aliases of the object entity
- `s_uri`: Wikidata URI of the subject entity
- `o_uri`: Wikidata URI of the object entity
- `s_wiki_title`: Wikipedia page title of the subject entity
- `o_wiki_title`: Wikipedia page title of the object entity
- `s_pop`: Wikipedia monthly pageview of the subject entity
- `o_pop`: Wikipedia monthly pageview of the object entity
- `question`: PopQA question
- `possible_answers`: a list of the gold answers.
## Citation Information
```
@article{ mallen2023llm_memorization ,
title={When Not to Trust Language Models: Investigating Effectiveness and Limitations of Parametric and Non-Parametric Memories },
author={ Mallen, Alex and Asai,Akari and Zhong, Victor and Das, Rajarshi and Hajishirzi, Hannaneh and Khashabi, Daniel},
journal={ arXiv preprint },
year={ 2022 }
}
```
| akariasai/PopQA | [
"region:us"
] | 2022-12-22T00:37:19+00:00 | {} | 2022-12-22T01:01:20+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for PopQA
## Dataset Summary
PopQA is a large-scale open-domain question answering (QA) dataset, consisting of 14k entity-centric QA pairs. Each question is created by converting a knowledge tuple retrieved from Wikidata using a template. Each question come with the original 'subject_entitiey', 'object_entity'and 'relationship_type' annotation, as well as Wikipedia monthly page views.
## Languages
The dataset contains samples in English only.
## Dataset Structure
### Data Instances
- Size of downloaded dataset file: 5.2 MB
## Data Fields
- 'id': question id
- 'subj': subject entity name
- 'prop': relationship type
- 'obj': object entity name
- 'subj_id': Wikidata ID of the subject entity
- 'prop_id': Wikidata relationship type ID
- 'obj_id': Wikidata ID of the object entity
- 's_aliases': aliases of the subject entity
- 'o_aliases': aliases of the object entity
- 's_uri': Wikidata URI of the subject entity
- 'o_uri': Wikidata URI of the object entity
- 's_wiki_title': Wikipedia page title of the subject entity
- 'o_wiki_title': Wikipedia page title of the object entity
- 's_pop': Wikipedia monthly pageview of the subject entity
- 'o_pop': Wikipedia monthly pageview of the object entity
- 'question': PopQA question
- 'possible_answers': a list of the gold answers.
| [
"# Dataset Card for PopQA",
"## Dataset Summary \nPopQA is a large-scale open-domain question answering (QA) dataset, consisting of 14k entity-centric QA pairs. Each question is created by converting a knowledge tuple retrieved from Wikidata using a template. Each question come with the original 'subject_entitiey', 'object_entity'and 'relationship_type' annotation, as well as Wikipedia monthly page views.",
"## Languages\nThe dataset contains samples in English only.",
"## Dataset Structure",
"### Data Instances\n- Size of downloaded dataset file: 5.2 MB",
"## Data Fields\n- 'id': question id\n- 'subj': subject entity name\n- 'prop': relationship type\n- 'obj': object entity name\n- 'subj_id': Wikidata ID of the subject entity\n- 'prop_id': Wikidata relationship type ID\n- 'obj_id': Wikidata ID of the object entity\n- 's_aliases': aliases of the subject entity\n- 'o_aliases': aliases of the object entity\n- 's_uri': Wikidata URI of the subject entity \n- 'o_uri': Wikidata URI of the object entity\n- 's_wiki_title': Wikipedia page title of the subject entity\n- 'o_wiki_title': Wikipedia page title of the object entity\n- 's_pop': Wikipedia monthly pageview of the subject entity\n- 'o_pop': Wikipedia monthly pageview of the object entity\n- 'question': PopQA question\n- 'possible_answers': a list of the gold answers."
] | [
"TAGS\n#region-us \n",
"# Dataset Card for PopQA",
"## Dataset Summary \nPopQA is a large-scale open-domain question answering (QA) dataset, consisting of 14k entity-centric QA pairs. Each question is created by converting a knowledge tuple retrieved from Wikidata using a template. Each question come with the original 'subject_entitiey', 'object_entity'and 'relationship_type' annotation, as well as Wikipedia monthly page views.",
"## Languages\nThe dataset contains samples in English only.",
"## Dataset Structure",
"### Data Instances\n- Size of downloaded dataset file: 5.2 MB",
"## Data Fields\n- 'id': question id\n- 'subj': subject entity name\n- 'prop': relationship type\n- 'obj': object entity name\n- 'subj_id': Wikidata ID of the subject entity\n- 'prop_id': Wikidata relationship type ID\n- 'obj_id': Wikidata ID of the object entity\n- 's_aliases': aliases of the subject entity\n- 'o_aliases': aliases of the object entity\n- 's_uri': Wikidata URI of the subject entity \n- 'o_uri': Wikidata URI of the object entity\n- 's_wiki_title': Wikipedia page title of the subject entity\n- 'o_wiki_title': Wikipedia page title of the object entity\n- 's_pop': Wikipedia monthly pageview of the subject entity\n- 'o_pop': Wikipedia monthly pageview of the object entity\n- 'question': PopQA question\n- 'possible_answers': a list of the gold answers."
] |
24ebc073e65279f0d2a8cd785ec5f9283a1ab7fa | # Dataset Card for "dataset-v-1.4_CLIP_identities_random_seeds"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | SDbiaseval/dataset-v-1.4_CLIP_identities_random_seeds | [
"region:us"
] | 2022-12-22T00:44:47+00:00 | {"dataset_info": {"features": [{"name": "adjective", "dtype": "string"}, {"name": "profession", "dtype": "string"}, {"name": "no", "dtype": "int32"}, {"name": "image_path", "dtype": "string"}, {"name": "image", "dtype": "image"}, {"name": "gender", "dtype": "string"}, {"name": "identity", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1172792739.5, "num_examples": 31500}], "download_size": 1167658244, "dataset_size": 1172792739.5}} | 2022-12-22T00:46:00+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "dataset-v-1.4_CLIP_identities_random_seeds"
More Information needed | [
"# Dataset Card for \"dataset-v-1.4_CLIP_identities_random_seeds\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"dataset-v-1.4_CLIP_identities_random_seeds\"\n\nMore Information needed"
] |
519a5e8ba1c5657282a90bb81a7bf4a10f200742 | # Dataset Card for "rick-and-morty-s06e01-blip-captions"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | juliaturc/rick-and-morty-s06e01-blip-captions | [
"region:us"
] | 2022-12-22T01:05:00+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 78803729.742, "num_examples": 1341}], "download_size": 78105717, "dataset_size": 78803729.742}} | 2022-12-22T01:05:09+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "rick-and-morty-s06e01-blip-captions"
More Information needed | [
"# Dataset Card for \"rick-and-morty-s06e01-blip-captions\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"rick-and-morty-s06e01-blip-captions\"\n\nMore Information needed"
] |
48df7abf0f64f9279b4ee04386272eb9dc89ef89 |
## Dataset Description
- **Repository:** https://github.com/shuyanzhou/docprompting
- **Paper:** [DocPrompting: Generating Code by Retrieving the Docs](https://arxiv.org/pdf/2207.05987.pdf)
### Dataset Summary
This is the re-split of [CoNaLa](https://conala-corpus.github.io/) dataset.
For each code snippet in the dev and test set, at least one function is held out from the training set.
This split aims at testing a code generation model's capacity in generating *unseen* functions
We further make sure that examples from the same StackOverflow post (same `question_id` before `-`) are in the same split.
### Supported Tasks and Leaderboards
This dataset is used to evaluate code generations.
### Languages
English - Python code.
## Dataset Structure
```python
dataset = load_dataset("neulab/docpromting-conala")
DatasetDict({
train: Dataset({
features: ['nl', 'cmd', 'question_id', 'cmd_name', 'oracle_man', 'canonical_cmd'],
num_rows: 2135
})
test: Dataset({
features: ['nl', 'cmd', 'question_id', 'cmd_name', 'oracle_man', 'canonical_cmd'],
num_rows: 543
})
validation: Dataset({
features: ['nl', 'cmd', 'question_id', 'cmd_name', 'oracle_man', 'canonical_cmd'],
num_rows: 201
})
})
})
code_docs = load_dataset("neulab/docprompting-conala", "docs")
DatasetDict({
train: Dataset({
features: ['doc_id', 'doc_content'],
num_rows: 34003
})
})
```
### Data Fields
train/dev/test:
- nl: The natural language intent
- cmd: The reference code snippet
- question_id: `x-y`where `x` is the StackOverflow post ID
- oracle_man: The `doc_id` of the functions used in the reference code snippet. The corresponding contents are in `doc` split
- canonical_cmd: The canonical version reference code snippet
docs:
- doc_id: the id of a doc
- doc_content: the content of the doc
## Dataset Creation
The dataset was crawled from Stack Overflow, automatically filtered, then curated by annotators. For more details, please refer to the original [paper](https://arxiv.org/pdf/1805.08949.pdf)
### Citation Information
```
@article{zhou2022doccoder,
title={DocCoder: Generating Code by Retrieving and Reading Docs},
author={Zhou, Shuyan and Alon, Uri and Xu, Frank F and JIang, Zhengbao and Neubig, Graham},
journal={arXiv preprint arXiv:2207.05987},
year={2022}
}
``` | neulab/docprompting-conala | [
"task_categories:text2text-generation",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:unknown",
"source_datasets:original",
"language:code",
"license:mit",
"code-generation",
"doc retrieval",
"retrieval augmented generation",
"arxiv:2207.05987",
"arxiv:1805.08949",
"region:us"
] | 2022-12-22T02:40:47+00:00 | {"annotations_creators": [], "language_creators": ["crowdsourced", "expert-generated"], "language": ["code"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["unknown"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "pretty_name": "DocPrompting-CoNaLa", "tags": ["code-generation", "doc retrieval", "retrieval augmented generation"]} | 2023-03-14T17:59:47+00:00 | [
"2207.05987",
"1805.08949"
] | [
"code"
] | TAGS
#task_categories-text2text-generation #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-unknown #source_datasets-original #language-code #license-mit #code-generation #doc retrieval #retrieval augmented generation #arxiv-2207.05987 #arxiv-1805.08949 #region-us
|
## Dataset Description
- Repository: URL
- Paper: DocPrompting: Generating Code by Retrieving the Docs
### Dataset Summary
This is the re-split of CoNaLa dataset.
For each code snippet in the dev and test set, at least one function is held out from the training set.
This split aims at testing a code generation model's capacity in generating *unseen* functions
We further make sure that examples from the same StackOverflow post (same 'question_id' before '-') are in the same split.
### Supported Tasks and Leaderboards
This dataset is used to evaluate code generations.
### Languages
English - Python code.
## Dataset Structure
### Data Fields
train/dev/test:
- nl: The natural language intent
- cmd: The reference code snippet
- question_id: 'x-y'where 'x' is the StackOverflow post ID
- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split
- canonical_cmd: The canonical version reference code snippet
docs:
- doc_id: the id of a doc
- doc_content: the content of the doc
## Dataset Creation
The dataset was crawled from Stack Overflow, automatically filtered, then curated by annotators. For more details, please refer to the original paper
| [
"## Dataset Description\n- Repository: URL\n- Paper: DocPrompting: Generating Code by Retrieving the Docs",
"### Dataset Summary\nThis is the re-split of CoNaLa dataset. \nFor each code snippet in the dev and test set, at least one function is held out from the training set. \nThis split aims at testing a code generation model's capacity in generating *unseen* functions\nWe further make sure that examples from the same StackOverflow post (same 'question_id' before '-') are in the same split.",
"### Supported Tasks and Leaderboards\nThis dataset is used to evaluate code generations.",
"### Languages\nEnglish - Python code.",
"## Dataset Structure",
"### Data Fields\ntrain/dev/test:\n- nl: The natural language intent\n- cmd: The reference code snippet\n- question_id: 'x-y'where 'x' is the StackOverflow post ID\n- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split\n- canonical_cmd: The canonical version reference code snippet\n \n\ndocs:\n- doc_id: the id of a doc\n- doc_content: the content of the doc",
"## Dataset Creation\nThe dataset was crawled from Stack Overflow, automatically filtered, then curated by annotators. For more details, please refer to the original paper"
] | [
"TAGS\n#task_categories-text2text-generation #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-unknown #source_datasets-original #language-code #license-mit #code-generation #doc retrieval #retrieval augmented generation #arxiv-2207.05987 #arxiv-1805.08949 #region-us \n",
"## Dataset Description\n- Repository: URL\n- Paper: DocPrompting: Generating Code by Retrieving the Docs",
"### Dataset Summary\nThis is the re-split of CoNaLa dataset. \nFor each code snippet in the dev and test set, at least one function is held out from the training set. \nThis split aims at testing a code generation model's capacity in generating *unseen* functions\nWe further make sure that examples from the same StackOverflow post (same 'question_id' before '-') are in the same split.",
"### Supported Tasks and Leaderboards\nThis dataset is used to evaluate code generations.",
"### Languages\nEnglish - Python code.",
"## Dataset Structure",
"### Data Fields\ntrain/dev/test:\n- nl: The natural language intent\n- cmd: The reference code snippet\n- question_id: 'x-y'where 'x' is the StackOverflow post ID\n- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split\n- canonical_cmd: The canonical version reference code snippet\n \n\ndocs:\n- doc_id: the id of a doc\n- doc_content: the content of the doc",
"## Dataset Creation\nThe dataset was crawled from Stack Overflow, automatically filtered, then curated by annotators. For more details, please refer to the original paper"
] |
1c75706ec89ed05fd07382a50cfed9b40847a21f | Sampled Data from AIforBharat corpora | aashay96/indic-gpt | [
"region:us"
] | 2022-12-22T06:55:12+00:00 | {} | 2023-04-21T19:45:09+00:00 | [] | [] | TAGS
#region-us
| Sampled Data from AIforBharat corpora | [] | [
"TAGS\n#region-us \n"
] |
2a397393d74975c43e0b64ff466fa839d1347eb8 | # Cleaned russian traffic sign images dataset
Dataset is generated from [Russian traffic sign images dataset](https://www.kaggle.com/datasets/watchman/rtsd-dataset) and [detected signs in the dataset](https://graphics.cs.msu.ru/projects/traffic-sign-recognition.html). | eleldar/rtsd_cleaned | [
"region:us"
] | 2022-12-22T07:09:31+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "sign_class", "dtype": "string"}, {"name": "sign_id", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": -515611439.904, "num_examples": 104358}], "download_size": 58343345, "dataset_size": -515611439.904}} | 2022-12-22T13:09:31+00:00 | [] | [] | TAGS
#region-us
| # Cleaned russian traffic sign images dataset
Dataset is generated from Russian traffic sign images dataset and detected signs in the dataset. | [
"# Cleaned russian traffic sign images dataset\n\nDataset is generated from Russian traffic sign images dataset and detected signs in the dataset."
] | [
"TAGS\n#region-us \n",
"# Cleaned russian traffic sign images dataset\n\nDataset is generated from Russian traffic sign images dataset and detected signs in the dataset."
] |
3028680b3026b981a4bf7bae6e6ba222077c1b90 | # Dataset Card for "diachronia-ocr-dev"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Zombely/diachronia-ocr-dev | [
"region:us"
] | 2022-12-22T09:47:47+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 22012156.0, "num_examples": 22}], "download_size": 22013969, "dataset_size": 22012156.0}} | 2022-12-22T09:48:12+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "diachronia-ocr-dev"
More Information needed | [
"# Dataset Card for \"diachronia-ocr-dev\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"diachronia-ocr-dev\"\n\nMore Information needed"
] |
46d69af917fa77d952c487f8e7f88c27d4cb848f | # Dataset Card for "fiszki-ocr-dev"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Zombely/fiszki-ocr-dev | [
"region:us"
] | 2022-12-22T12:40:51+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 430755239.0, "num_examples": 102}], "download_size": 430685891, "dataset_size": 430755239.0}} | 2022-12-22T12:43:44+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "fiszki-ocr-dev"
More Information needed | [
"# Dataset Card for \"fiszki-ocr-dev\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"fiszki-ocr-dev\"\n\nMore Information needed"
] |
a68ec0be8993c24c10c71fc75c02e38526b80aa0 | # Dataset Card for "fiszki-ocr"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Zombely/fiszki-ocr-test-A | [
"region:us"
] | 2022-12-22T12:43:49+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 379639540.0, "num_examples": 91}], "download_size": 379576204, "dataset_size": 379639540.0}} | 2022-12-22T12:46:07+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "fiszki-ocr"
More Information needed | [
"# Dataset Card for \"fiszki-ocr\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"fiszki-ocr\"\n\nMore Information needed"
] |
55142e51668207b21f68e40a99e9aed0655e3a3f |
# Dataset Card for Horse-30
## Dataset Description
- **Homepage:** horse10.deeplabcut.org
- **Repository:** https://github.com/DeepLabCut/DeepLabCut
- **Paper:** Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868}
- **Leaderboard:** https://paperswithcode.com/sota/animal-pose-estimation-on-horse-10?p=pretraining-boosts-out-of-domain-robustness
- **Point of Contact:** Mackenzie Mathis
### Dataset Summary
Pose estimation is an important tool for measuring behavior, and thus widely used in technology, medicine and biology. Due to innovations in both deep learning algorithms and large-scale datasets pose estimation on humans has gotten very powerful. However, typical human pose estimation benchmarks, such as MPII pose and COCO, contain many different individuals (>10K) in different contexts, but only very few example postures per individual. In real world application of pose estimation, users want to estimate the location of user-defined bodyparts by only labeling a few hundred frames on a small subset of individuals, yet want this to generalize to new individuals. Thus, one naturally asks the following question: Assume you have trained an algorithm that performs with high accuracy on a given (individual) animal for the whole repertoire of movement - how well will it generalize to different individuals that have slightly or a dramatically different appearance? Unlike in common human pose estimation benchmarks here the setting is that datasets have many (annotated) poses per individual (>200) but only few individuals (1-25).
To allow the field to tackle this challenge, we developed a novel benchmark, called Horse-10, comprising 30 diverse Thoroughbred horses, for which 22 body parts were labeled by an expert in 8,114 frames. Horses have various coat colors and the โin-the-wildโ aspect of the collected data at various Thoroughbred yearling sales and farms added additional complexity.
- **Homepage:** horse10.deeplabcut.org
- **Repository:** https://github.com/DeepLabCut/DeepLabCut
- **Paper:** `{Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} `
- **Leaderboard:** https://paperswithcode.com/sota/animal-pose-estimation-on-horse-10?p=pretraining-boosts-out-of-domain-robustness
- **Point of Contact:** Mackenzie Mathis
### Supported Tasks and Leaderboards
Horse-10 task: Train on a subset of individuals (10) and evaluate on held-out โout-of-domainโ horses (20).
### Languages
Python, deeplabcut, tensorflow, pytorch
## Dataset Structure
### Data Instances
Over 8,000 expertly labeled frames across 30 individual thoroughbred horses
### Data Splits
The ground truth training data is provided as 3 splits of 10 Horses each. The download provides you a project compatible with loading into the deeplabcut framework, but ground truth labels/training data can be easily loaded in pandas to accommodate your framework (example loader here).
Please do NOT train on all three splits simultaneously. You must train independently (as some horses can be considered out-of-domain in other splits for evaluation!). Integrity matters!
The download also includes all of Horse-30 images and annotations (thus is ~850MB).
| mwmathis/Horse-30 | [
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2022-12-22T13:50:49+00:00 | {"license": "cc-by-nc-sa-4.0"} | 2023-04-19T08:59:42+00:00 | [] | [] | TAGS
#license-cc-by-nc-sa-4.0 #region-us
|
# Dataset Card for Horse-30
## Dataset Description
- Homepage: URL
- Repository: URL
- Paper: Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868}
- Leaderboard: URL
- Point of Contact: Mackenzie Mathis
### Dataset Summary
Pose estimation is an important tool for measuring behavior, and thus widely used in technology, medicine and biology. Due to innovations in both deep learning algorithms and large-scale datasets pose estimation on humans has gotten very powerful. However, typical human pose estimation benchmarks, such as MPII pose and COCO, contain many different individuals (>10K) in different contexts, but only very few example postures per individual. In real world application of pose estimation, users want to estimate the location of user-defined bodyparts by only labeling a few hundred frames on a small subset of individuals, yet want this to generalize to new individuals. Thus, one naturally asks the following question: Assume you have trained an algorithm that performs with high accuracy on a given (individual) animal for the whole repertoire of movement - how well will it generalize to different individuals that have slightly or a dramatically different appearance? Unlike in common human pose estimation benchmarks here the setting is that datasets have many (annotated) poses per individual (>200) but only few individuals (1-25).
To allow the field to tackle this challenge, we developed a novel benchmark, called Horse-10, comprising 30 diverse Thoroughbred horses, for which 22 body parts were labeled by an expert in 8,114 frames. Horses have various coat colors and the โin-the-wildโ aspect of the collected data at various Thoroughbred yearling sales and farms added additional complexity.
- Homepage: URL
- Repository: URL
- Paper: '{Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} '
- Leaderboard: URL
- Point of Contact: Mackenzie Mathis
### Supported Tasks and Leaderboards
Horse-10 task: Train on a subset of individuals (10) and evaluate on held-out โout-of-domainโ horses (20).
### Languages
Python, deeplabcut, tensorflow, pytorch
## Dataset Structure
### Data Instances
Over 8,000 expertly labeled frames across 30 individual thoroughbred horses
### Data Splits
The ground truth training data is provided as 3 splits of 10 Horses each. The download provides you a project compatible with loading into the deeplabcut framework, but ground truth labels/training data can be easily loaded in pandas to accommodate your framework (example loader here).
Please do NOT train on all three splits simultaneously. You must train independently (as some horses can be considered out-of-domain in other splits for evaluation!). Integrity matters!
The download also includes all of Horse-30 images and annotations (thus is ~850MB).
| [
"# Dataset Card for Horse-30",
"## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} \n- Leaderboard: URL\n- Point of Contact: Mackenzie Mathis",
"### Dataset Summary\n\nPose estimation is an important tool for measuring behavior, and thus widely used in technology, medicine and biology. Due to innovations in both deep learning algorithms and large-scale datasets pose estimation on humans has gotten very powerful. However, typical human pose estimation benchmarks, such as MPII pose and COCO, contain many different individuals (>10K) in different contexts, but only very few example postures per individual. In real world application of pose estimation, users want to estimate the location of user-defined bodyparts by only labeling a few hundred frames on a small subset of individuals, yet want this to generalize to new individuals. Thus, one naturally asks the following question: Assume you have trained an algorithm that performs with high accuracy on a given (individual) animal for the whole repertoire of movement - how well will it generalize to different individuals that have slightly or a dramatically different appearance? Unlike in common human pose estimation benchmarks here the setting is that datasets have many (annotated) poses per individual (>200) but only few individuals (1-25). \nTo allow the field to tackle this challenge, we developed a novel benchmark, called Horse-10, comprising 30 diverse Thoroughbred horses, for which 22 body parts were labeled by an expert in 8,114 frames. Horses have various coat colors and the โin-the-wildโ aspect of the collected data at various Thoroughbred yearling sales and farms added additional complexity. \n\n- Homepage: URL\n- Repository: URL\n- Paper: '{Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} '\n- Leaderboard: URL\n- Point of Contact: Mackenzie Mathis",
"### Supported Tasks and Leaderboards\n\nHorse-10 task: Train on a subset of individuals (10) and evaluate on held-out โout-of-domainโ horses (20).",
"### Languages\n\nPython, deeplabcut, tensorflow, pytorch",
"## Dataset Structure",
"### Data Instances\n\nOver 8,000 expertly labeled frames across 30 individual thoroughbred horses",
"### Data Splits\n\nThe ground truth training data is provided as 3 splits of 10 Horses each. The download provides you a project compatible with loading into the deeplabcut framework, but ground truth labels/training data can be easily loaded in pandas to accommodate your framework (example loader here). \n\nPlease do NOT train on all three splits simultaneously. You must train independently (as some horses can be considered out-of-domain in other splits for evaluation!). Integrity matters!\n\nThe download also includes all of Horse-30 images and annotations (thus is ~850MB)."
] | [
"TAGS\n#license-cc-by-nc-sa-4.0 #region-us \n",
"# Dataset Card for Horse-30",
"## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} \n- Leaderboard: URL\n- Point of Contact: Mackenzie Mathis",
"### Dataset Summary\n\nPose estimation is an important tool for measuring behavior, and thus widely used in technology, medicine and biology. Due to innovations in both deep learning algorithms and large-scale datasets pose estimation on humans has gotten very powerful. However, typical human pose estimation benchmarks, such as MPII pose and COCO, contain many different individuals (>10K) in different contexts, but only very few example postures per individual. In real world application of pose estimation, users want to estimate the location of user-defined bodyparts by only labeling a few hundred frames on a small subset of individuals, yet want this to generalize to new individuals. Thus, one naturally asks the following question: Assume you have trained an algorithm that performs with high accuracy on a given (individual) animal for the whole repertoire of movement - how well will it generalize to different individuals that have slightly or a dramatically different appearance? Unlike in common human pose estimation benchmarks here the setting is that datasets have many (annotated) poses per individual (>200) but only few individuals (1-25). \nTo allow the field to tackle this challenge, we developed a novel benchmark, called Horse-10, comprising 30 diverse Thoroughbred horses, for which 22 body parts were labeled by an expert in 8,114 frames. Horses have various coat colors and the โin-the-wildโ aspect of the collected data at various Thoroughbred yearling sales and farms added additional complexity. \n\n- Homepage: URL\n- Repository: URL\n- Paper: '{Mathis, Alexander and Biasi, Thomas and Schneider, Steffen and Yuksekgonul, Mert and Rogers, Byron and Bethge, Matthias and Mathis, Mackenzie W.}, title = {Pretraining Boosts Out-of-Domain Robustness for Pose Estimation}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {1859-1868} '\n- Leaderboard: URL\n- Point of Contact: Mackenzie Mathis",
"### Supported Tasks and Leaderboards\n\nHorse-10 task: Train on a subset of individuals (10) and evaluate on held-out โout-of-domainโ horses (20).",
"### Languages\n\nPython, deeplabcut, tensorflow, pytorch",
"## Dataset Structure",
"### Data Instances\n\nOver 8,000 expertly labeled frames across 30 individual thoroughbred horses",
"### Data Splits\n\nThe ground truth training data is provided as 3 splits of 10 Horses each. The download provides you a project compatible with loading into the deeplabcut framework, but ground truth labels/training data can be easily loaded in pandas to accommodate your framework (example loader here). \n\nPlease do NOT train on all three splits simultaneously. You must train independently (as some horses can be considered out-of-domain in other splits for evaluation!). Integrity matters!\n\nThe download also includes all of Horse-30 images and annotations (thus is ~850MB)."
] |
45f8e5d34bb4489aac82fb6944ef7f73aaaba45e | # Dataset Card for "aveyron_test"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | yacine-djm/aveyron_test | [
"region:us"
] | 2022-12-22T16:22:27+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "label", "sequence": "string"}, {"name": "date", "dtype": "string"}, {"name": "sheet_id", "dtype": "string"}, {"name": "group_id", "dtype": "string"}, {"name": "source", "dtype": "string"}, {"name": "est", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 132770347, "num_examples": 530910}], "download_size": 49780765, "dataset_size": 132770347}} | 2022-12-22T16:22:53+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "aveyron_test"
More Information needed | [
"# Dataset Card for \"aveyron_test\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"aveyron_test\"\n\nMore Information needed"
] |
64fd53cc91f7cb73b283a6e4f661205e277d23c9 | # Dataset Card for "rm-static"
Split of [hh-static](https://huggingface.co/datasets/Dahoas/static-hh) used for training reward models after supervised fine-tuning. | Dahoas/rm-static | [
"region:us"
] | 2022-12-22T16:50:14+00:00 | {"dataset_info": {"features": [{"name": "prompt", "dtype": "string"}, {"name": "response", "dtype": "string"}, {"name": "chosen", "dtype": "string"}, {"name": "rejected", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 113850006, "num_examples": 76256}, {"name": "test", "num_bytes": 7649255, "num_examples": 5103}], "download_size": 73006535, "dataset_size": 121499261}} | 2023-03-06T00:13:07+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "rm-static"
Split of hh-static used for training reward models after supervised fine-tuning. | [
"# Dataset Card for \"rm-static\"\n\nSplit of hh-static used for training reward models after supervised fine-tuning."
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"rm-static\"\n\nSplit of hh-static used for training reward models after supervised fine-tuning."
] |
981e3f95845b9cdd54de5847725da3f12dd9da84 |
# Dataset Card for OLM December 2022 Wikipedia
Pretraining dataset, created with the OLM repo [here](https://github.com/huggingface/olm-datasets) from a December 2022 Wikipedia snapshot. | olm/olm-wikipedia-20221220 | [
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1M<n<10M",
"language:en",
"pretraining",
"language modelling",
"wikipedia",
"web",
"region:us"
] | 2022-12-22T17:38:13+00:00 | {"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": [], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": [], "task_categories": [], "task_ids": [], "pretty_name": "OLM December 2022 Wikipedia", "tags": ["pretraining", "language modelling", "wikipedia", "web"]} | 2022-12-29T03:12:35+00:00 | [] | [
"en"
] | TAGS
#annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #language-English #pretraining #language modelling #wikipedia #web #region-us
|
# Dataset Card for OLM December 2022 Wikipedia
Pretraining dataset, created with the OLM repo here from a December 2022 Wikipedia snapshot. | [
"# Dataset Card for OLM December 2022 Wikipedia\n\nPretraining dataset, created with the OLM repo here from a December 2022 Wikipedia snapshot."
] | [
"TAGS\n#annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #language-English #pretraining #language modelling #wikipedia #web #region-us \n",
"# Dataset Card for OLM December 2022 Wikipedia\n\nPretraining dataset, created with the OLM repo here from a December 2022 Wikipedia snapshot."
] |
62e85e84102b5ebb6a412cf0f4245227ea897a08 |
## Dataset Description
- **Repository:** https://github.com/shuyanzhou/docprompting
- **Paper:** [DocPrompting: Generating Code by Retrieving the Docs](https://arxiv.org/pdf/2207.05987.pdf)
### Dataset Summary
This is the natural language to bash generation dataset we harvested from the English subset of [`tldr`](https://github.com/tldr-pages/tldr)
We split the dataset by bash commands. Every command in the dev and test set is held out from the training set.
### Supported Tasks and Leaderboards
This dataset is used to evaluate code generations.
### Languages
English - Bash
## Dataset Structure
```python
dataset = load_dataset("neulab/tldr")
DatasetDict({
train: Dataset({
features: ['question_id', 'nl', 'cmd', 'oracle_man', 'cmd_name', 'tldr_cmd_name', 'manual_exist', 'matching_info'],
num_rows: 6414
})
test: Dataset({
features: ['question_id', 'nl', 'cmd', 'oracle_man', 'cmd_name', 'tldr_cmd_name', 'manual_exist', 'matching_info'],
num_rows: 928
})
validation: Dataset({
features: ['question_id', 'nl', 'cmd', 'oracle_man', 'cmd_name', 'tldr_cmd_name', 'manual_exist', 'matching_info'],
num_rows: 1845
})
})
code_docs = load_dataset("neulab/docprompting-conala", "docs")
DatasetDict({
train: Dataset({
features: ['doc_id', 'doc_content'],
num_rows: 439064
})
})
```
### Data Fields
train/dev/test:
- nl: The natural language intent
- cmd: The reference code snippet
- question_id: the unique id of a question
- oracle_man: The `doc_id` of the functions used in the reference code snippet. The corresponding contents are in `doc` split
- cmd_name: the bash command of this code snippet
- tldr_cmd_name: the bash command used in tldr github repo. The `cmd_name` and `tldr_cmd_name` can be different due to naming difference
- manual_exist: whether the manual exists in https://manned.org
- matching_info: each code snippets have multiple tokens, this is the detailed reference doc matching on each token.
docs:
- doc_id: the id of a doc
- doc_content: the content of the doc
## Dataset Creation
The dataset was curated from [`tldr`](https://github.com/tldr-pages/tldr).
The project aims to provide frequent usage of bash commands with natural language intents.
For more details, please check the repo.
### Citation Information
```
@article{zhou2022doccoder,
title={DocCoder: Generating Code by Retrieving and Reading Docs},
author={Zhou, Shuyan and Alon, Uri and Xu, Frank F and Jiang, Zhengbao and Neubig, Graham},
journal={arXiv preprint arXiv:2207.05987},
year={2022}
}
``` | neulab/tldr | [
"task_categories:text2text-generation",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:unknown",
"source_datasets:original",
"language:code",
"license:mit",
"code-generation",
"doc retrieval",
"retrieval augmented generation",
"arxiv:2207.05987",
"region:us"
] | 2022-12-22T17:58:43+00:00 | {"annotations_creators": [], "language_creators": ["crowdsourced", "expert-generated"], "language": ["code"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["unknown"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "pretty_name": "DocPrompting-CoNaLa", "tags": ["code-generation", "doc retrieval", "retrieval augmented generation"]} | 2022-12-22T19:47:11+00:00 | [
"2207.05987"
] | [
"code"
] | TAGS
#task_categories-text2text-generation #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-unknown #source_datasets-original #language-code #license-mit #code-generation #doc retrieval #retrieval augmented generation #arxiv-2207.05987 #region-us
|
## Dataset Description
- Repository: URL
- Paper: DocPrompting: Generating Code by Retrieving the Docs
### Dataset Summary
This is the natural language to bash generation dataset we harvested from the English subset of 'tldr'
We split the dataset by bash commands. Every command in the dev and test set is held out from the training set.
### Supported Tasks and Leaderboards
This dataset is used to evaluate code generations.
### Languages
English - Bash
## Dataset Structure
### Data Fields
train/dev/test:
- nl: The natural language intent
- cmd: The reference code snippet
- question_id: the unique id of a question
- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split
- cmd_name: the bash command of this code snippet
- tldr_cmd_name: the bash command used in tldr github repo. The 'cmd_name' and 'tldr_cmd_name' can be different due to naming difference
- manual_exist: whether the manual exists in URL
- matching_info: each code snippets have multiple tokens, this is the detailed reference doc matching on each token.
docs:
- doc_id: the id of a doc
- doc_content: the content of the doc
## Dataset Creation
The dataset was curated from 'tldr'.
The project aims to provide frequent usage of bash commands with natural language intents.
For more details, please check the repo.
| [
"## Dataset Description\n- Repository: URL\n- Paper: DocPrompting: Generating Code by Retrieving the Docs",
"### Dataset Summary\nThis is the natural language to bash generation dataset we harvested from the English subset of 'tldr'\nWe split the dataset by bash commands. Every command in the dev and test set is held out from the training set.",
"### Supported Tasks and Leaderboards\nThis dataset is used to evaluate code generations.",
"### Languages\nEnglish - Bash",
"## Dataset Structure",
"### Data Fields\ntrain/dev/test:\n- nl: The natural language intent\n- cmd: The reference code snippet\n- question_id: the unique id of a question\n- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split\n- cmd_name: the bash command of this code snippet\n- tldr_cmd_name: the bash command used in tldr github repo. The 'cmd_name' and 'tldr_cmd_name' can be different due to naming difference\n- manual_exist: whether the manual exists in URL\n- matching_info: each code snippets have multiple tokens, this is the detailed reference doc matching on each token.\n \n\ndocs:\n- doc_id: the id of a doc\n- doc_content: the content of the doc",
"## Dataset Creation\nThe dataset was curated from 'tldr'. \nThe project aims to provide frequent usage of bash commands with natural language intents. \nFor more details, please check the repo."
] | [
"TAGS\n#task_categories-text2text-generation #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-unknown #source_datasets-original #language-code #license-mit #code-generation #doc retrieval #retrieval augmented generation #arxiv-2207.05987 #region-us \n",
"## Dataset Description\n- Repository: URL\n- Paper: DocPrompting: Generating Code by Retrieving the Docs",
"### Dataset Summary\nThis is the natural language to bash generation dataset we harvested from the English subset of 'tldr'\nWe split the dataset by bash commands. Every command in the dev and test set is held out from the training set.",
"### Supported Tasks and Leaderboards\nThis dataset is used to evaluate code generations.",
"### Languages\nEnglish - Bash",
"## Dataset Structure",
"### Data Fields\ntrain/dev/test:\n- nl: The natural language intent\n- cmd: The reference code snippet\n- question_id: the unique id of a question\n- oracle_man: The 'doc_id' of the functions used in the reference code snippet. The corresponding contents are in 'doc' split\n- cmd_name: the bash command of this code snippet\n- tldr_cmd_name: the bash command used in tldr github repo. The 'cmd_name' and 'tldr_cmd_name' can be different due to naming difference\n- manual_exist: whether the manual exists in URL\n- matching_info: each code snippets have multiple tokens, this is the detailed reference doc matching on each token.\n \n\ndocs:\n- doc_id: the id of a doc\n- doc_content: the content of the doc",
"## Dataset Creation\nThe dataset was curated from 'tldr'. \nThe project aims to provide frequent usage of bash commands with natural language intents. \nFor more details, please check the repo."
] |
5da017b06ca90943494f28c8333a71077e3736b3 | https://mrcheeze.github.io/musenet-midi/ is used to get the musenet encoding.
where i got the midi files I used to convert into "codes" are all from https://bitmidi.com
| breadlicker45/midi-music-codes | [
"region:us"
] | 2022-12-22T18:18:41+00:00 | {} | 2023-01-10T12:37:38+00:00 | [] | [] | TAGS
#region-us
| URL is used to get the musenet encoding.
where i got the midi files I used to convert into "codes" are all from URL
| [] | [
"TAGS\n#region-us \n"
] |
547fa98342d2c80734a1c841dbf3de2cfcaaab95 | # Dataset Card for "financial_news_sentiment"
Manually validated sentiment for ~2000 Canadian news articles.
The dataset also include a column topic which contains one of the following value:
* acquisition
* other
* quaterly financial release
* appointment to new position
* dividend
* corporate update
* drillings results
* conference
* share repurchase program
* grant of stocks
This was generated automatically using a zero-shot classification model and **was not** reviewed manually. | Jean-Baptiste/financial_news_sentiment | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"task_ids:sentiment-classification",
"annotations_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"language:en",
"license:mit",
"region:us"
] | 2022-12-22T18:49:05+00:00 | {"annotations_creators": ["expert-generated"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification", "sentiment-classification"], "pretty_name": "financial_news_sentiment", "dataset_info": {"splits": [{"name": "test", "num_examples": 267}, {"name": "train", "num_examples": 1512}]}, "tags": []} | 2022-12-29T03:14:44+00:00 | [] | [
"en"
] | TAGS
#task_categories-text-classification #task_ids-multi-class-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #language-English #license-mit #region-us
| # Dataset Card for "financial_news_sentiment"
Manually validated sentiment for ~2000 Canadian news articles.
The dataset also include a column topic which contains one of the following value:
* acquisition
* other
* quaterly financial release
* appointment to new position
* dividend
* corporate update
* drillings results
* conference
* share repurchase program
* grant of stocks
This was generated automatically using a zero-shot classification model and was not reviewed manually. | [
"# Dataset Card for \"financial_news_sentiment\"\n\nManually validated sentiment for ~2000 Canadian news articles.\n\nThe dataset also include a column topic which contains one of the following value:\n* acquisition\n* other\n* quaterly financial release\n* appointment to new position\n* dividend\n* corporate update\n* drillings results\n* conference\n* share repurchase program\n* grant of stocks\n\nThis was generated automatically using a zero-shot classification model and was not reviewed manually."
] | [
"TAGS\n#task_categories-text-classification #task_ids-multi-class-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #language-English #license-mit #region-us \n",
"# Dataset Card for \"financial_news_sentiment\"\n\nManually validated sentiment for ~2000 Canadian news articles.\n\nThe dataset also include a column topic which contains one of the following value:\n* acquisition\n* other\n* quaterly financial release\n* appointment to new position\n* dividend\n* corporate update\n* drillings results\n* conference\n* share repurchase program\n* grant of stocks\n\nThis was generated automatically using a zero-shot classification model and was not reviewed manually."
] |
dab57d94d2f44f286477933ae39f7764eca37c2c | # Dataset Card for "dataset-v-1.4_CLIP_us_identities_random_seeds"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | SDbiaseval/dataset-v-1.4_CLIP_us_identities_random_seeds | [
"region:us"
] | 2022-12-22T19:10:30+00:00 | {"dataset_info": {"features": [{"name": "adjective", "dtype": "string"}, {"name": "profession", "dtype": "string"}, {"name": "no", "dtype": "int32"}, {"name": "image_path", "dtype": "string"}, {"name": "image", "dtype": "image"}, {"name": "gender", "dtype": "string"}, {"name": "identity", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1172794597.5, "num_examples": 31500}], "download_size": 1167645236, "dataset_size": 1172794597.5}} | 2022-12-22T19:11:43+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "dataset-v-1.4_CLIP_us_identities_random_seeds"
More Information needed | [
"# Dataset Card for \"dataset-v-1.4_CLIP_us_identities_random_seeds\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"dataset-v-1.4_CLIP_us_identities_random_seeds\"\n\nMore Information needed"
] |
54fb216b93e85cbeef482ce1fd13194a519c6382 | # AutoTrain Dataset for project: auto-arabic-summarization
## Dataset Description
This dataset has been automatically processed by AutoTrain for project auto-arabic-summarization.
### Languages
The BCP-47 code for the dataset's language is unk.
## Dataset Structure
### Data Instances
A sample from this dataset looks as follows:
```json
[
{
"text": "\u0627\u0643\u062f \u0648\u0632\u064a\u0631 \u0627\u0644\u0635\u0646\u0627\u0639\u0647 \u0648\u0627\u0644\u0637\u0627\u0642\u0647 \u0648\u0627\u0644\u0645\u0646\u0627\u062c\u0645 \u0632\u0643\u0631\u064a\u0627 \u062d\u0645\u062f \u0627\u0646\u0647 \u062a\u0645 \u0627\u0644\u064a\u0648\u0645 \u0627\u0644\u062e\u0645\u064a\u0633 \u062e\u0644\u0627\u0644 \u062c\u0644\u0633\u0647 \u0627\u0644\u062a\u0627\u0645\u062a \u0628\u0627\u0644\u0639\u0627\u0635\u0645\u0647 \u0648\u0632\u064a\u0631 \u0627\u0644\u0637\u0627\u0642\u0647 \u0627\u0644\u062c\u0632\u0627\u0626\u064a \u0635\u0627\u0644\u062d \u062e\u0628\u0631\u064a \u0628\u062e\u0635\u0648\u0635 \u0627\u0634\u063a\u0627\u0644 \u0627\u0644\u0644\u062c\u0646\u0647 \u0627\u0644\u062a\u0648\u0646\u0633\u064a\u0647 \u0627\u0644\u062c\u0632\u0627\u0626\u0631\u064a\u0647 \u0645\u062c\u0627\u0644 \u0627\u0644\u0637\u0627\u0642\u0647 \u0644\u062a\u0642\u064a\u064a\u0645 \u0645\u062f\u0649 \u062a\u0637\u0628\u064a\u0642 \u0627\u0644\u0628\u0631\u0627\u0645\u062c \u0627\u0644\u0645\u062a\u0641\u0642 \u0639\u0644\u064a\u0647\u0627 \u062e\u0628\u0631\u0627\u0621 \u0627\u0644\u0628\u0644\u062f\u064a\u0646 \u0627\u0644\u0627\u062a\u0641\u0627\u0642 \u062a\u0632\u0648\u064a\u062f \u0627\u0644\u0645\u0646\u0627\u0637\u0642 \u0627\u0644\u062d\u062f\u0648\u062f\u064a\u0647 \u0627\u0644\u062a\u0648\u0646\u0633\u064a\u0647 \u0628\u0627\u0644\u0643\u0645\u064a\u0627\u062a \u0627\u0644\u0643\u0627\u0641\u064a\u0647 \u0642\u0648\u0627\u0631\u064a\u0631 \u0627\u0644\u063a\u0627\u0632 \u0627\u0644\u0645\u0646\u0632\u0644\u064a \u062a\u0642\u062f\u0631 \u0628\u062d\u0648\u0627\u0644\u064a \u0637\u0646 \u0627\u0644\u0642\u0648\u0627\u0631\u064a\u0631 \u0648\u0627\u0636\u0627\u0641 \u062d\u0645\u062f \u0627\u0646\u0647 \u0627\u0644\u0646\u0642\u0627\u0637 \u062a\u0645 \u0627\u0644\u0627\u062a\u0641\u0627\u0642 \u0628\u0634\u0627\u0646\u0647\u0627 \u062c\u0644\u0633\u0647 \u0627\u0644\u064a\u0648\u0645 \u062a\u0632\u0648\u064a\u062f \u0627\u0644\u0633\u0648\u0642 \u0627\u0644\u062a\u0648\u0646\u0633\u064a\u0647 \u0628\u0627\u0644\u063a\u0627\u0632 \u0627\u0644\u0637\u0628\u064a\u0639\u064a \u0639\u0628\u0631 \u0627\u0644\u0627\u0646\u0627\u0628\u064a\u0628 \u0648\u062a\u0632\u0648\u064a\u062f \u0627\u0644\u0645\u0646\u0627\u0637\u0642 \u0628\u0627\u0644\u0628\u062a\u0631\u0648\u0644 \u0627\u0644\u0645\u0633\u0627\u0644 \u0627\u0636\u0627\u0641\u0647 \u0627\u0644\u0649 \u062f\u0639\u0645 \u0627\u0644\u062a\u0639\u0627\u0648\u0646 \u0627\u0644\u0645\u062c\u0627\u0644 \u0627\u0644\u062a\u062c\u0627\u0631\u064a \u062a\u0645 \u0627\u0645\u0636\u0627\u0621 \u0645\u0630\u0643\u0631\u0647 \u062a\u0641\u0627\u0647\u0645 \u0639\u0642\u062f \u0644\u062a\u0643\u0648\u064a\u0646 \u062a\u0642\u0646\u0646\u064a\u0646 \u062a\u0648\u0646\u0633\u064a\u064a\u0646 \u0627\u0644\u062c\u0632\u0627\u0626\u0631",
"target": "\u0643\u0645\u0627 \u062a\u0645 \u0627\u0645\u0636\u0627\u0621 \u0645\u0630\u0643\u0631\u0629 \u062a\u0641\u0627\u0647\u0645 \u0639\u0642\u062f \u0644\u062a\u0643\u0648\u064a\u0646 \u062a\u0642\u0646\u0646\u064a\u0646 \u062a\u0648\u0646\u0633\u064a\u064a\u0646 \u0641\u064a \u0627\u0644\u062c\u0632\u0627\u0626\u0631 ."
},
{
"text": "\u0642\u0627\u0644 \u0627\u0644\u0648\u0632\u064a\u0631 \u0627\u0644\u0627\u0648\u0644 \u0627\u0644\u062c\u0632\u0627\u0626\u0631\u064a \u0639\u0628\u062f \u0627\u0644\u0645\u0627\u0644\u0643 \u0633\u0644\u0627\u0644 \u0627\u062b\u0631 \u0644\u0642\u0627\u0621 \u062c\u0645\u0639\u0647 \u0628\u0631\u0626\u064a\u0633 \u0645\u062c\u0644\u0633 \u0646\u0648\u0627\u0628 \u0627\u0644\u0634\u0639\u0628 \u0645\u062d\u0645\u062f \u0627\u0644\u0646\u0627\u0635\u0631 \u0627\u0644\u0639\u0644\u0627\u0642\u0627\u062a \u0627\u0644\u062b\u0646\u0627\u0626\u064a\u0647 \u0627\u0644\u0628\u0644\u062f\u064a\u0646 \u0645\u0645\u064a\u0632\u0647 \u0648\u0633\u062a\u0643\u0648\u0646 \u0627\u062d\u0633\u0646 \u062e\u0644\u0627\u0644 \u0627\u0644\u0641\u062a\u0631\u0647 \u0627\u0644\u0642\u0627\u062f\u0645\u0647 \u0648\u0627\u0636\u0627\u0641 \u062a\u0635\u0631\u064a\u062d \u0644\u0645\u0631\u0627\u0633\u0644 \u0627\u0644\u062c\u0648\u0647\u0631\u0647 \u0627\u0641 \u0627\u0645 \u0627\u0646\u0647 \u0639\u0627\u0647\u062f \u0631\u0626\u064a\u0633 \u0627\u0644\u0645\u062c\u0644\u0633 \u0628\u0627\u0644\u0645\u062d\u0627\u0641\u0638\u0647 \u0645\u062a\u0627\u0646\u0647 \u0627\u0644\u0639\u0644\u0627\u0642\u0647 \u0627\u0644\u0628\u0644\u062f\u064a\u0646 \u0648\u0645\u0648\u0627\u0635\u0644\u0647 \u0627\u0644\u062a\u0642\u062f\u0645 \u0648\u0627\u0644\u0639\u0645\u0644 \u0645\u0639\u0627 \u0648\u0627\u0648\u0636\u062d \u0639\u0628\u062f \u0627\u0644\u0645\u0627\u0644\u0643 \u0633\u0644\u0627\u0644 \u0645\u062d\u0645\u062f \u0627\u0644\u0646\u0627\u0635\u0631 \u0627\u0628\u062f\u0649 \u062f\u0639\u0645\u0647 \u0644\u0644\u0645\u0646\u0647\u062c \u062a\u0646\u062a\u0647\u062c\u0647 \u0627\u0644\u062c\u0632\u0627\u0626\u0631 \u0648\u0639\u0645\u0644\u0647\u0627 \u0648\u064a\u0627\u062a\u064a \u0627\u062c\u062a\u0645\u0627\u0639 \u0627\u0644\u0648\u0632\u064a\u0631 \u0627\u0644\u0627\u0648\u0644 \u0627\u0644\u062c\u0632\u0627\u0626\u0631\u064a \u0628\u0631\u0626\u064a\u0633 \u0627\u0644\u0645\u062c\u0644\u0633 \u0647\u0627\u0645\u0634 \u0632\u064a\u0627\u0631\u0647 \u0639\u0645\u0644 \u0627\u062f\u0627\u0647\u0627 \u0627\u0644\u064a\u0648\u0645 \u0627\u0644\u062e\u0645\u064a\u0633 \u062a\u0648\u0646\u0633 \u062a\u0631\u0627\u0633 \u062e\u0644\u0627\u0644\u0647\u0627 \u0627\u0634\u063a\u0627\u0644 \u0627\u0644\u062f\u0648\u0631\u0647 \u0627\u0644 \u0644\u0644\u062c\u0646\u0647 \u0627\u0644\u0645\u062e\u062a\u0644\u0637\u0647 \u0627\u0644\u0639\u0644\u064a\u0627 \u0627\u0644\u062a\u0648\u0646\u0633\u064a\u0647 \u0627\u0644\u062c\u0632\u0627\u0626\u0631\u064a\u0647 \u0631\u0641\u0642\u0647 \u0631\u0626\u064a\u0633 \u0627\u0644\u062d\u0643\u0648\u0645\u0647 \u064a\u0648\u0633\u0641 \u0627\u0644\u0634\u0627\u0647\u062f \u0648\u0627\u0644\u062a\u064a \u0627\u0646\u062a\u0647\u062a \u0628\u0627\u0644\u0645\u0635\u0627\u062f\u0642\u0647 \u0639\u062f\u064a\u062f \u0627\u0644\u0627\u062a\u0641\u0627\u0642\u064a\u0627\u062a \u062a\u0648\u0646\u0633 \u0648\u0627\u0644\u062c\u0632\u0627\u0626\u0631",
"target": "\n\u0642\u0627\u0644 \u0627\u0644\u0648\u0632\u064a\u0631 \u0627\u0644\u0623\u0648\u0644 \u0627\u0644\u062c\u0632\u0627\u0626\u0631\u064a \u0639\u0628\u062f \u0627\u0644\u0645\u0627\u0644\u0643 \u0633\u0644\u0627\u0644 \u0627\u062b\u0631 \u0644\u0642\u0627\u0621 \u062c\u0645\u0639\u0647 \u0628\u0631\u0626\u064a\u0633 \u0645\u062c\u0644\u0633 \u0646\u0648\u0627\u0628 \u0627\u0644\u0634\u0639\u0628 \u0645\u062d\u0645\u062f \u0627\u0644\u0646\u0627\u0635\u0631\u060c \u0625\u0646 \u0627\u0644\u0639\u0644\u0627\u0642\u0627\u062a \u0627\u0644\u062b\u0646\u0627\u0626\u064a\u0629 \u0628\u064a\u0646 \u0627\u0644\u0628\u0644\u062f\u064a\u0646 \u0645\u0645\u064a\u0632\u0629 \u0648\u0633\u062a\u0643\u0648\u0646 \u0623\u062d\u0633\u0646 \u062e\u0644\u0627\u0644 \u0627\u0644\u0641\u062a\u0631\u0629 \u0627\u0644\u0642\u0627\u062f\u0645\u0629."
}
]
```
### Dataset Fields
The dataset has the following fields (also called "features"):
```json
{
"text": "Value(dtype='string', id=None)",
"target": "Value(dtype='string', id=None)"
}
```
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| Split name | Num samples |
| ------------ | ------------------- |
| train | 5102 |
| valid | 1276 |
| abdalrahmanshahrour/autotrain-data-auto-arabic-summarization | [
"task_categories:summarization",
"region:us"
] | 2022-12-22T19:15:40+00:00 | {"task_categories": ["summarization"]} | 2022-12-22T19:20:04+00:00 | [] | [] | TAGS
#task_categories-summarization #region-us
| AutoTrain Dataset for project: auto-arabic-summarization
========================================================
Dataset Description
-------------------
This dataset has been automatically processed by AutoTrain for project auto-arabic-summarization.
### Languages
The BCP-47 code for the dataset's language is unk.
Dataset Structure
-----------------
### Data Instances
A sample from this dataset looks as follows:
### Dataset Fields
The dataset has the following fields (also called "features"):
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| [
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] | [
"TAGS\n#task_categories-summarization #region-us \n",
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
] |
64bb548ebe3f6db721e96520b8b0b18d7633aa47 |
# DreamBooth model for pugsly trained by lewtun on the Shirleyphd/Pug dataset.
This is a Stable Diffusion model fine-tuned the ccorgi concept taught to Stable Diffusion with DreamBooth.
It can be used by modifying the `instance_prompt`: **a photo of pugsly dog**
This model was created as part of the DreamBooth Hackathon ๐ฅ. Visit the [organisation page](https://huggingface.co/dreambooth-hackathon) for instructions on how to take part!
## Description
This is a Stable Diffusion model fine-tuned on `dog` images for the animal theme.
## Usage
```python
from diffusers import StableDiffusionPipeline
pipeline = StableDiffusionPipeline.from_pretrained('Shirleyphd/Pug-dog')
image = pipeline().images[0]
image
``` | Shirleyphd/Pug | [
"license:creativeml-openrail-m",
"pytorch",
"diffusers",
"stable-diffusion",
"text-to-image",
"diffusion-models-class",
"dreambooth-hackathon",
"animal",
"region:us"
] | 2022-12-22T21:05:20+00:00 | {"license": "creativeml-openrail-m", "tags": ["pytorch", "diffusers", "stable-diffusion", "text-to-image", "diffusion-models-class", "dreambooth-hackathon", "animal"], "widget": [{"text": "a photo of pug dog in a cup"}]} | 2022-12-22T21:39:10+00:00 | [] | [] | TAGS
#license-creativeml-openrail-m #pytorch #diffusers #stable-diffusion #text-to-image #diffusion-models-class #dreambooth-hackathon #animal #region-us
|
# DreamBooth model for pugsly trained by lewtun on the Shirleyphd/Pug dataset.
This is a Stable Diffusion model fine-tuned the ccorgi concept taught to Stable Diffusion with DreamBooth.
It can be used by modifying the 'instance_prompt': a photo of pugsly dog
This model was created as part of the DreamBooth Hackathon . Visit the organisation page for instructions on how to take part!
## Description
This is a Stable Diffusion model fine-tuned on 'dog' images for the animal theme.
## Usage
| [
"# DreamBooth model for pugsly trained by lewtun on the Shirleyphd/Pug dataset.\n\nThis is a Stable Diffusion model fine-tuned the ccorgi concept taught to Stable Diffusion with DreamBooth.\nIt can be used by modifying the 'instance_prompt': a photo of pugsly dog\n\nThis model was created as part of the DreamBooth Hackathon . Visit the organisation page for instructions on how to take part!",
"## Description\n\n\nThis is a Stable Diffusion model fine-tuned on 'dog' images for the animal theme.",
"## Usage"
] | [
"TAGS\n#license-creativeml-openrail-m #pytorch #diffusers #stable-diffusion #text-to-image #diffusion-models-class #dreambooth-hackathon #animal #region-us \n",
"# DreamBooth model for pugsly trained by lewtun on the Shirleyphd/Pug dataset.\n\nThis is a Stable Diffusion model fine-tuned the ccorgi concept taught to Stable Diffusion with DreamBooth.\nIt can be used by modifying the 'instance_prompt': a photo of pugsly dog\n\nThis model was created as part of the DreamBooth Hackathon . Visit the organisation page for instructions on how to take part!",
"## Description\n\n\nThis is a Stable Diffusion model fine-tuned on 'dog' images for the animal theme.",
"## Usage"
] |
ab589dfd37cb6b1c3cd420c8c922389e36546b9a | # Dataset Card for "ev-skins-blip-lg"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | jubba/ev-skins-blip-lg | [
"region:us"
] | 2022-12-22T21:10:48+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 13554378.0, "num_examples": 215}], "download_size": 13363408, "dataset_size": 13554378.0}} | 2022-12-22T21:10:54+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "ev-skins-blip-lg"
More Information needed | [
"# Dataset Card for \"ev-skins-blip-lg\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"ev-skins-blip-lg\"\n\nMore Information needed"
] |
5c157f65faed02d46d37c30068d727a4a8dad7bf | # Dataset Card for "booksum-fullbooks"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | abertsch/booksum-fullbooks | [
"region:us"
] | 2022-12-22T21:43:49+00:00 | {"dataset_info": {"features": [{"name": "bid", "dtype": "string"}, {"name": "source", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "book", "dtype": "string"}], "splits": [{"name": "validation", "num_bytes": 23586559, "num_examples": 45}, {"name": "train", "num_bytes": 165182724, "num_examples": 314}, {"name": "test", "num_bytes": 31094987, "num_examples": 46}], "download_size": 60336046, "dataset_size": 219864270}} | 2022-12-22T21:44:19+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "booksum-fullbooks"
More Information needed | [
"# Dataset Card for \"booksum-fullbooks\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"booksum-fullbooks\"\n\nMore Information needed"
] |
ebe18349a55891244f20dbcd8513f1c349c0c4b4 | # Dataset Card for "stable-diffusion-prompts"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | yizhangliu/stable-diffusion-prompts | [
"region:us"
] | 2022-12-23T00:17:08+00:00 | {"dataset_info": {"features": [{"name": "prompt", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 284636284, "num_examples": 1819808}], "download_size": 101931398, "dataset_size": 284636284}} | 2022-12-23T00:17:18+00:00 | [] | [] | TAGS
#region-us
| # Dataset Card for "stable-diffusion-prompts"
More Information needed | [
"# Dataset Card for \"stable-diffusion-prompts\"\n\nMore Information needed"
] | [
"TAGS\n#region-us \n",
"# Dataset Card for \"stable-diffusion-prompts\"\n\nMore Information needed"
] |
088804c86ffc82367bf84c51500f09701919cada | TBA | research-backup/qa_squadshifts_synthetic_random | [
"region:us"
] | 2022-12-23T03:26:02+00:00 | {} | 2023-01-15T18:58:41+00:00 | [] | [] | TAGS
#region-us
| TBA | [] | [
"TAGS\n#region-us \n"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.