id
stringlengths
2
115
private
bool
1 class
tags
sequence
description
stringlengths
0
5.93k
downloads
int64
0
1.14M
likes
int64
0
1.79k
g4849d/abcde
false
[]
null
0
0
kentsui/basemath
false
[]
null
0
0
vldsavelyev/murakami
false
[]
Russian translations of Murakami novels, to fine-tune a generative language model. Source is FB2 files from http://flibusta.is/a/8570.
11
0
g4849d/nva-Mai
false
[]
null
0
0
jamescalam/langchain-docs
false
[]
null
409
4
cahya/instructions-all
false
[]
null
17
0
kilog/LUJH3
false
[]
null
0
0
michaelthwan/wiki_qa_bart_10000row
false
[ "license:mit" ]
null
0
0
merror/custom
false
[ "license:other" ]
null
0
0
semeru/completeformer_java_data
false
[ "license:cc-by-4.0" ]
This new dataset is designed to solve this great NLP task and is crafted with a lot of care.
17
0
Sleoruiz/disc_cla_primera-2
false
[]
null
0
0
Sleoruiz/disc_cla_segunda-2
false
[]
null
0
0
Sleoruiz/disc_cla_tercera-2
false
[]
null
0
0
Sleoruiz/disc_cla_cuarta-2
false
[]
null
0
0
Sleoruiz/disc_cla_quinta-2
false
[]
null
0
0
Sleoruiz/disc_cla_sexta-2
false
[]
null
0
0
Sleoruiz/disc_cla_septima-2
false
[]
null
3
0
cyanic-selkie/aida-conll-yago-wikidata
false
[ "task_categories:token-classification", "size_categories:10K<n<100K", "language:en", "license:cc-by-sa-3.0", "wikidata", "wikipedia", "named-entity-linking" ]
null
20
2
minnq/dataset
false
[ "license:mit" ]
null
0
0
michaelthwan/oa_wiki_qa_bart_10000row
false
[]
null
527
1
Thewillonline/gpt4
false
[ "language:en" ]
null
255
0
Mesutt/deneme2
false
[]
null
0
0
marasama/Ajax_azurlane
false
[]
null
0
0
brainyletter/nancy_drew
false
[]
null
0
0
Nastul/ddpm-butterflies-128
false
[]
null
0
0
albertvillanova/test2
false
[]
Lorem ipsum
0
0
marasama/nva-ajax_azurlane
false
[]
null
0
0
siemvaessen/iati
false
[ "license:other" ]
null
0
0
TSawyer/rku
false
[]
null
0
0
ywpl/SD_config
false
[]
null
0
0
Cyberzos/your-dataset-name
false
[]
null
0
0
jonathan-roberts1/SATIN
false
[ "task_categories:image-classification", "task_categories:zero-shot-image-classification", "size_categories:100K<n<1M", "language:en", "license:other" ]
null
4
0
prvInSpace/banc-trawsgrifiadau-bangor
false
[ "task_categories:automatic-speech-recognition", "size_categories:10K<n<100K", "language:cy", "license:cc0-1.0" ]
Huggingface Dataset version of Banc Trawsgrifiadau Bangor
4
0
HiTZ/alpaca_mt
false
[ "task_categories:text-generation", "task_ids:dialogue-modeling", "annotations_creators:no-annotation", "language_creators:machine-generated", "multilinguality:multilingual", "multilinguality:translation", "size_categories:10K<n<100K", "source_datasets:tatsu-lab/alpaca", "language:en", "language:pt", "language:es", "language:ca", "language:eu", "language:gl", "language:at", "license:cc-by-nc-4.0", "instruction-finetuning" ]
Alpaca is a dataset of 52,000 instructions and demonstrations generated by OpenAI's text-davinci-003 engine. This instruction data can be used to conduct instruction-tuning for language models and make the language model follow instruction better. This dataset also includes machine-translated data for 6 Iberian languages: Portuguese, Spanish, Catalan, Basque, Galician and Asturian.
82
4
DanielVelaJ/codeparrot-train
false
[]
null
0
0
homangab/robotcombinedroboset
false
[]
null
18
0
cahya/instructions-ar
false
[]
null
0
0
cahya/instructions-bg
false
[]
null
0
0
cahya/instructions-bn
false
[]
null
0
0
cahya/instructions-ca
false
[]
null
0
0
cahya/instructions-el
false
[]
null
0
0
cahya/instructions-et
false
[]
null
0
0
cahya/instructions-fi
false
[]
null
0
0
cahya/instructions-ht
false
[]
null
0
0
cahya/instructions-it
false
[]
null
0
0
cahya/instructions-ko
false
[]
null
0
0
cahya/instructions-sw
false
[]
null
0
0
cahya/instructions-ta
false
[]
null
0
0
cahya/instructions-tr
false
[]
null
0
0
cahya/instructions-ur
false
[]
null
0
0
ItchyFingaz/data
false
[]
null
0
0
pythainlp/thainer-corpus-v2
false
[ "task_categories:token-classification", "language:th", "license:cc-by-3.0" ]
null
5
0
semeru/code-text-go
false
[ "license:mit", "arxiv:1909.09436" ]
null
0
0
cambridgeltl/vsr_random
false
[ "task_categories:text-classification", "task_categories:question-answering", "size_categories:10K<n<100K", "language:en", "license:cc-by-4.0", "multimodality", "vision-and-language", "arxiv:2205.00363" ]
null
11
1
SunDog/Files
false
[]
null
0
0
semeru/code-text-java
false
[ "license:mit", "arxiv:1909.09436" ]
null
1
0
semeru/code-text-javascript
false
[ "license:mit", "arxiv:1909.09436" ]
null
8
0
semeru/code-text-php
false
[ "license:mit", "arxiv:1909.09436" ]
null
0
0
semeru/code-text-python
false
[ "license:mit", "arxiv:1909.09436" ]
null
10
1
semeru/code-text-ruby
false
[ "license:mit", "arxiv:1909.09436" ]
null
0
0
cambridgeltl/vsr_zeroshot
false
[ "task_categories:text-classification", "task_categories:question-answering", "size_categories:1K<n<10K", "language:en", "license:cc-by-4.0", "multimodal", "vision-and-language", "arxiv:2205.00363" ]
null
14
1
DanielVelaJ/codeparrot-valid
false
[]
null
0
0
cQueenccc/Vivian-Blip-Captions
false
[ "task_categories:text-to-image", "annotations_creators:machine-generated", "size_categories:n<1K", "language:en" ]
null
7
5
cahya/alpaca-id
false
[ "license:mit" ]
null
0
0
alamshoaib134/Metrics-Dataset
false
[ "size_categories:1K<n<10K" ]
null
0
0
Hrishikesh332/autotrain-data-meme-classification
false
[ "task_categories:image-classification" ]
null
0
0
kilog/LX5
false
[]
null
0
0
genjib/qvh
false
[]
null
0
0
marcus2000/Names2chinese
false
[]
null
0
0
ethanbodnar/emojis
false
[]
null
6
0
cybfi/cyber-2006-unrated
false
[ "license:mit" ]
null
0
0
lab156/github-issues
false
[]
null
2
0
Shreddertheman/Model_CFG_and_Denoise_Comaprisons
false
[]
null
0
1
ds3lab/instructions
false
[ "language:en" ]
An open instructions dataset.
0
0
davanstrien/fuego-20230322-205020-91962a
false
[ "fuego" ]
null
0
0
acul3/instruct_id_chatgpt
false
[ "license:cc-by-4.0" ]
null
2
0
davanstrien/fuego-20230322-205425-d25ee6
false
[ "fuego" ]
null
0
0
davanstrien/ia-loaded
false
[]
null
0
0
MichiganNLP/svo_probes
false
[ "language:en", "license:cc-by-4.0" ]
null
1
1
davanstrien/fuego-20230322-205840-8c6f25
false
[ "fuego" ]
null
0
0
gabeorlanski/bc-tp3
false
[ "license:apache-2.0" ]
null
0
0
davanstrien/fuego-20230322-211033-00ad7c
false
[ "fuego" ]
null
0
0
dvilasuero/somos-alpaca-es
false
[]
null
0
0
davanstrien/fuego-20230322-212050-904d5b
false
[ "fuego" ]
null
0
0
dvilasuero/somos-alpaca-es-rg
false
[]
null
0
0
Norian/Concepts
false
[]
null
0
0
somosnlp/somos-alpaca-es
false
[ "task_categories:text-generation", "size_categories:10K<n<100K", "language:es" ]
null
23
7
davanstrien/ia-loaded2
false
[]
null
0
0
Fathi123/fathi
false
[]
null
0
0
enoreyes/imdb_3000_sphere
false
[ "task_categories:text-classification", "size_categories:1K<n<10K", "language:en", "license:mit" ]
null
0
0
davanstrien/fuego-20230322-220223-7c74cf
false
[ "fuego" ]
null
0
0
cybfi/cyber-2007-unrated
false
[ "license:mit" ]
null
0
0
davanstrien/fuego-20230322-221516-695ee7
false
[ "fuego" ]
null
0
0
aegrif/CIS6930_DAAGR_Empathetic_Dialogues
false
[ "doi:10.57967/hf/0550" ]
null
91
0
pimentooliver/fungifutures2
false
[]
null
0
0
pimentooliver/fungi
false
[]
null
11
0
luciferxf/WikipediaUpdated
false
[]
Wikipedia dataset containing cleaned articles of all languages. The datasets are built from the Wikipedia dump (https://dumps.wikimedia.org/) with one split per language. Each example contains the content of one full Wikipedia article with cleaning to strip markdown and unwanted sections (references, etc.).
2
0
davanstrien/ia-loaded-embedded-gpu
false
[]
null
1
0
CalibraGPT/Fact-Completion
false
[ "task_categories:text-generation", "task_categories:fill-mask", "task_categories:text2text-generation", "language_creators:expert-generated", "language_creators:machine-generated", "multilinguality:multilingual", "size_categories:100K<n<1M", "language:en", "language:fr", "language:es", "language:de", "language:uk", "language:bg", "language:ca", "language:da", "language:hr", "language:hu", "language:it", "language:nl", "language:pl", "language:pt", "language:ro", "language:ru", "language:sl", "language:sr", "language:sv", "language:cs", "license:apache-2.0", "natural-language-understanding", "arxiv:2302.13971", "arxiv:2210.03329", "arxiv:2210.07229" ]
null
257
3
slhenty/climate-fever-nli-stsb
false
[ "license:unknown" ]
A modified CLIMATE-FEVER dataset that includes NLI-style features and STSb-features suitable for SentenceBERT training scripts.
24
1