sha
stringlengths
40
40
text
stringlengths
1
13.4M
id
stringlengths
2
117
tags
sequencelengths
1
7.91k
created_at
stringlengths
25
25
metadata
stringlengths
2
875k
last_modified
stringlengths
25
25
arxiv
sequencelengths
0
25
languages
sequencelengths
0
7.91k
tags_str
stringlengths
17
159k
text_str
stringlengths
1
447k
text_lists
sequencelengths
0
352
processed_texts
sequencelengths
1
353
77cafc116bb191b377b32d2dd153425aa4fc8379
# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Female ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/935?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Female audio data of American English,. It is recorded by American English native speaker, with authentic accent and sweet sound. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/935?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages American English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/American_English_Speech_Synthesis_Corpus-Female
[ "region:us" ]
2022-06-22T06:55:03+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-25T03:23:04+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Female ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Female audio data of American English,. It is recorded by American English native speaker, with authentic accent and sweet sound. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages American English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Female", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nFemale audio data of American English,. It is recorded by American English native speaker, with authentic accent and sweet sound. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Female", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nFemale audio data of American English,. It is recorded by American English native speaker, with authentic accent and sweet sound. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
92d4285f739618e2ff3112ed10d1390680a50a34
# Dataset Card for Nexdata/Scene_Noise_Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/178?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Scene noise data, with a duration of 1,297 hours. The data covers multiple scenarios, including subways, supermarkets, restaurants, roads, etc.; audio is recorded using professional recorders, high sampling rate, dual-channel format collection; time and type of non-noise are annotated. this data set can be used for noise modeling. For more details, please refer to the link: https://www.nexdata.ai/datasets/178?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Noise Data ### Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Scene_Noise_Data
[ "region:us" ]
2022-06-22T06:56:34+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:51:03+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Scene_Noise_Data ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Scene noise data, with a duration of 1,297 hours. The data covers multiple scenarios, including subways, supermarkets, restaurants, roads, etc.; audio is recorded using professional recorders, high sampling rate, dual-channel format collection; time and type of non-noise are annotated. this data set can be used for noise modeling. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Noise Data ### Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Scene_Noise_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nScene noise data, with a duration of 1,297 hours. The data covers multiple scenarios, including subways, supermarkets, restaurants, roads, etc.; audio is recorded using professional recorders, high sampling rate, dual-channel format collection; time and type of non-noise are annotated. this data set can be used for noise modeling.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nNoise Data", "### Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Scene_Noise_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nScene noise data, with a duration of 1,297 hours. The data covers multiple scenarios, including subways, supermarkets, restaurants, roads, etc.; audio is recorded using professional recorders, high sampling rate, dual-channel format collection; time and type of non-noise are annotated. this data set can be used for noise modeling.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nNoise Data", "### Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
418bcc31ec41f7eb730b8795e11b85d7af182194
# Dataset Card for Nexdata/English_Emotional_Speech_Data_by_Microphone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/179?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary English emotional audio data captured by microphone, 20 American native speakers participate in the recording, 2,100 sentences per person; the recorded script covers 10 emotions such as anger, happiness, sadness; the voice is recorded by high-fidelity microphone therefore has high quality; it is used for analytical detection of emotional speech. For more details, please refer to the link: https://www.nexdata.ai/datasets/179?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, emotion-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/English_Emotional_Speech_Data_by_Microphone
[ "region:us" ]
2022-06-22T06:57:59+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:54:34+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/English_Emotional_Speech_Data_by_Microphone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary English emotional audio data captured by microphone, 20 American native speakers participate in the recording, 2,100 sentences per person; the recorded script covers 10 emotions such as anger, happiness, sadness; the voice is recorded by high-fidelity microphone therefore has high quality; it is used for analytical detection of emotional speech. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, emotion-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/English_Emotional_Speech_Data_by_Microphone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nEnglish emotional audio data captured by microphone, 20 American native speakers participate in the recording, 2,100 sentences per person; the recorded script covers 10 emotions such as anger, happiness, sadness; the voice is recorded by high-fidelity microphone therefore has high quality; it is used for analytical detection of emotional speech.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, emotion-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/English_Emotional_Speech_Data_by_Microphone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nEnglish emotional audio data captured by microphone, 20 American native speakers participate in the recording, 2,100 sentences per person; the recorded script covers 10 emotions such as anger, happiness, sadness; the voice is recorded by high-fidelity microphone therefore has high quality; it is used for analytical detection of emotional speech.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, emotion-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
54d1322eecde28608b0cbbd270876f0de5942827
# Dataset Card for Nexdata/Mandarin_Speech_Data_in_Noisy_Environment_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/191?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Spoken Mandarin audio data under noisy environment captured by mobile phone, it is recorded by 203 speakers from all over China, covering all major dialect regions; and a variety of noise scenes such as subways, supermarkets, restaurants, etc., more suitable for real application scenes; it can be used for automatic speech recognition, machine translation, and voiceprint recognition. For more details, please refer to the link: https://www.nexdata.ai/datasets/191?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mandarin_Speech_Data_in_Noisy_Environment_by_Mobile_Phone
[ "region:us" ]
2022-06-22T06:59:26+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:54:09+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mandarin_Speech_Data_in_Noisy_Environment_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Spoken Mandarin audio data under noisy environment captured by mobile phone, it is recorded by 203 speakers from all over China, covering all major dialect regions; and a variety of noise scenes such as subways, supermarkets, restaurants, etc., more suitable for real application scenes; it can be used for automatic speech recognition, machine translation, and voiceprint recognition. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mandarin_Speech_Data_in_Noisy_Environment_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nSpoken Mandarin audio data under noisy environment captured by mobile phone, it is recorded by 203 speakers from all over China, covering all major dialect regions; and a variety of noise scenes such as subways, supermarkets, restaurants, etc., more suitable for real application scenes; it can be used for automatic speech recognition, machine translation, and voiceprint recognition.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mandarin_Speech_Data_in_Noisy_Environment_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nSpoken Mandarin audio data under noisy environment captured by mobile phone, it is recorded by 203 speakers from all over China, covering all major dialect regions; and a variety of noise scenes such as subways, supermarkets, restaurants, etc., more suitable for real application scenes; it can be used for automatic speech recognition, machine translation, and voiceprint recognition.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
ff2777fc92a7a6a292e2d6e5f381481e90570cb6
# Dataset Card for Nexdata/German_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/949?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary German audio data captured by mobile phone, 1,796 hours in total, recorded by 3,442 German native speakers. The recorded text is designed by linguistic experts, covering generic, interactive, on-board, home and other categories. The text has been proofread manually with high accuracy; this data can be used for automatic speech recognition, machine translation, and voiceprint recognition. For more details, please refer to the link: https://www.nexdata.ai/datasets/949?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages German ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/German_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:00:38+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:51:26+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/German_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary German audio data captured by mobile phone, 1,796 hours in total, recorded by 3,442 German native speakers. The recorded text is designed by linguistic experts, covering generic, interactive, on-board, home and other categories. The text has been proofread manually with high accuracy; this data can be used for automatic speech recognition, machine translation, and voiceprint recognition. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages German ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/German_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nGerman audio data captured by mobile phone, 1,796 hours in total, recorded by 3,442 German native speakers. The recorded text is designed by linguistic experts, covering generic, interactive, on-board, home and other categories. The text has been proofread manually with high accuracy; this data can be used for automatic speech recognition, machine translation, and voiceprint recognition.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nGerman", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/German_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nGerman audio data captured by mobile phone, 1,796 hours in total, recorded by 3,442 German native speakers. The recorded text is designed by linguistic experts, covering generic, interactive, on-board, home and other categories. The text has been proofread manually with high accuracy; this data can be used for automatic speech recognition, machine translation, and voiceprint recognition.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nGerman", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
57bebc3e965385c1845953f16e62e3f55ea9d7f2
# Dataset Card for Nexdata/Far-filed_Noise_Speech_Data_in_Home_Environment_by_Mic-Array ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/255?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The data consists of multiple sets of products, each with a different type of microphone arrays. Noise data is collected from real home scenes of the indoor residence of ordinary residents. The data set can be used for tasks such as voice enhancement and automatic speech recognition in a home scene. For more details, please refer to the link: https://www.nexdata.ai/datasets/255?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Noise Data ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Far-filed_Noise_Speech_Data_in_Home_Environment_by_Mic-Array
[ "region:us" ]
2022-06-22T07:03:31+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:53:41+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Far-filed_Noise_Speech_Data_in_Home_Environment_by_Mic-Array ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary The data consists of multiple sets of products, each with a different type of microphone arrays. Noise data is collected from real home scenes of the indoor residence of ordinary residents. The data set can be used for tasks such as voice enhancement and automatic speech recognition in a home scene. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Noise Data ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Far-filed_Noise_Speech_Data_in_Home_Environment_by_Mic-Array", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe data consists of multiple sets of products, each with a different type of microphone arrays. Noise data is collected from real home scenes of the indoor residence of ordinary residents. The data set can be used for tasks such as voice enhancement and automatic speech recognition in a home scene.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nNoise Data", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Far-filed_Noise_Speech_Data_in_Home_Environment_by_Mic-Array", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe data consists of multiple sets of products, each with a different type of microphone arrays. Noise data is collected from real home scenes of the indoor residence of ordinary residents. The data set can be used for tasks such as voice enhancement and automatic speech recognition in a home scene.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition,noisy-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nNoise Data", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
f3ae902e93f1e09d642f02fa8dd4c49ab0d292da
# Dataset Card for Nexdata/Japanese_Speech_Datae ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/934?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1006 Japanese native speakers participated in the recording, coming from eastern, western, and Kyushu regions, while the eastern region accounting for the largest proportion. The recording content is rich and all texts have been manually transferred with high accuracy. For more details, please refer to the link: https://www.nexdata.ai/datasets/934?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Japanese ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Japanese_Speech_Data
[ "region:us" ]
2022-06-22T07:05:33+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:57:36+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Japanese_Speech_Datae ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1006 Japanese native speakers participated in the recording, coming from eastern, western, and Kyushu regions, while the eastern region accounting for the largest proportion. The recording content is rich and all texts have been manually transferred with high accuracy. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Japanese ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Japanese_Speech_Datae", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1006 Japanese native speakers participated in the recording, coming from eastern, western, and Kyushu regions, while the eastern region accounting for the largest proportion. The recording content is rich and all texts have been manually transferred with high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nJapanese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Japanese_Speech_Datae", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1006 Japanese native speakers participated in the recording, coming from eastern, western, and Kyushu regions, while the eastern region accounting for the largest proportion. The recording content is rich and all texts have been manually transferred with high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nJapanese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
fbf40b235992c6e772f01b6fe5e555208143f205
# Dataset Card for Nexdata/American_English_Spontaneous_Speech_Data ## Description The 1,178-hour American English Spontaneous Speech Data is a collection of speech clips, covering multiple topics. Audio is transcribed into text, and speaker identity and other more attributes are annotated. This data set can be used for voiceprint recognition model training, construction of corpus for machine translation and algorithm research. For more details, please refer to the link: https://www.nexdata.ai/datasets/1115?source=Huggingface # Specifications ## Format 16kHz, 16bit, mono channel; ## Content category including self-media, conversation, live, lecture, variety-show. ## Language American English; ## Annotation annotation for the transcription text, speaker identification, gender; ## Application scenarios speech recognition, video caption generation and video content review; ## Accuracy at a Sentence Accuracy Rate (SAR) of being no less than 95%. # Licensing Information Commercial License
Nexdata/American_English_Spontaneous_Speech_Data
[ "task_categories:automatic-speech-recognition", "language:en", "region:us" ]
2022-06-22T07:07:16+00:00
{"language": ["en"], "task_categories": ["automatic-speech-recognition"]}
2023-11-10T07:29:24+00:00
[]
[ "en" ]
TAGS #task_categories-automatic-speech-recognition #language-English #region-us
# Dataset Card for Nexdata/American_English_Spontaneous_Speech_Data ## Description The 1,178-hour American English Spontaneous Speech Data is a collection of speech clips, covering multiple topics. Audio is transcribed into text, and speaker identity and other more attributes are annotated. This data set can be used for voiceprint recognition model training, construction of corpus for machine translation and algorithm research. For more details, please refer to the link: URL # Specifications ## Format 16kHz, 16bit, mono channel; ## Content category including self-media, conversation, live, lecture, variety-show. ## Language American English; ## Annotation annotation for the transcription text, speaker identification, gender; ## Application scenarios speech recognition, video caption generation and video content review; ## Accuracy at a Sentence Accuracy Rate (SAR) of being no less than 95%. # Licensing Information Commercial License
[ "# Dataset Card for Nexdata/American_English_Spontaneous_Speech_Data", "## Description\nThe 1,178-hour American English Spontaneous Speech Data is a collection of speech clips, covering multiple topics. Audio is transcribed into text, and speaker identity and other more attributes are annotated. This data set can be used for voiceprint recognition model training, construction of corpus for machine translation and algorithm research.\n\nFor more details, please refer to the link: URL", "# Specifications", "## Format\n16kHz, 16bit, mono channel;", "## Content category\nincluding self-media, conversation, live, lecture, variety-show.", "## Language\nAmerican English;", "## Annotation\nannotation for the transcription text, speaker identification, gender;", "## Application scenarios\nspeech recognition, video caption generation and video content review;", "## Accuracy\nat a Sentence Accuracy Rate (SAR) of being no less than 95%.", "# Licensing Information\nCommercial License" ]
[ "TAGS\n#task_categories-automatic-speech-recognition #language-English #region-us \n", "# Dataset Card for Nexdata/American_English_Spontaneous_Speech_Data", "## Description\nThe 1,178-hour American English Spontaneous Speech Data is a collection of speech clips, covering multiple topics. Audio is transcribed into text, and speaker identity and other more attributes are annotated. This data set can be used for voiceprint recognition model training, construction of corpus for machine translation and algorithm research.\n\nFor more details, please refer to the link: URL", "# Specifications", "## Format\n16kHz, 16bit, mono channel;", "## Content category\nincluding self-media, conversation, live, lecture, variety-show.", "## Language\nAmerican English;", "## Annotation\nannotation for the transcription text, speaker identification, gender;", "## Application scenarios\nspeech recognition, video caption generation and video content review;", "## Accuracy\nat a Sentence Accuracy Rate (SAR) of being no less than 95%.", "# Licensing Information\nCommercial License" ]
41b08241e6492a98f87196babed9828ab7fc2e5f
# Dataset Card for Nexdata/Non-Hispanic_Spanish_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/970?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1,630 non-Spanish nationality native Spanish speakers such as Mexicans and Colombians participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, in-vehicle and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: https://www.nexdata.ai/datasets/970?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Spanish ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Non-Hispanic_Spanish_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:08:37+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:59:27+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Non-Hispanic_Spanish_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1,630 non-Spanish nationality native Spanish speakers such as Mexicans and Colombians participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, in-vehicle and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Spanish ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Non-Hispanic_Spanish_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1,630 non-Spanish nationality native Spanish speakers such as Mexicans and Colombians participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, in-vehicle and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nSpanish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Non-Hispanic_Spanish_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1,630 non-Spanish nationality native Spanish speakers such as Mexicans and Colombians participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, in-vehicle and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nSpanish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
76a1238a2fff1d5b8a5cfd16a49a3022fecba103
# Dataset Card for Nexdata/Emotional_Video_Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/977?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1,003 People - Emotional Video Data. The data diversity includes multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions). For each sentence in each video, emotion types (including facial emotions and inner emotions), start & end time, and text transcription were annotated.This dataset can be used for tasks such as emotion recognition and sentiment analysis. For more details, please refer to the link: https://www.nexdata.ai/datasets/977?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, sentiment-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages English, Chinese ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Emotional_Video_Data
[ "region:us" ]
2022-06-22T07:09:51+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:59:50+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Emotional_Video_Data ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1,003 People - Emotional Video Data. The data diversity includes multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions). For each sentence in each video, emotion types (including facial emotions and inner emotions), start & end time, and text transcription were annotated.This dataset can be used for tasks such as emotion recognition and sentiment analysis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification, sentiment-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages English, Chinese ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Emotional_Video_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1,003 People - Emotional Video Data. The data diversity includes multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions). For each sentence in each video, emotion types (including facial emotions and inner emotions), start & end time, and text transcription were annotated.This dataset can be used for tasks such as emotion recognition and sentiment analysis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, sentiment-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nEnglish, Chinese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Emotional_Video_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1,003 People - Emotional Video Data. The data diversity includes multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions). For each sentence in each video, emotion types (including facial emotions and inner emotions), start & end time, and text transcription were annotated.This dataset can be used for tasks such as emotion recognition and sentiment analysis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification, sentiment-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nEnglish, Chinese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
2fcd1dac9d2caad93f62e91869d1f59d42b71081
# Dataset Card for Nexdata/Mandarin_Interactive_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/981?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Mandarin home interaction mobile phone language audio data (Far-field home collected audio data subset), with duration of 849 hours, recorded in the real home scene; content focuses on home instructions, functional assistants and wake-up words, specially designed for smart home, more close to data application scenes. For more details, please refer to the link: https://www.nexdata.ai/datasets/981?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mandarin_Interactive_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:13:02+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:57:12+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mandarin_Interactive_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Mandarin home interaction mobile phone language audio data (Far-field home collected audio data subset), with duration of 849 hours, recorded in the real home scene; content focuses on home instructions, functional assistants and wake-up words, specially designed for smart home, more close to data application scenes. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mandarin_Interactive_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMandarin home interaction mobile phone language audio data (Far-field home collected audio data subset), with duration of 849 hours, recorded in the real home scene; content focuses on home instructions, functional assistants and wake-up words, specially designed for smart home, more close to data application scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mandarin_Interactive_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMandarin home interaction mobile phone language audio data (Far-field home collected audio data subset), with duration of 849 hours, recorded in the real home scene; content focuses on home instructions, functional assistants and wake-up words, specially designed for smart home, more close to data application scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
4cfcbc7f1d04f7d40ca1796ca15a3ee82db61cbb
# Dataset Card for Nexdata/French_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/989?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1089 French native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: https://www.nexdata.ai/datasets/989?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages French English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/French_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:14:22+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-31T01:31:13+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/French_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1089 French native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages French English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/French_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1089 French native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nFrench English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/French_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1089 French native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nFrench English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
c74c84bb5b95b99ce6e8f6e6c3e83ab76101c6cb
# Dataset Card for Nexdata/Spanish_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/990?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 891 Spanish native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: https://www.nexdata.ai/datasets/990?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Spanish English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Spanish_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:16:02+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:56:48+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Spanish_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 891 Spanish native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Spanish English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Spanish_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n891 Spanish native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSpanish English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Spanish_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n891 Spanish native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSpanish English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
ccd1bb12ac2b0a445d722369788e599ab7909b74
# Dataset Card for Nexdata/Singaporean_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1045?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset is recorded by 452 native Singaporean speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy. For more details, please refer to the link: https://www.nexdata.ai/datasets/1045?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Singaporean English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Singaporean_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:17:36+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-25T03:17:28+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Singaporean_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary This dataset is recorded by 452 native Singaporean speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Singaporean English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Singaporean_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis dataset is recorded by 452 native Singaporean speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSingaporean English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Singaporean_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis dataset is recorded by 452 native Singaporean speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSingaporean English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
aa5bfb4c655749760616b0ce09ba96cf17d9e2b3
# Dataset Card for Nexdata/Australian_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1046?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary It collects 2.956 speakers from Shanghai and is recorded in quiet indoor environment. The recorded content includes multi-domain customer consultation, short messages, numbers, Shanghai POI, etc. The corpus has no repetition and the average sentence length is 12.68 words. Recording devices are mainstream Android phones and iPhones. For more details, please refer to the link: https://www.nexdata.ai/datasets/1046?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Australian English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Australian_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:19:25+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-25T03:19:47+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Australian_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary It collects 2.956 speakers from Shanghai and is recorded in quiet indoor environment. The recorded content includes multi-domain customer consultation, short messages, numbers, Shanghai POI, etc. The corpus has no repetition and the average sentence length is 12.68 words. Recording devices are mainstream Android phones and iPhones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Australian English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Australian_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nIt collects 2.956 speakers from Shanghai and is recorded in quiet indoor environment. The recorded content includes multi-domain customer consultation, short messages, numbers, Shanghai POI, etc. The corpus has no repetition and the average sentence length is 12.68 words. Recording devices are mainstream Android phones and iPhones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nAustralian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Australian_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nIt collects 2.956 speakers from Shanghai and is recorded in quiet indoor environment. The recorded content includes multi-domain customer consultation, short messages, numbers, Shanghai POI, etc. The corpus has no repetition and the average sentence length is 12.68 words. Recording devices are mainstream Android phones and iPhones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nAustralian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
9a20c993a0eab537b9c0d40fb34d7d8dc7d13ae9
# Dataset Card for Nexdata/Indonesian_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/991?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1285 Indonesian native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: https://www.nexdata.ai/datasets/991?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Indonesian ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Indonesian_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:20:41+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:58:32+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Indonesian_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1285 Indonesian native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Indonesian ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Indonesian_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1285 Indonesian native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nIndonesian", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Indonesian_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1285 Indonesian native speakers participated in the recording with authentic accent. The recorded script is designed by linguists and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. The data set can be applied for automatic speech recognition, and machine translation scenes.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nIndonesian", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
6fbbd53218b7aaf4ed7563b6f61cb0f8b37eb4ad
# 201-People-Infant-Cry-Speech-Data-by-Mobile-Phone ## Description Crying sound of 201 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's crying sound in smart home projects. For more details, please refer to the link: https://www.nexdata.ai/datasets/998?source=Huggingface ## Format 16kHz, 16bit, uncompressed wav, mono channel ## Recording Environment relatively quiet indoor environment, without echo ## Recording Content infant cry ## Population 201 people; 105 boys, 96 girls; 132 people aged 0~1, 58 people aged 1~2, 11 people aged 2~3 ## Device iPhone, Android mobile phone ## Application scene abnormal voice recognition,smart home # Licensing Information Commercial License
Nexdata/Infant_Cry_Speech_Data_by_Mobile_Phone
[ "task_categories:automatic-speech-recognition", "task_categories:voice-activity-detection", "region:us" ]
2022-06-22T07:21:57+00:00
{"task_categories": ["automatic-speech-recognition", "voice-activity-detection"], "YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2024-01-26T09:44:10+00:00
[]
[]
TAGS #task_categories-automatic-speech-recognition #task_categories-voice-activity-detection #region-us
# 201-People-Infant-Cry-Speech-Data-by-Mobile-Phone ## Description Crying sound of 201 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's crying sound in smart home projects. For more details, please refer to the link: URL ## Format 16kHz, 16bit, uncompressed wav, mono channel ## Recording Environment relatively quiet indoor environment, without echo ## Recording Content infant cry ## Population 201 people; 105 boys, 96 girls; 132 people aged 0~1, 58 people aged 1~2, 11 people aged 2~3 ## Device iPhone, Android mobile phone ## Application scene abnormal voice recognition,smart home # Licensing Information Commercial License
[ "# 201-People-Infant-Cry-Speech-Data-by-Mobile-Phone", "## Description\nCrying sound of 201 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's crying sound in smart home projects.\n\nFor more details, please refer to the link: URL", "## Format\n16kHz, 16bit, uncompressed wav, mono channel", "## Recording Environment\nrelatively quiet indoor environment, without echo", "## Recording Content\ninfant cry", "## Population\n201 people; 105 boys, 96 girls; 132 people aged 0~1, 58 people aged 1~2, 11 people aged 2~3", "## Device\niPhone, Android mobile phone", "## Application scene\nabnormal voice recognition,smart home", "# Licensing Information\nCommercial License" ]
[ "TAGS\n#task_categories-automatic-speech-recognition #task_categories-voice-activity-detection #region-us \n", "# 201-People-Infant-Cry-Speech-Data-by-Mobile-Phone", "## Description\nCrying sound of 201 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's crying sound in smart home projects.\n\nFor more details, please refer to the link: URL", "## Format\n16kHz, 16bit, uncompressed wav, mono channel", "## Recording Environment\nrelatively quiet indoor environment, without echo", "## Recording Content\ninfant cry", "## Population\n201 people; 105 boys, 96 girls; 132 people aged 0~1, 58 people aged 1~2, 11 people aged 2~3", "## Device\niPhone, Android mobile phone", "## Application scene\nabnormal voice recognition,smart home", "# Licensing Information\nCommercial License" ]
59f4bca87b641d9fa74fb99832146f24c302e5e3
# Dataset Card for Nexdata/American_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/999?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1842 American native speakers participated in the recording with authentic accent. The recorded script is designed by linguists, based on scenes, and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: https://www.nexdata.ai/datasets/999?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages American English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/American_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:23:17+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:55:49+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/American_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1842 American native speakers participated in the recording with authentic accent. The recorded script is designed by linguists, based on scenes, and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages American English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/American_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1842 American native speakers participated in the recording with authentic accent. The recorded script is designed by linguists, based on scenes, and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/American_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1842 American native speakers participated in the recording with authentic accent. The recorded script is designed by linguists, based on scenes, and cover a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
7c4fc4bde54b57c54d736f748fc22178925578b8
# Dataset Card for Nexdata/Chinese_Children_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1001?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Children read English audio data, covering ages from preschool (3-5 years old) to post-school (6-12 years old) , with children's speech features; content accurately matches children's actual scenes of speaking English. It provides data support for children's smart home, automatic speech recognition and oral assessment in intelligent education scene, . For more details, please refer to the link: https://www.nexdata.ai/datasets/1001?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Children_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:24:33+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-24T09:37:42+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Children_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Children read English audio data, covering ages from preschool (3-5 years old) to post-school (6-12 years old) , with children's speech features; content accurately matches children's actual scenes of speaking English. It provides data support for children's smart home, automatic speech recognition and oral assessment in intelligent education scene, . For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Children_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nChildren read English audio data, covering ages from preschool (3-5 years old) to post-school (6-12 years old) , with children's speech features; content accurately matches children's actual scenes of speaking English. It provides data support for children's smart home, automatic speech recognition and oral assessment in intelligent education scene, .\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Children_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nChildren read English audio data, covering ages from preschool (3-5 years old) to post-school (6-12 years old) , with children's speech features; content accurately matches children's actual scenes of speaking English. It provides data support for children's smart home, automatic speech recognition and oral assessment in intelligent education scene, .\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
0910b09a0850ed7ecc8a65035b86647a8b622345
# Dataset Card for Nexdata/Mandarin_Strong_Accent_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1003?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary More than 2,000 Chinese native speakers participated in the recording with equal gender. Speakers are mainly from the southern China, and some of them are from the provinces of northern China with Strong accents. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers and other fields, which is accurately matching the smart home, intelligent car and other practical application scenarios. For more details, please refer to the link: https://www.nexdata.ai/datasets/1003?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mandarin_Strong_Accent_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:25:45+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:58:10+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mandarin_Strong_Accent_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary More than 2,000 Chinese native speakers participated in the recording with equal gender. Speakers are mainly from the southern China, and some of them are from the provinces of northern China with Strong accents. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers and other fields, which is accurately matching the smart home, intelligent car and other practical application scenarios. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mandarin_Strong_Accent_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMore than 2,000 Chinese native speakers participated in the recording with equal gender. Speakers are mainly from the southern China, and some of them are from the provinces of northern China with Strong accents. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers and other fields, which is accurately matching the smart home, intelligent car and other practical application scenarios.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mandarin_Strong_Accent_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMore than 2,000 Chinese native speakers participated in the recording with equal gender. Speakers are mainly from the southern China, and some of them are from the provinces of northern China with Strong accents. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers and other fields, which is accurately matching the smart home, intelligent car and other practical application scenarios.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
e6847b0e86454869eff6262ef1f9c1ce0d47acab
# Dataset Card for Nexdata/Korean_Speech_Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1008?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Korean audio data with duration of 516 hours, recorded texts include: daily language, various interactive sentences, home commands, on-board commands, etc. Among 1,077 speakers, male and female speakers are 49% and 51%. The duration of each speaker is around half an hour. For more details, please refer to the link: https://www.nexdata.ai/datasets/1008?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Korean ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Korean_Speech_Data
[ "region:us" ]
2022-06-22T07:27:21+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:29:18+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Korean_Speech_Data ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Korean audio data with duration of 516 hours, recorded texts include: daily language, various interactive sentences, home commands, on-board commands, etc. Among 1,077 speakers, male and female speakers are 49% and 51%. The duration of each speaker is around half an hour. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Korean ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Korean_Speech_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nKorean audio data with duration of 516 hours, recorded texts include: daily language, various interactive sentences, home commands, on-board commands, etc. Among 1,077 speakers, male and female speakers are 49% and 51%. The duration of each speaker is around half an hour.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nKorean", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Korean_Speech_Data", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nKorean audio data with duration of 516 hours, recorded texts include: daily language, various interactive sentences, home commands, on-board commands, etc. Among 1,077 speakers, male and female speakers are 49% and 51%. The duration of each speaker is around half an hour.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nKorean", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
178e835a2a49cc6aa511cb68ee4560f13bb50e55
# Dataset Card for Nexdata/Mandarin_Voiceprint_Recognition_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1009?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Each person's time span is very long, which can better cover the sound features of a person in different periods and different states. For more details, please refer to the link: https://www.nexdata.ai/datasets/1009?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mandarin_Voiceprint_Recognition_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:28:38+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:32:14+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mandarin_Voiceprint_Recognition_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Each person's time span is very long, which can better cover the sound features of a person in different periods and different states. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mandarin_Voiceprint_Recognition_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nEach person's time span is very long, which can better cover the sound features of a person in different periods and different states.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mandarin_Voiceprint_Recognition_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nEach person's time span is very long, which can better cover the sound features of a person in different periods and different states.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
2c5ee81bdc0bdb98eeef43427bf7f06783f02348
# Dataset Card for Nexdata/Italian_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1022?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 497 Italians recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: https://www.nexdata.ai/datasets/1022?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Italian English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Italian_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:30:05+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:28:55+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Italian_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 497 Italians recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Italian English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Italian_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n497 Italians recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nItalian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Italian_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n497 Italians recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nItalian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
6f2cdc8d30cdf2dfd64f5c8cf019ac5123b20764
# Dataset Card for Nexdata/Portuguese_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1023?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 532 Portuguese recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: https://www.nexdata.ai/datasets/1023?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Portuguese English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Portuguese_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:32:21+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:31:51+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Portuguese_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 532 Portuguese recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Portuguese English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Portuguese_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n532 Portuguese recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nPortuguese English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Portuguese_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n532 Portuguese recorded in a relatively quiet environment in authentic English. The recorded script is designed by linguists and covers a wide range of topics including generic, interactive, on-board and home. The text is manually proofread with high accuracy. It matches with mainstream Android and Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nPortuguese English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
1488a05544980c6b607089dcf54d54642ca647dc
# Dataset Card for Nexdata/Korean_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1041?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The product contains the speech data recorded by 400 native Korean speakers, with roughly equal gender distribution. The corpus covers a wide domain with rich content of generic category, human-machine interaction category, in-car category, smart home category, etc. The corpus text was manually checked to ensure the high accuracy. For more details, please refer to the link: https://www.nexdata.ai/datasets/1041?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Korean English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Korean_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:33:45+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-25T03:16:15+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Korean_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary The product contains the speech data recorded by 400 native Korean speakers, with roughly equal gender distribution. The corpus covers a wide domain with rich content of generic category, human-machine interaction category, in-car category, smart home category, etc. The corpus text was manually checked to ensure the high accuracy. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Korean English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Korean_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe product contains the speech data recorded by 400 native Korean speakers, with roughly equal gender distribution. The corpus covers a wide domain with rich content of generic category, human-machine interaction category, in-car category, smart home category, etc. The corpus text was manually checked to ensure the high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nKorean English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Korean_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe product contains the speech data recorded by 400 native Korean speakers, with roughly equal gender distribution. The corpus covers a wide domain with rich content of generic category, human-machine interaction category, in-car category, smart home category, etc. The corpus text was manually checked to ensure the high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nKorean English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
5d81e703f0ebabf303f53f5be6ed7af7fa6dd4bc
# Dataset Card for Nexdata/Russian_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1042?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset is recorded by 498 native Russian speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy. For more details, please refer to the link: https://www.nexdata.ai/datasets/1042?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Russian English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Russian_Speaking_English_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:49:02+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-24T09:36:43+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Russian_Speaking_English_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary This dataset is recorded by 498 native Russian speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Russian English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Russian_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis dataset is recorded by 498 native Russian speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nRussian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Russian_Speaking_English_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis dataset is recorded by 498 native Russian speakers with a balanced gender. It is rich in content and it covers generic command and control;human-machine interaction; smart home command and control;in-car command and control categories. The transcription corpus has been manually proofread to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nRussian English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
160da285bc293c4a7b138337af50be4931fca0a5
# Dataset Card for Nexdata/Sichuan_Dialect_Conversational_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1065?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 1730 Sichuan native speakers participated in the recording and face-to-face free talking in a natural way in wide fields without the topic specified. It is natural and fluency in speech, and in line with the actual dialogue scene. We transcribed the speech into text manually to ensure high accuracy. For more details, please refer to the link: https://www.nexdata.ai/datasets/1065?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Sichuan Dialect ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Sichuan_Dialect_Conversational_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:50:19+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-25T03:20:54+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Sichuan_Dialect_Conversational_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 1730 Sichuan native speakers participated in the recording and face-to-face free talking in a natural way in wide fields without the topic specified. It is natural and fluency in speech, and in line with the actual dialogue scene. We transcribed the speech into text manually to ensure high accuracy. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Sichuan Dialect ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Sichuan_Dialect_Conversational_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1730 Sichuan native speakers participated in the recording and face-to-face free talking in a natural way in wide fields without the topic specified. It is natural and fluency in speech, and in line with the actual dialogue scene. We transcribed the speech into text manually to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSichuan Dialect", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Sichuan_Dialect_Conversational_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n1730 Sichuan native speakers participated in the recording and face-to-face free talking in a natural way in wide fields without the topic specified. It is natural and fluency in speech, and in line with the actual dialogue scene. We transcribed the speech into text manually to ensure high accuracy.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nSichuan Dialect", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
20072a3a2cb5f6351e9b196666746510f8e39976
# Dataset Card for Nexdata/Chines_Digital_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1072?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 11,010 People - Chines Digital Speech Data by Mobile Phone was recorded by 11,010 voice recorders in Mandarin. It's collected from each person 30 sentences of 4-8 digits. For more details, please refer to the link: https://www.nexdata.ai/datasets/1072?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Digital_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:51:32+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:30:48+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chines_Digital_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 11,010 People - Chines Digital Speech Data by Mobile Phone was recorded by 11,010 voice recorders in Mandarin. It's collected from each person 30 sentences of 4-8 digits. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chines_Digital_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n11,010 People - Chines Digital Speech Data by Mobile Phone was recorded by 11,010 voice recorders in Mandarin. It's collected from each person 30 sentences of 4-8 digits.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chines_Digital_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n11,010 People - Chines Digital Speech Data by Mobile Phone was recorded by 11,010 voice recorders in Mandarin. It's collected from each person 30 sentences of 4-8 digits.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
1270f1b5b9a8c20714acf5ec93a91260788d3285
# Dataset Card for Nexdata/Wake-up_Words_Speech_Data_by_Microphone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1076?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary More than 1,000 recorders read the specified wake-up words, covering slow, normal, and fast three speeds. Audios are recorded in the professional recording studio using the microphone. For more details, please refer to the link: https://www.nexdata.ai/datasets/1076?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Wake-up_Words_Speech_Data_by_Microphone
[ "region:us" ]
2022-06-22T07:55:22+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:28:28+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Wake-up_Words_Speech_Data_by_Microphone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary More than 1,000 recorders read the specified wake-up words, covering slow, normal, and fast three speeds. Audios are recorded in the professional recording studio using the microphone. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Wake-up_Words_Speech_Data_by_Microphone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMore than 1,000 recorders read the specified wake-up words, covering slow, normal, and fast three speeds. Audios are recorded in the professional recording studio using the microphone.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Wake-up_Words_Speech_Data_by_Microphone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMore than 1,000 recorders read the specified wake-up words, covering slow, normal, and fast three speeds. Audios are recorded in the professional recording studio using the microphone.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
792aae595a83a121e7945f7cc8d33fce4222ec92
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655888220
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T07:57:00+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T07:57:03+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
1a5cc7b7ff302eb102ec7698fbd31184f45c64d2
# Dataset Card for Nexdata/Mixed_Speech_with_Chinese_and_English_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1080?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The data is recorded by 1113 Chinese native speakers with accents covering seven major dialect areas. The recorded text is a mixture of Chinese and English sentences, covering general scenes and human-computer interaction scenes. It is rich in content and accurate in transcription. It can be used for improving the recognition effect of the speech recognition system on Chinese-English mixed reading speech. For more details, please refer to the link: https://www.nexdata.ai/datasets/1080?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese, English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mixed_Speech_with_Chinese_and_English_Data
[ "region:us" ]
2022-06-22T07:57:18+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:06:07+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mixed_Speech_with_Chinese_and_English_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary The data is recorded by 1113 Chinese native speakers with accents covering seven major dialect areas. The recorded text is a mixture of Chinese and English sentences, covering general scenes and human-computer interaction scenes. It is rich in content and accurate in transcription. It can be used for improving the recognition effect of the speech recognition system on Chinese-English mixed reading speech. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese, English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mixed_Speech_with_Chinese_and_English_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe data is recorded by 1113 Chinese native speakers with accents covering seven major dialect areas. The recorded text is a mixture of Chinese and English sentences, covering general scenes and human-computer interaction scenes. It is rich in content and accurate in transcription. It can be used for improving the recognition effect of the speech recognition system on Chinese-English mixed reading speech.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese, English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mixed_Speech_with_Chinese_and_English_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe data is recorded by 1113 Chinese native speakers with accents covering seven major dialect areas. The recorded text is a mixture of Chinese and English sentences, covering general scenes and human-computer interaction scenes. It is rich in content and accurate in transcription. It can be used for improving the recognition effect of the speech recognition system on Chinese-English mixed reading speech.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese, English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
ff441ddfaf2dae6b03bfb4d31947b18ae7324443
# Dataset Card for Nexdata/Mandarin_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1081?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 4,787 Chinese native speakers participated in the recording with equal gender. Speakers are from various provinces of China. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers, and other fields, which is accurately matching the smart home, intelligent car, and other practical application scenarios. For more details, please refer to the link: https://www.nexdata.ai/datasets/1081?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Mandarin_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:58:40+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:15:29+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Mandarin_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 4,787 Chinese native speakers participated in the recording with equal gender. Speakers are from various provinces of China. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers, and other fields, which is accurately matching the smart home, intelligent car, and other practical application scenarios. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Mandarin_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n4,787 Chinese native speakers participated in the recording with equal gender. Speakers are from various provinces of China. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers, and other fields, which is accurately matching the smart home, intelligent car, and other practical application scenarios.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Mandarin_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n4,787 Chinese native speakers participated in the recording with equal gender. Speakers are from various provinces of China. The recording content is rich, covering mobile phone voice assistant interaction, smart home command and control, In-car command and control, numbers, and other fields, which is accurately matching the smart home, intelligent car, and other practical application scenarios.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
476f51b81ea43a5fc716d515ce3347fd2d98c50b
# Dataset Card for Nexdata/Infant_Laugh_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1090?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Laugh sound of 20 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's laugh sound in smart home projects. For more details, please refer to the link: https://www.nexdata.ai/datasets/1090?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Infant Cry ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Infant_Laugh_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T07:59:55+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:14:57+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Infant_Laugh_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Laugh sound of 20 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's laugh sound in smart home projects. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Infant Cry ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Infant_Laugh_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nLaugh sound of 20 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's laugh sound in smart home projects.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nInfant Cry", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Infant_Laugh_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nLaugh sound of 20 infants and young children aged 0~3 years old, a number of paragraphs from each of them; It provides data support for detecting children's laugh sound in smart home projects.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\nInfant Cry", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
cf60d22dab38728a9331532bdae04e03649bd357
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Data_Female_Customer_Service ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1098?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 26.1 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, It is recorded by Chinese native speakers, with lively and frindly voice. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1098?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Synthesis_Data_Female_Customer_Service
[ "region:us" ]
2022-06-22T08:02:00+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-30T09:51:48+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Data_Female_Customer_Service ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 26.1 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, It is recorded by Chinese native speakers, with lively and frindly voice. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Data_Female_Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n26.1 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, It is recorded by Chinese native speakers, with lively and frindly voice. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Data_Female_Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n26.1 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, It is recorded by Chinese native speakers, with lively and frindly voice. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
729e0463bdba8f54c61d271e50c519aec3334909
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Male_Customer_Service ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1099?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 20 Hours - Chinese Mandarin Synthesis Corpus-Male, Customer Service. It is recorded by Chinese native speakers, the voice of the full of magnetism. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1099?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Synthesis_Corpus-Male_Customer_Service
[ "region:us" ]
2022-06-22T08:03:49+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:07:55+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Male_Customer_Service ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 20 Hours - Chinese Mandarin Synthesis Corpus-Male, Customer Service. It is recorded by Chinese native speakers, the voice of the full of magnetism. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Male_Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n20 Hours - Chinese Mandarin Synthesis Corpus-Male, Customer Service. It is recorded by Chinese native speakers, the voice of the full of magnetism. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Male_Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n20 Hours - Chinese Mandarin Synthesis Corpus-Male, Customer Service. It is recorded by Chinese native speakers, the voice of the full of magnetism. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
254c3402c1f52a25e90b4f8c7bd73004870a582d
# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus-Customer_Service ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1100?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 150 People - Chinese Mandarin Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1100?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus-Customer_Service
[ "region:us" ]
2022-06-22T08:05:20+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:09:07+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus-Customer_Service ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 150 People - Chinese Mandarin Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus-Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n150 People - Chinese Mandarin Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus-Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n150 People - Chinese Mandarin Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
b159a2ec6a53a601cc3f9ece18457d0704152a9c
# Dataset Card for Nexdata/Chinese-English_Mixed_Average_Tone_Speech_Synthesis_Corpus-Customer_Service ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1118?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 50 People - Chinese-English Mixed Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1118?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese, English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese-English_Mixed_Average_Tone_Speech_Synthesis_Corpus-Customer_Service
[ "region:us" ]
2022-06-22T08:09:56+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:16:27+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese-English_Mixed_Average_Tone_Speech_Synthesis_Corpus-Customer_Service ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 50 People - Chinese-English Mixed Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese, English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese-English_Mixed_Average_Tone_Speech_Synthesis_Corpus-Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n50 People - Chinese-English Mixed Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese, English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese-English_Mixed_Average_Tone_Speech_Synthesis_Corpus-Customer_Service", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n50 People - Chinese-English Mixed Average Tone Speech Synthesis Corpus-Customer Service. It is recorded by Chinese native speakers,customer service text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese, English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
d4d108abc7855e35d218e8209976b2e6bdd531d1
# Dataset Card for Nexdata/Filipino_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1126?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 500 Hours - Filipino Speech Data by Mobile Phone,the data were recorded by Filipino speakers with authentic Filipino accents.The text is manually proofread with high accuracy. Match mainstream Android, Apple system phones. For more details, please refer to the link: https://www.nexdata.ai/datasets/1126?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Filipino ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Filipino_Speech_Data_by_Mobile_Phone
[ "region:us" ]
2022-06-22T08:16:44+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:20:55+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Filipino_Speech_Data_by_Mobile_Phone ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 500 Hours - Filipino Speech Data by Mobile Phone,the data were recorded by Filipino speakers with authentic Filipino accents.The text is manually proofread with high accuracy. Match mainstream Android, Apple system phones. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Filipino ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Filipino_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n500 Hours - Filipino Speech Data by Mobile Phone,the data were recorded by Filipino speakers with authentic Filipino accents.The text is manually proofread with high accuracy. Match mainstream Android, Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nFilipino", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Filipino_Speech_Data_by_Mobile_Phone", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n500 Hours - Filipino Speech Data by Mobile Phone,the data were recorded by Filipino speakers with authentic Filipino accents.The text is manually proofread with high accuracy. Match mainstream Android, Apple system phones.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\nautomatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR).", "### Languages\n\nFilipino", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
219cceddba05880ec920cfecc73e8afb93585a7f
# Dataset Card for Nexdata/Chinese_Mandarin_Entertainment_anchor_Style_Multi-emotional_Synthesis_Corpus ## Description 12 Hours - Chinese Mandarin Entertainment anchor Style Multi-emotional Synthesis Corpus. It is recorded by Chinese native speaker. six emotional text+modal particles, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1304?source=Huggingface # Specifications ## Format 48,000Hz, 24bit, uncompressed wav, mono channel ## Recording environment professional recording studio ## Recording content seven emotions (happiness, anger, sadness, surprise, fear, disgust)+sentences with filler word ## Speaker professional CharacterVoice; Role: An 18-year-old girl who works as an entertainment anchor and enjoys singing and dancing ## Device microphone ## Language Mandarin ## Annotation word and pinyin transcription, prosodic boundary annotation, phoneme boundary annotation ## The amount of data The amount of neutral data is not less than 1.6 hours; the amount of data with filler word is not less than 0.4 hours; and the remaining six types of emotional data is not less than 1.67 hours each # Licensing Information Commercial License
Nexdata/Chinese_Mandarin_Entertainment_anchor_Style_Multi-emotional_Synthesis_Corpus
[ "task_categories:text-to-speech", "language:zh", "region:us" ]
2022-06-22T08:18:30+00:00
{"language": ["zh"], "task_categories": ["text-to-speech"]}
2023-11-10T07:22:29+00:00
[]
[ "zh" ]
TAGS #task_categories-text-to-speech #language-Chinese #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Entertainment_anchor_Style_Multi-emotional_Synthesis_Corpus ## Description 12 Hours - Chinese Mandarin Entertainment anchor Style Multi-emotional Synthesis Corpus. It is recorded by Chinese native speaker. six emotional text+modal particles, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL # Specifications ## Format 48,000Hz, 24bit, uncompressed wav, mono channel ## Recording environment professional recording studio ## Recording content seven emotions (happiness, anger, sadness, surprise, fear, disgust)+sentences with filler word ## Speaker professional CharacterVoice; Role: An 18-year-old girl who works as an entertainment anchor and enjoys singing and dancing ## Device microphone ## Language Mandarin ## Annotation word and pinyin transcription, prosodic boundary annotation, phoneme boundary annotation ## The amount of data The amount of neutral data is not less than 1.6 hours; the amount of data with filler word is not less than 0.4 hours; and the remaining six types of emotional data is not less than 1.67 hours each # Licensing Information Commercial License
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Entertainment_anchor_Style_Multi-emotional_Synthesis_Corpus", "## Description\n12 Hours - Chinese Mandarin Entertainment anchor Style Multi-emotional Synthesis Corpus. It is recorded by Chinese native speaker. six emotional text+modal particles, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n\nFor more details, please refer to the link: URL", "# Specifications", "## Format\n48,000Hz, 24bit, uncompressed wav, mono channel", "## Recording environment\nprofessional recording studio", "## Recording content\nseven emotions (happiness, anger, sadness, surprise, fear, disgust)+sentences with filler word", "## Speaker\nprofessional CharacterVoice; Role: An 18-year-old girl who works as an entertainment anchor and enjoys singing and dancing", "## Device\nmicrophone", "## Language\nMandarin", "## Annotation\nword and pinyin transcription, prosodic boundary annotation, phoneme boundary annotation", "## The amount of data\nThe amount of neutral data is not less than 1.6 hours; the amount of data with filler word is not less than 0.4 hours; and the remaining six types of emotional data is not less than 1.67 hours each", "# Licensing Information\nCommercial License" ]
[ "TAGS\n#task_categories-text-to-speech #language-Chinese #region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Entertainment_anchor_Style_Multi-emotional_Synthesis_Corpus", "## Description\n12 Hours - Chinese Mandarin Entertainment anchor Style Multi-emotional Synthesis Corpus. It is recorded by Chinese native speaker. six emotional text+modal particles, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n\nFor more details, please refer to the link: URL", "# Specifications", "## Format\n48,000Hz, 24bit, uncompressed wav, mono channel", "## Recording environment\nprofessional recording studio", "## Recording content\nseven emotions (happiness, anger, sadness, surprise, fear, disgust)+sentences with filler word", "## Speaker\nprofessional CharacterVoice; Role: An 18-year-old girl who works as an entertainment anchor and enjoys singing and dancing", "## Device\nmicrophone", "## Language\nMandarin", "## Annotation\nword and pinyin transcription, prosodic boundary annotation, phoneme boundary annotation", "## The amount of data\nThe amount of neutral data is not less than 1.6 hours; the amount of data with filler word is not less than 0.4 hours; and the remaining six types of emotional data is not less than 1.67 hours each", "# Licensing Information\nCommercial License" ]
c7f1176cdd74f48750b8a4edf5113a017445536a
# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus_General ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1139?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 100 People - Chinese Mandarin Average Tone Speech Synthesis Corpus, General. It is recorded by Chinese native speaker. It covers news, dialogue, audio books, poetry, advertising, news broadcasting, entertainment; and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1139?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus_General
[ "region:us" ]
2022-06-22T08:20:03+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:17:22+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus_General ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 100 People - Chinese Mandarin Average Tone Speech Synthesis Corpus, General. It is recorded by Chinese native speaker. It covers news, dialogue, audio books, poetry, advertising, news broadcasting, entertainment; and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus_General", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n100 People - Chinese Mandarin Average Tone Speech Synthesis Corpus, General. It is recorded by Chinese native speaker. It covers news, dialogue, audio books, poetry, advertising, news broadcasting, entertainment; and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Average_Tone_Speech_Synthesis_Corpus_General", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n100 People - Chinese Mandarin Average Tone Speech Synthesis Corpus, General. It is recorded by Chinese native speaker. It covers news, dialogue, audio books, poetry, advertising, news broadcasting, entertainment; and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
a691a39a35ff98c465ea47ce5ab9359ce3dabe00
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_General ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1140?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Chinese Mandarin Synthesis Corpus-Female, General. It is recorded by Chinese native speaker. It covers oral sentences, audio books, news, advertising, customer service and movie commentary, and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1140?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_General
[ "region:us" ]
2022-06-22T08:21:37+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:20:31+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_General ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Chinese Mandarin Synthesis Corpus-Female, General. It is recorded by Chinese native speaker. It covers oral sentences, audio books, news, advertising, customer service and movie commentary, and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_General", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nChinese Mandarin Synthesis Corpus-Female, General. It is recorded by Chinese native speaker. It covers oral sentences, audio books, news, advertising, customer service and movie commentary, and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_General", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nChinese Mandarin Synthesis Corpus-Female, General. It is recorded by Chinese native speaker. It covers oral sentences, audio books, news, advertising, customer service and movie commentary, and the phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
59e0b24c5371be5e3360d64ef2a8a7a982039260
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Emotional ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1141?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 13.3 Hours - Chinese Mandarin Synthesis Corpus-Female, Emotional. It is recorded by Chinese native speaker,emotional text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1141?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Emotional
[ "region:us" ]
2022-06-22T08:22:59+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:20:00+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Emotional ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 13.3 Hours - Chinese Mandarin Synthesis Corpus-Female, Emotional. It is recorded by Chinese native speaker,emotional text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Emotional", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n13.3 Hours - Chinese Mandarin Synthesis Corpus-Female, Emotional. It is recorded by Chinese native speaker,emotional text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Emotional", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n13.3 Hours - Chinese Mandarin Synthesis Corpus-Female, Emotional. It is recorded by Chinese native speaker,emotional text, and the syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
26c6a7bfb716ca032e7cc51ae1ffc67ce0bb3a51
# Dataset Card for Nexdata/Chinese_Average_Tone_Speech_Synthesis_Corpus-Three_Styles ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1144?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 50 People - Chinese Average Tone Speech Synthesis Corpus-Three Styles.It is recorded by Chinese native speakers. Corpus includes cunstomer service,news and story. The syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1144?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Average_Tone_Speech_Synthesis_Corpus-Three_Styles
[ "region:us" ]
2022-06-22T08:24:17+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:18:03+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Average_Tone_Speech_Synthesis_Corpus-Three_Styles ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 50 People - Chinese Average Tone Speech Synthesis Corpus-Three Styles.It is recorded by Chinese native speakers. Corpus includes cunstomer service,news and story. The syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Average_Tone_Speech_Synthesis_Corpus-Three_Styles", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n50 People - Chinese Average Tone Speech Synthesis Corpus-Three Styles.It is recorded by Chinese native speakers. Corpus includes cunstomer service,news and story. The syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Average_Tone_Speech_Synthesis_Corpus-Three_Styles", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n50 People - Chinese Average Tone Speech Synthesis Corpus-Three Styles.It is recorded by Chinese native speakers. Corpus includes cunstomer service,news and story. The syllables, phonemes and tones are balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
3a11c416ea0cd6073c6bec31cd49225e84f7306b
# Dataset Card for Nexdata/Chinese_Mandarin_Songs_in_Acapella__Female ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1151?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 103 Chinese Mandarin Songs in Acapella - Female. It is recorded by Chinese professional singer, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the song synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1151?source=Huggingface ### Supported Tasks and Leaderboards tts,: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Songs_in_Acapella__Female
[ "region:us" ]
2022-06-22T08:25:46+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:18:29+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Songs_in_Acapella__Female ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 103 Chinese Mandarin Songs in Acapella - Female. It is recorded by Chinese professional singer, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the song synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts,: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Songs_in_Acapella__Female", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n103 Chinese Mandarin Songs in Acapella - Female. It is recorded by Chinese professional singer, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the song synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts,: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Songs_in_Acapella__Female", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n103 Chinese Mandarin Songs in Acapella - Female. It is recorded by Chinese professional singer, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the song synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts,: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
5fdc226629dc028980e4fe94370203445c9d7d71
# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Male ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1159?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Male audio data of American English. It is recorded by American English native speakers, with authentic accent. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1159?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages American English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/American_English_Speech_Synthesis_Corpus-Male
[ "region:us" ]
2022-06-22T08:27:09+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:19:26+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Male ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary Male audio data of American English. It is recorded by American English native speakers, with authentic accent. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages American English ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Male", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMale audio data of American English. It is recorded by American English native speakers, with authentic accent. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/American_English_Speech_Synthesis_Corpus-Male", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nMale audio data of American English. It is recorded by American English native speakers, with authentic accent. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nAmerican English", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
d7673e82101404e9096b10280954a2d8fd17b200
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Customer_Service_Conversational_Speech ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1167?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 20 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, Conversational Speech, It is recorded by Chinese native speakers, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1167?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Customer_Service_Conversational_Speech
[ "region:us" ]
2022-06-22T08:28:32+00:00
{"YAML tags": [{"copy-paste the tags obtained with the tagging app": "https://github.com/huggingface/datasets-tagging"}]}
2023-08-28T07:18:57+00:00
[]
[]
TAGS #region-us
# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Customer_Service_Conversational_Speech ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary 20 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, Conversational Speech, It is recorded by Chinese native speakers, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: URL ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Chinese Mandarin ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Commerical License: URL ### Contributions
[ "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Customer_Service_Conversational_Speech", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n20 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, Conversational Speech, It is recorded by Chinese native speakers, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
[ "TAGS\n#region-us \n", "# Dataset Card for Nexdata/Chinese_Mandarin_Synthesis_Corpus-Female_Customer_Service_Conversational_Speech", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\n20 Hours - Chinese Mandarin Synthesis Corpus-Female, Customer Service, Conversational Speech, It is recorded by Chinese native speakers, with sweet voice. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis.\n \nFor more details, please refer to the link: URL", "### Supported Tasks and Leaderboards\n\ntts: The dataset can be used to train a model for Text to Speech (TTS).", "### Languages\n\nChinese Mandarin", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCommerical License: URL", "### Contributions" ]
4b37cb089a33454dcfd5c1af2902e58464a41fdb
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655900658
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T11:24:18+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T11:24:21+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
e529817c203e680865a51ea9940f2ee1eb85b2af
# About Dataset ### Context Text summarization is a way to condense the large amount of information into a concise form by the process of selection of important information and discarding unimportant and redundant information. With the amount of textual information present in the world wide web the area of text summarization is becoming very important. The extractive summarization is the one where the exact sentences present in the document are used as summaries. The extractive summarization is simpler and is the general practice among the automatic text summarization researchers at the present time. Extractive summarization process involves giving scores to sentences using some method and then using the sentences that achieve highest scores as summaries. As the exact sentence present in the document is used the semantic factor can be ignored which results in generation of less calculation intensive summarization procedure. This kind of summary is generally completely unsupervised and language independent too. Although this kind of summary does its job in conveying the essential information it may not be necessarily smooth or fluent. Sometimes there can be almost no connection between adjacent sentences in the summary resulting in the text lacking in readability. Content This dataset for extractive text summarization has four hundred and seventeen political news articles of BBC from 2004 to 2005 in the News Articles folder. For each articles, five summaries are provided in the Summaries folder. The first clause of the text of articles is the respective title. Acknowledgements This dataset was created using a dataset used for data categorization that onsists of 2225 documents from the BBC news website corresponding to stories in five topical areas from 2004-2005 used in the paper of D. Greene and P. Cunningham. "Practical Solutions to the Problem of Diagonal Dominance in Kernel Document Clustering", Proc. ICML 2006; whose all rights, including copyright, in the content of the original articles are owned by the BBC. More at http://mlg.ucd.ie/datasets/bbc.html **Kaggle Link:** https://www.kaggle.com/datasets/pariza/bbc-news-summary
gopalkalpande/bbc-news-summary
[ "license:cc0-1.0", "region:us" ]
2022-06-22T11:56:16+00:00
{"license": "cc0-1.0"}
2022-06-22T12:08:15+00:00
[]
[]
TAGS #license-cc0-1.0 #region-us
# About Dataset ### Context Text summarization is a way to condense the large amount of information into a concise form by the process of selection of important information and discarding unimportant and redundant information. With the amount of textual information present in the world wide web the area of text summarization is becoming very important. The extractive summarization is the one where the exact sentences present in the document are used as summaries. The extractive summarization is simpler and is the general practice among the automatic text summarization researchers at the present time. Extractive summarization process involves giving scores to sentences using some method and then using the sentences that achieve highest scores as summaries. As the exact sentence present in the document is used the semantic factor can be ignored which results in generation of less calculation intensive summarization procedure. This kind of summary is generally completely unsupervised and language independent too. Although this kind of summary does its job in conveying the essential information it may not be necessarily smooth or fluent. Sometimes there can be almost no connection between adjacent sentences in the summary resulting in the text lacking in readability. Content This dataset for extractive text summarization has four hundred and seventeen political news articles of BBC from 2004 to 2005 in the News Articles folder. For each articles, five summaries are provided in the Summaries folder. The first clause of the text of articles is the respective title. Acknowledgements This dataset was created using a dataset used for data categorization that onsists of 2225 documents from the BBC news website corresponding to stories in five topical areas from 2004-2005 used in the paper of D. Greene and P. Cunningham. "Practical Solutions to the Problem of Diagonal Dominance in Kernel Document Clustering", Proc. ICML 2006; whose all rights, including copyright, in the content of the original articles are owned by the BBC. More at URL Kaggle Link: URL
[ "# About Dataset", "### Context\n\nText summarization is a way to condense the large amount of information into a concise form by the process of selection of important information and discarding unimportant and redundant information. With the amount of textual information present in the world wide web the area of text summarization is becoming very important. The extractive summarization is the one where the exact sentences present in the document are used as summaries. The extractive summarization is simpler and is the general practice among the automatic text summarization researchers at the present time. Extractive summarization process involves giving scores to sentences using some method and then using the sentences that achieve highest scores as summaries. As the exact sentence present in the document is used the semantic factor can be ignored which results in generation of less calculation intensive summarization procedure. This kind of summary is generally completely unsupervised and language independent too. Although this kind of summary does its job in conveying the essential information it may not be necessarily smooth or fluent. Sometimes there can be almost no connection between adjacent sentences in the summary resulting in the text lacking in readability.\n\nContent\nThis dataset for extractive text summarization has four hundred and seventeen political news articles of BBC from 2004 to 2005 in the News Articles folder. For each articles, five summaries are provided in the Summaries folder. The first clause of the text of articles is the respective title.\n\nAcknowledgements\nThis dataset was created using a dataset used for data categorization that onsists of 2225 documents from the BBC news website corresponding to stories in five topical areas from 2004-2005 used in the paper of D. Greene and P. Cunningham. \"Practical Solutions to the Problem of Diagonal Dominance in Kernel Document Clustering\", Proc. ICML 2006; whose all rights, including copyright, in the content of the original articles are owned by the BBC. More at URL\n\nKaggle Link: URL" ]
[ "TAGS\n#license-cc0-1.0 #region-us \n", "# About Dataset", "### Context\n\nText summarization is a way to condense the large amount of information into a concise form by the process of selection of important information and discarding unimportant and redundant information. With the amount of textual information present in the world wide web the area of text summarization is becoming very important. The extractive summarization is the one where the exact sentences present in the document are used as summaries. The extractive summarization is simpler and is the general practice among the automatic text summarization researchers at the present time. Extractive summarization process involves giving scores to sentences using some method and then using the sentences that achieve highest scores as summaries. As the exact sentence present in the document is used the semantic factor can be ignored which results in generation of less calculation intensive summarization procedure. This kind of summary is generally completely unsupervised and language independent too. Although this kind of summary does its job in conveying the essential information it may not be necessarily smooth or fluent. Sometimes there can be almost no connection between adjacent sentences in the summary resulting in the text lacking in readability.\n\nContent\nThis dataset for extractive text summarization has four hundred and seventeen political news articles of BBC from 2004 to 2005 in the News Articles folder. For each articles, five summaries are provided in the Summaries folder. The first clause of the text of articles is the respective title.\n\nAcknowledgements\nThis dataset was created using a dataset used for data categorization that onsists of 2225 documents from the BBC news website corresponding to stories in five topical areas from 2004-2005 used in the paper of D. Greene and P. Cunningham. \"Practical Solutions to the Problem of Diagonal Dominance in Kernel Document Clustering\", Proc. ICML 2006; whose all rights, including copyright, in the content of the original articles are owned by the BBC. More at URL\n\nKaggle Link: URL" ]
6951a96ab68c2a37040ee9912bb326bd6d29a41e
# Dataset Card for #PraCegoVer ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Citation Information](#citation-information) ## Dataset Description - **Dataset repository:** [#PraCegoVer dataset](https://doi.org/10.5281/zenodo.5710562) - **Github repository:** [PraCegoVer](https://github.com/larocs/PraCegoVer) - **Paper:** [#PraCegoVer: A Large Dataset for Image Captioning in Portuguese](https://doi.org/10.3390/data7020013) - **Contact:** [Gabriel Oliveira]([email protected]) ### Dataset Summary \#PraCegoVer is a multi-modal dataset with Portuguese captions based on posts from Instagram. It is the first large dataset for image captioning in Portuguese with freely annotated images. The dataset has been created to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks. \#PraCegoVer comprehends 533,523 instances that represent public posts collected from Instagram tagged with #PraCegoVer. the data were collected from more than 14 thousand different profiles. This dataset contains images of people, and it consists of data collected from public profiles on Instagram. Thus, the images and raw captions might contain sensitive data that reveal racial or ethnic origins, sexual orientations, and religious beliefs. Hence, under Brazilian Law No. 13,709, to avoid the unintended use of our dataset, we decided to restrict its access, ensuring that the dataset will be used for **research purposes only**. ### Supported Tasks and Leaderboards - `image-captioning`, `image-to-text`: the dataset can be used to train models for Image Captioning, which consists in generating a short description of the visual content of a given image. The model performance is typically measured using [ROUGE](https://huggingface.co/metrics/rouge), [METEOR](https://huggingface.co/spaces/evaluate-metric/meteor), [**CIDEr-D**](https://openaccess.thecvf.com/content_cvpr_2015/html/Vedantam_CIDEr_Consensus-Based_Image_2015_CVPR_paper.html), [**CIDEr-R**](https://aclanthology.org/2021.wnut-1.39/) and [**SPICE**](https://link.springer.com/chapter/10.1007/978-3-319-46454-1_24). ### Languages The captions in this dataset are in Brazilian Portuguese (pt-BR). ## Dataset Structure ### Data Instances \#PraCegoVer dataset is composed of the main file `dataset.json` and a directory with images, `images`. The instances in `dataset.json` have the following format: ``` {'user': '16247e952a987935792d1d9d937eeb8413e0367cfb9c5e640db1d1bc4a58dc01', 'filename': 'i-00518416.jpg', 'raw_caption': 'Com mais de 12 milhões de habitantes 👨🏾👩🏼🦰, #SãoPaulo é a maior e mais populosa cidade do Brasil 🇧🇷 , além de ser a primeira metrópole da América e do hemisfério sul.\n\nSe você mora nessa cidade incrível, comente um ❤️ nesta imagem.\n\n📸Bruno Mancini\n\n#MaisSegurosJuntos #Segurança #Aplicativo #Metrópole #Brasil\n\n#PraCegoVer #PraTodosVerem: Foto do Rio Pinheiros, em São Paulo, mostrando a Ponte Estaiada e vários prédios dos dois lados do rio. Com um tom azulado a imagem possui quatro i´s transparentes bem suaves cobrindo-a toda', 'caption': 'Foto do Rio Pinheiros, em São Paulo, mostrando a Ponte Estaiada e vários prédios dos dois lados do rio. Com um tom azulado a imagem possui quatro i´s transparentes bem suaves cobrindo-a toda.', 'date': '25-09-2020'}, ``` ### Data Fields user: anonymized user that created the post; filename: image file name, which indicates the image in the `images` directory; raw_caption: raw caption; caption: clean caption; date: post date. ### Data Splits This dataset comes with two specified train/validation/test splits, one for #PraCegoVer-63K (train/validation/test: 37,881/12,442/12,612) and another for #PraCegoVer-173K (train/validation/test: 104,004/34,452/34,882). These splits are subsets of the whole dataset. ## Dataset Creation ### Curation Rationale Automatically describing images using natural sentences is an essential task to visually impaired people's inclusion on the Internet. Although there are many datasets in the literature, most of them contain only English captions, whereas datasets with captions described in other languages are scarce. Then, inspired by the movement [PraCegoVer](https://mwpt.com.br/criadora-do-projeto-pracegover-incentiva-descricao-de-imagens-na-web/), #PraCegoVer dataset has been created to provide images annotated with descriptions in Portuguese for the image captioning task. With this dataset, we aim to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks. ### Source Data #### Initial Data Collection and Normalization The data were collected from posts on Instagram that tagged #PraCegoVer. Then, descriptions are extracted from the raw image caption by using regular expressions. The script to download more data is available in the [\#PraCegoVer repository](https://github.com/larocs/PraCegoVer). We collected the data on a daily basis from 2020 to 2021, but the posts can have been created at any time before this period. #### Who are the source language producers? The language producers are Instagram users that post images tagging #PraCegoVer. #### Who are the annotators? The Instagram users that tag \#PraCegoVer spontaneously add a short description of the image content in their posts. ### Personal and Sensitive Information The usernames were anonymized in order to make it difficult to directly identify the individuals. However, we don't anonymize the remaining data, thus the individuals present in the images can be identified. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop better image captioning models in Portuguese. Such models are essential to help the inclusion of visually impaired people on the Internet, making it more inclusive and democratic. However, it is worth noting that the data were collected from public profiles on Instagram, and were not thoroughly validated. Thus, there might be more examples of offensive and insulting content, albeit an exploratory analysis shows that although there exist words that can be offensive they are insignificant because they occur rarely. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships. ### Discussion of Biases We collected the data from public posts on Instagram. Thus, the data is susceptible to the bias of its algorithm and stereotypes. We conducted an initial analysis of the bias within our dataset. [Figure 20](https://www.mdpi.com/data/data-07-00013/article_deploy/html/images/data-07-00013-g020-550.jpg) from the dataset's paper shows that women are frequently associated with beauty, cosmetic products, and domestic violence. Moreover, black women co-occur more often with terms such as "racism", "discrimination", "prejudice"and "consciousness", whereas white women appear with "spa", "hair", and "lipstick", and indigenous women are mostly associated with beauty products. Similarly, black men frequently appear together with the terms "Zumbi dos Palmares", "consciousness", "racism", "United States"and "justice", while white men are associated with "theatre", "wage", "benefit"and "social security". In addition, [Table 4](https://www.mdpi.com/2306-5729/7/2/13/htm) from the dataset's paper shows that women are more frequently associated with physical words (e.g., thin, fat); still, fat people appear more frequently than thin people. [Figure 21](https://www.mdpi.com/data/data-07-00013/article_deploy/html/images/data-07-00013-g021-550.jpg) from the dataset's paper illustrates that fat women are also related to swearing words, "mental harassment", and "boss", while thin women are associated with "vitamin", "fruits", and "healthy skin". To sum up, depending on the usage of this dataset, future users may take these aspects into account. ## Additional Information ### Dataset Curators The dataset was created by [Gabriel Oliveira](https://orcid.org/0000-0003-2835-1331), [Esther Colombini](https://orcid.org/0000-0003-0467-3133) and [Sandra Avila](https://orcid.org/0000-0001-9068-938X). ### Citation Information If you use \#PraCegoVer dataset, please cite as: ``` @article{pracegover2022, AUTHOR = {dos Santos, Gabriel Oliveira and Colombini, Esther Luna and Avila, Sandra}, TITLE = {#PraCegoVer: A Large Dataset for Image Captioning in Portuguese}, JOURNAL = {Data}, VOLUME = {7}, YEAR = {2022}, NUMBER = {2}, ARTICLE-NUMBER = {13}, URL = {https://www.mdpi.com/2306-5729/7/2/13}, ISSN = {2306-5729}, DOI = {10.3390/data7020013} } ```
gabrielsantosrv/pracegover
[ "language:pt", "region:us" ]
2022-06-22T11:57:45+00:00
{"language": ["pt"]}
2023-03-13T15:25:01+00:00
[]
[ "pt" ]
TAGS #language-Portuguese #region-us
# Dataset Card for #PraCegoVer ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Additional Information - Dataset Curators - Citation Information ## Dataset Description - Dataset repository: #PraCegoVer dataset - Github repository: PraCegoVer - Paper: #PraCegoVer: A Large Dataset for Image Captioning in Portuguese - Contact: Gabriel Oliveira ### Dataset Summary \#PraCegoVer is a multi-modal dataset with Portuguese captions based on posts from Instagram. It is the first large dataset for image captioning in Portuguese with freely annotated images. The dataset has been created to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks. \#PraCegoVer comprehends 533,523 instances that represent public posts collected from Instagram tagged with #PraCegoVer. the data were collected from more than 14 thousand different profiles. This dataset contains images of people, and it consists of data collected from public profiles on Instagram. Thus, the images and raw captions might contain sensitive data that reveal racial or ethnic origins, sexual orientations, and religious beliefs. Hence, under Brazilian Law No. 13,709, to avoid the unintended use of our dataset, we decided to restrict its access, ensuring that the dataset will be used for research purposes only. ### Supported Tasks and Leaderboards - 'image-captioning', 'image-to-text': the dataset can be used to train models for Image Captioning, which consists in generating a short description of the visual content of a given image. The model performance is typically measured using ROUGE, METEOR, CIDEr-D, CIDEr-R and SPICE. ### Languages The captions in this dataset are in Brazilian Portuguese (pt-BR). ## Dataset Structure ### Data Instances \#PraCegoVer dataset is composed of the main file 'URL' and a directory with images, 'images'. The instances in 'URL' have the following format: ### Data Fields user: anonymized user that created the post; filename: image file name, which indicates the image in the 'images' directory; raw_caption: raw caption; caption: clean caption; date: post date. ### Data Splits This dataset comes with two specified train/validation/test splits, one for #PraCegoVer-63K (train/validation/test: 37,881/12,442/12,612) and another for #PraCegoVer-173K (train/validation/test: 104,004/34,452/34,882). These splits are subsets of the whole dataset. ## Dataset Creation ### Curation Rationale Automatically describing images using natural sentences is an essential task to visually impaired people's inclusion on the Internet. Although there are many datasets in the literature, most of them contain only English captions, whereas datasets with captions described in other languages are scarce. Then, inspired by the movement PraCegoVer, #PraCegoVer dataset has been created to provide images annotated with descriptions in Portuguese for the image captioning task. With this dataset, we aim to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks. ### Source Data #### Initial Data Collection and Normalization The data were collected from posts on Instagram that tagged #PraCegoVer. Then, descriptions are extracted from the raw image caption by using regular expressions. The script to download more data is available in the \#PraCegoVer repository. We collected the data on a daily basis from 2020 to 2021, but the posts can have been created at any time before this period. #### Who are the source language producers? The language producers are Instagram users that post images tagging #PraCegoVer. #### Who are the annotators? The Instagram users that tag \#PraCegoVer spontaneously add a short description of the image content in their posts. ### Personal and Sensitive Information The usernames were anonymized in order to make it difficult to directly identify the individuals. However, we don't anonymize the remaining data, thus the individuals present in the images can be identified. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop better image captioning models in Portuguese. Such models are essential to help the inclusion of visually impaired people on the Internet, making it more inclusive and democratic. However, it is worth noting that the data were collected from public profiles on Instagram, and were not thoroughly validated. Thus, there might be more examples of offensive and insulting content, albeit an exploratory analysis shows that although there exist words that can be offensive they are insignificant because they occur rarely. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships. ### Discussion of Biases We collected the data from public posts on Instagram. Thus, the data is susceptible to the bias of its algorithm and stereotypes. We conducted an initial analysis of the bias within our dataset. Figure 20 from the dataset's paper shows that women are frequently associated with beauty, cosmetic products, and domestic violence. Moreover, black women co-occur more often with terms such as "racism", "discrimination", "prejudice"and "consciousness", whereas white women appear with "spa", "hair", and "lipstick", and indigenous women are mostly associated with beauty products. Similarly, black men frequently appear together with the terms "Zumbi dos Palmares", "consciousness", "racism", "United States"and "justice", while white men are associated with "theatre", "wage", "benefit"and "social security". In addition, Table 4 from the dataset's paper shows that women are more frequently associated with physical words (e.g., thin, fat); still, fat people appear more frequently than thin people. Figure 21 from the dataset's paper illustrates that fat women are also related to swearing words, "mental harassment", and "boss", while thin women are associated with "vitamin", "fruits", and "healthy skin". To sum up, depending on the usage of this dataset, future users may take these aspects into account. ## Additional Information ### Dataset Curators The dataset was created by Gabriel Oliveira, Esther Colombini and Sandra Avila. If you use \#PraCegoVer dataset, please cite as:
[ "# Dataset Card for #PraCegoVer", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n- Additional Information\n - Dataset Curators\n - Citation Information", "## Dataset Description\n\n- Dataset repository: #PraCegoVer dataset\n- Github repository: PraCegoVer\n- Paper: #PraCegoVer: A Large Dataset for Image Captioning in Portuguese\n- Contact: Gabriel Oliveira", "### Dataset Summary\n\n\\#PraCegoVer is a multi-modal dataset with Portuguese captions based on posts from Instagram. It is the first large dataset for image captioning in Portuguese with freely annotated images. The dataset has been created to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks.\n\n\\#PraCegoVer comprehends 533,523 instances that represent public posts collected from Instagram tagged with #PraCegoVer. the data were collected from more than 14 thousand different profiles.\n\nThis dataset contains images of people, and it consists of data collected from public profiles on Instagram. Thus, the images and raw captions might contain sensitive data that reveal racial or ethnic origins, sexual orientations, and religious beliefs. Hence, under Brazilian Law No. 13,709, to avoid the unintended use of our dataset, we decided to restrict its access, ensuring that the dataset will be used for research purposes only.", "### Supported Tasks and Leaderboards\n\n- 'image-captioning', 'image-to-text': the dataset can be used to train models for Image Captioning, which consists in generating a short description of the visual content of a given image. The model performance is typically measured using ROUGE, METEOR, CIDEr-D, CIDEr-R and SPICE.", "### Languages\n\nThe captions in this dataset are in Brazilian Portuguese (pt-BR).", "## Dataset Structure", "### Data Instances\n\n\\#PraCegoVer dataset is composed of the main file 'URL' and a directory with images, 'images'. The instances in 'URL' have the following format:", "### Data Fields\n\n\nuser: anonymized user that created the post;\nfilename: image file name, which indicates the image in the 'images' directory;\nraw_caption: raw caption;\ncaption: clean caption;\ndate: post date.", "### Data Splits\n\nThis dataset comes with two specified train/validation/test splits, one for #PraCegoVer-63K (train/validation/test: 37,881/12,442/12,612) and another for #PraCegoVer-173K (train/validation/test: 104,004/34,452/34,882). These splits are subsets of the whole dataset.", "## Dataset Creation", "### Curation Rationale\n\nAutomatically describing images using natural sentences is an essential task to visually impaired people's inclusion on the Internet. Although there are many datasets in the literature, most of them contain only English captions, whereas datasets with captions described in other languages are scarce.\n\nThen, inspired by the movement PraCegoVer, #PraCegoVer dataset has been created to provide images annotated with descriptions in Portuguese for the image captioning task. With this dataset, we aim to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data were collected from posts on Instagram that tagged #PraCegoVer. Then, descriptions are extracted from the raw image caption by using regular expressions. The script to download more data is available in the \\#PraCegoVer repository. We collected the data on a daily basis from 2020 to 2021, but the posts can have been created at any time before this period.", "#### Who are the source language producers?\n\nThe language producers are Instagram users that post images tagging #PraCegoVer.", "#### Who are the annotators?\n\nThe Instagram users that tag \\#PraCegoVer spontaneously add a short description of the image content in their posts.", "### Personal and Sensitive Information\n\nThe usernames were anonymized in order to make it difficult to directly identify the individuals. However, we don't anonymize the remaining data, thus the individuals present in the images can be identified. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nThe purpose of this dataset is to help develop better image captioning models in Portuguese. Such models are essential to help the inclusion of visually impaired people on the Internet, making it more inclusive and democratic.\n\nHowever, it is worth noting that the data were collected from public profiles on Instagram, and were not thoroughly validated. Thus, there might be more examples of offensive and insulting content, albeit an exploratory analysis shows that although there exist words that can be offensive they are insignificant because they occur rarely. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships.", "### Discussion of Biases\n\nWe collected the data from public posts on Instagram. Thus, the data is susceptible to the bias of its algorithm and stereotypes. We conducted an initial analysis of the bias within our dataset. Figure 20 from the dataset's paper shows that women are frequently associated with beauty, cosmetic products, and domestic violence. Moreover, black women co-occur more often with terms such as \"racism\", \"discrimination\", \"prejudice\"\u001dand \"consciousness\", whereas white women appear with \"spa\", \"hair\", and \"lipstick\", and indigenous women are mostly associated with beauty products. Similarly, black men frequently appear together with the terms \"Zumbi dos Palmares\", \"consciousness\", \"racism\", \"United States\"\u001dand \"justice\", while white men are associated with \"theatre\", \"wage\", \"benefit\"\u001dand \"social security\". In addition, Table 4 from the dataset's paper shows that women are more frequently associated with physical words (e.g., thin, fat); still, fat people appear more frequently than thin people. Figure 21 from the dataset's paper illustrates that fat women are also related to swearing words, \"mental harassment\", and \"boss\", while thin women are associated with \"vitamin\", \"fruits\", and \"healthy skin\". To sum up, depending on the usage of this dataset, future users may take these aspects into account.", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Gabriel Oliveira, Esther Colombini and Sandra Avila.\n\n\n\n\nIf you use \\#PraCegoVer dataset, please cite as:" ]
[ "TAGS\n#language-Portuguese #region-us \n", "# Dataset Card for #PraCegoVer", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n- Additional Information\n - Dataset Curators\n - Citation Information", "## Dataset Description\n\n- Dataset repository: #PraCegoVer dataset\n- Github repository: PraCegoVer\n- Paper: #PraCegoVer: A Large Dataset for Image Captioning in Portuguese\n- Contact: Gabriel Oliveira", "### Dataset Summary\n\n\\#PraCegoVer is a multi-modal dataset with Portuguese captions based on posts from Instagram. It is the first large dataset for image captioning in Portuguese with freely annotated images. The dataset has been created to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks.\n\n\\#PraCegoVer comprehends 533,523 instances that represent public posts collected from Instagram tagged with #PraCegoVer. the data were collected from more than 14 thousand different profiles.\n\nThis dataset contains images of people, and it consists of data collected from public profiles on Instagram. Thus, the images and raw captions might contain sensitive data that reveal racial or ethnic origins, sexual orientations, and religious beliefs. Hence, under Brazilian Law No. 13,709, to avoid the unintended use of our dataset, we decided to restrict its access, ensuring that the dataset will be used for research purposes only.", "### Supported Tasks and Leaderboards\n\n- 'image-captioning', 'image-to-text': the dataset can be used to train models for Image Captioning, which consists in generating a short description of the visual content of a given image. The model performance is typically measured using ROUGE, METEOR, CIDEr-D, CIDEr-R and SPICE.", "### Languages\n\nThe captions in this dataset are in Brazilian Portuguese (pt-BR).", "## Dataset Structure", "### Data Instances\n\n\\#PraCegoVer dataset is composed of the main file 'URL' and a directory with images, 'images'. The instances in 'URL' have the following format:", "### Data Fields\n\n\nuser: anonymized user that created the post;\nfilename: image file name, which indicates the image in the 'images' directory;\nraw_caption: raw caption;\ncaption: clean caption;\ndate: post date.", "### Data Splits\n\nThis dataset comes with two specified train/validation/test splits, one for #PraCegoVer-63K (train/validation/test: 37,881/12,442/12,612) and another for #PraCegoVer-173K (train/validation/test: 104,004/34,452/34,882). These splits are subsets of the whole dataset.", "## Dataset Creation", "### Curation Rationale\n\nAutomatically describing images using natural sentences is an essential task to visually impaired people's inclusion on the Internet. Although there are many datasets in the literature, most of them contain only English captions, whereas datasets with captions described in other languages are scarce.\n\nThen, inspired by the movement PraCegoVer, #PraCegoVer dataset has been created to provide images annotated with descriptions in Portuguese for the image captioning task. With this dataset, we aim to alleviate the lack of datasets with Portuguese captions for visual-linguistic tasks.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data were collected from posts on Instagram that tagged #PraCegoVer. Then, descriptions are extracted from the raw image caption by using regular expressions. The script to download more data is available in the \\#PraCegoVer repository. We collected the data on a daily basis from 2020 to 2021, but the posts can have been created at any time before this period.", "#### Who are the source language producers?\n\nThe language producers are Instagram users that post images tagging #PraCegoVer.", "#### Who are the annotators?\n\nThe Instagram users that tag \\#PraCegoVer spontaneously add a short description of the image content in their posts.", "### Personal and Sensitive Information\n\nThe usernames were anonymized in order to make it difficult to directly identify the individuals. However, we don't anonymize the remaining data, thus the individuals present in the images can be identified. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nThe purpose of this dataset is to help develop better image captioning models in Portuguese. Such models are essential to help the inclusion of visually impaired people on the Internet, making it more inclusive and democratic.\n\nHowever, it is worth noting that the data were collected from public profiles on Instagram, and were not thoroughly validated. Thus, there might be more examples of offensive and insulting content, albeit an exploratory analysis shows that although there exist words that can be offensive they are insignificant because they occur rarely. Moreover, the images and raw captions might contain data revealing racial or ethnic origins, sexual orientations, religious beliefs, political opinions, or union memberships.", "### Discussion of Biases\n\nWe collected the data from public posts on Instagram. Thus, the data is susceptible to the bias of its algorithm and stereotypes. We conducted an initial analysis of the bias within our dataset. Figure 20 from the dataset's paper shows that women are frequently associated with beauty, cosmetic products, and domestic violence. Moreover, black women co-occur more often with terms such as \"racism\", \"discrimination\", \"prejudice\"\u001dand \"consciousness\", whereas white women appear with \"spa\", \"hair\", and \"lipstick\", and indigenous women are mostly associated with beauty products. Similarly, black men frequently appear together with the terms \"Zumbi dos Palmares\", \"consciousness\", \"racism\", \"United States\"\u001dand \"justice\", while white men are associated with \"theatre\", \"wage\", \"benefit\"\u001dand \"social security\". In addition, Table 4 from the dataset's paper shows that women are more frequently associated with physical words (e.g., thin, fat); still, fat people appear more frequently than thin people. Figure 21 from the dataset's paper illustrates that fat women are also related to swearing words, \"mental harassment\", and \"boss\", while thin women are associated with \"vitamin\", \"fruits\", and \"healthy skin\". To sum up, depending on the usage of this dataset, future users may take these aspects into account.", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Gabriel Oliveira, Esther Colombini and Sandra Avila.\n\n\n\n\nIf you use \\#PraCegoVer dataset, please cite as:" ]
a5e415dfc7d7b5c370a6f8a4d18ffb679aa61f04
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655905032
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T12:37:13+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T12:37:16+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
a465c3749f98da4c8e222aa2a97b6c6c20b711b3
# Splits for the Ben-Mordecai and Elhadad Hebrew NER Corpus (BMC) In order to evaluate performance in accordance with the original Ben-Mordecai and Elhadad (2005) work, we provide three 75%-25% random splits. * Only the 7 entity categories viable for evaluation were kept (DATE, LOC, MONEY, ORG, PER, PERCENT, TIME) --- all MISC entities were filtered out. * Sequence label scheme was changed from IOB to BIOES * The dev sets are 10% taken out of the 75% ## Citation If you use use the BMC corpus, please cite the original paper as well as our paper which describes the splits: * Ben-Mordecai and Elhadad (2005): ```console @mastersthesis{naama, title={Hebrew Named Entity Recognition}, author={Ben-Mordecai, Naama}, advisor={Elhadad, Michael}, year={2005}, url="https://www.cs.bgu.ac.il/~elhadad/nlpproj/naama/", institution={Department of Computer Science, Ben-Gurion University}, school={Department of Computer Science, Ben-Gurion University}, } ``` * Bareket and Tsarfaty (2020) ```console @misc{bareket2020neural, title={Neural Modeling for Named Entities and Morphology (NEMO^2)}, author={Dan Bareket and Reut Tsarfaty}, year={2020}, eprint={2007.15620}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
imvladikon/bmc
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-reuters-corpus", "language:he", "license:other", "arxiv:2007.15620", "region:us" ]
2022-06-22T14:39:14+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["he"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["extended|other-reuters-corpus"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "train-eval-index": [{"config": "bmc", "task": "token-classification", "task_id": "entity_extraction", "splits": {"train_split": "train", "eval_split": "validation", "test_split": "test"}, "col_mapping": {"tokens": "tokens", "ner_tags": "tags"}, "metrics": [{"type": "seqeval", "name": "seqeval"}]}]}
2022-11-17T16:52:43+00:00
[ "2007.15620" ]
[ "he" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-reuters-corpus #language-Hebrew #license-other #arxiv-2007.15620 #region-us
# Splits for the Ben-Mordecai and Elhadad Hebrew NER Corpus (BMC) In order to evaluate performance in accordance with the original Ben-Mordecai and Elhadad (2005) work, we provide three 75%-25% random splits. * Only the 7 entity categories viable for evaluation were kept (DATE, LOC, MONEY, ORG, PER, PERCENT, TIME) --- all MISC entities were filtered out. * Sequence label scheme was changed from IOB to BIOES * The dev sets are 10% taken out of the 75% If you use use the BMC corpus, please cite the original paper as well as our paper which describes the splits: * Ben-Mordecai and Elhadad (2005): * Bareket and Tsarfaty (2020)
[ "# Splits for the Ben-Mordecai and Elhadad Hebrew NER Corpus (BMC)\n\nIn order to evaluate performance in accordance with the original Ben-Mordecai and Elhadad (2005) work, we provide three 75%-25% random splits. \n* Only the 7 entity categories viable for evaluation were kept (DATE, LOC, MONEY, ORG, PER, PERCENT, TIME) --- all MISC entities were filtered out.\n* Sequence label scheme was changed from IOB to BIOES\n* The dev sets are 10% taken out of the 75%\n\n\nIf you use use the BMC corpus, please cite the original paper as well as our paper which describes the splits:\n\n* Ben-Mordecai and Elhadad (2005):\n\n\n* Bareket and Tsarfaty (2020)" ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-reuters-corpus #language-Hebrew #license-other #arxiv-2007.15620 #region-us \n", "# Splits for the Ben-Mordecai and Elhadad Hebrew NER Corpus (BMC)\n\nIn order to evaluate performance in accordance with the original Ben-Mordecai and Elhadad (2005) work, we provide three 75%-25% random splits. \n* Only the 7 entity categories viable for evaluation were kept (DATE, LOC, MONEY, ORG, PER, PERCENT, TIME) --- all MISC entities were filtered out.\n* Sequence label scheme was changed from IOB to BIOES\n* The dev sets are 10% taken out of the 75%\n\n\nIf you use use the BMC corpus, please cite the original paper as well as our paper which describes the splits:\n\n* Ben-Mordecai and Elhadad (2005):\n\n\n* Bareket and Tsarfaty (2020)" ]
c80a08ca133af5409d996361a5ae8fd57e2a3e38
# Dataset Card for Animal Crossing New Horizons Catalog ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://kaggle.com/datasets/jessicali9530/animal-crossing-new-horizons-nookplaza-dataset - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary ### Context This dataset comes from this [spreadsheet](https://tinyurl.com/acnh-sheet), a comprehensive Item Catalog for Animal Crossing New Horizons (ACNH). As described by [Wikipedia](https://en.wikipedia.org/wiki/Animal_Crossing:_New_Horizons), &gt; ACNH is a life simulation game released by Nintendo for Nintendo Switch on March 20, 2020. It is the fifth main series title in the Animal Crossing series and, with 5 million digital copies sold, has broken the record for Switch title with most digital units sold in a single month. In New Horizons, the player assumes the role of a customizable character who moves to a deserted island. Taking place in real-time, the player can explore the island in a nonlinear fashion, gathering and crafting items, catching insects and fish, and developing the island into a community of anthropomorphic animals. ### Content There are 30 csvs each listing various items, villagers, clothing, and other collectibles from the game. The data was collected by a dedicated group of AC fans who continue to collaborate and build this [spreadsheet](https://tinyurl.com/acnh-sheet) for public use. The database contains the original data and full list of contributors and raw data. At the time of writing, the only difference between the spreadsheet and this version is that the Kaggle version omits all columns with images of the items, but is otherwise identical. ### Acknowledgements Thanks to every contributor listed on the [spreadsheet!](https://tinyurl.com/acnh-sheet) Please attribute this spreadsheet and group for any use of the data. They also have a Discord server linked in the spreadsheet in case you want to contact them. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset was shared by [@jessicali9530](https://kaggle.com/jessicali9530) ### Licensing Information The license for this dataset is cc0-1.0 ### Citation Information ```bibtex [More Information Needed] ``` ### Contributions [More Information Needed]
osanseviero/kaggle-animal-crossing-new-horizons-nookplaza-dataset
[ "license:cc0-1.0", "region:us" ]
2022-06-22T14:45:17+00:00
{"license": ["cc0-1.0"], "kaggle_id": "jessicali9530/animal-crossing-new-horizons-nookplaza-dataset"}
2022-10-25T09:32:48+00:00
[]
[]
TAGS #license-cc0-1.0 #region-us
# Dataset Card for Animal Crossing New Horizons Catalog ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary ### Context This dataset comes from this spreadsheet, a comprehensive Item Catalog for Animal Crossing New Horizons (ACNH). As described by Wikipedia, &gt; ACNH is a life simulation game released by Nintendo for Nintendo Switch on March 20, 2020. It is the fifth main series title in the Animal Crossing series and, with 5 million digital copies sold, has broken the record for Switch title with most digital units sold in a single month. In New Horizons, the player assumes the role of a customizable character who moves to a deserted island. Taking place in real-time, the player can explore the island in a nonlinear fashion, gathering and crafting items, catching insects and fish, and developing the island into a community of anthropomorphic animals. ### Content There are 30 csvs each listing various items, villagers, clothing, and other collectibles from the game. The data was collected by a dedicated group of AC fans who continue to collaborate and build this spreadsheet for public use. The database contains the original data and full list of contributors and raw data. At the time of writing, the only difference between the spreadsheet and this version is that the Kaggle version omits all columns with images of the items, but is otherwise identical. ### Acknowledgements Thanks to every contributor listed on the spreadsheet! Please attribute this spreadsheet and group for any use of the data. They also have a Discord server linked in the spreadsheet in case you want to contact them. ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators This dataset was shared by @jessicali9530 ### Licensing Information The license for this dataset is cc0-1.0 ### Contributions
[ "# Dataset Card for Animal Crossing New Horizons Catalog", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Context\nThis dataset comes from this spreadsheet, a comprehensive Item Catalog for Animal Crossing New Horizons (ACNH). As described by Wikipedia,\n&gt; ACNH is a life simulation game released by Nintendo for Nintendo Switch on March 20, 2020. It is the fifth main series title in the Animal Crossing series and, with 5 million digital copies sold, has broken the record for Switch title with most digital units sold in a single month. In New Horizons, the player assumes the role of a customizable character who moves to a deserted island. Taking place in real-time, the player can explore the island in a nonlinear fashion, gathering and crafting items, catching insects and fish, and developing the island into a community of anthropomorphic animals.", "### Content\nThere are 30 csvs each listing various items, villagers, clothing, and other collectibles from the game. The data was collected by a dedicated group of AC fans who continue to collaborate and build this spreadsheet for public use. The database contains the original data and full list of contributors and raw data. At the time of writing, the only difference between the spreadsheet and this version is that the Kaggle version omits all columns with images of the items, but is otherwise identical.", "### Acknowledgements\nThanks to every contributor listed on the spreadsheet! Please attribute this spreadsheet and group for any use of the data. They also have a Discord server linked in the spreadsheet in case you want to contact them.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset was shared by @jessicali9530", "### Licensing Information\n\nThe license for this dataset is cc0-1.0", "### Contributions" ]
[ "TAGS\n#license-cc0-1.0 #region-us \n", "# Dataset Card for Animal Crossing New Horizons Catalog", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Context\nThis dataset comes from this spreadsheet, a comprehensive Item Catalog for Animal Crossing New Horizons (ACNH). As described by Wikipedia,\n&gt; ACNH is a life simulation game released by Nintendo for Nintendo Switch on March 20, 2020. It is the fifth main series title in the Animal Crossing series and, with 5 million digital copies sold, has broken the record for Switch title with most digital units sold in a single month. In New Horizons, the player assumes the role of a customizable character who moves to a deserted island. Taking place in real-time, the player can explore the island in a nonlinear fashion, gathering and crafting items, catching insects and fish, and developing the island into a community of anthropomorphic animals.", "### Content\nThere are 30 csvs each listing various items, villagers, clothing, and other collectibles from the game. The data was collected by a dedicated group of AC fans who continue to collaborate and build this spreadsheet for public use. The database contains the original data and full list of contributors and raw data. At the time of writing, the only difference between the spreadsheet and this version is that the Kaggle version omits all columns with images of the items, but is otherwise identical.", "### Acknowledgements\nThanks to every contributor listed on the spreadsheet! Please attribute this spreadsheet and group for any use of the data. They also have a Discord server linked in the spreadsheet in case you want to contact them.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset was shared by @jessicali9530", "### Licensing Information\n\nThe license for this dataset is cc0-1.0", "### Contributions" ]
934f9ae2fe4a4bacc3fa69d6a0aeefccf247377a
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655913671
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T15:01:14+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T15:01:18+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
6fc1cfe32c1b2f07ddbdbbf7c800826ee2781f12
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655913794
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T15:03:17+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T15:03:20+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
a7dcebed891356a4bf9963ca8519f7bef271698b
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655913835
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T15:03:58+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T15:04:02+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
19ce9b32b6f7a676acee224d2b986636e583f3d3
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655913900
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T15:05:03+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T15:05:06+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
0a1065290fa91ec7b57e3d5ebea57f985b0d106f
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655914374
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T15:12:57+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T15:13:01+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
9753139e0b9d454ab4fd22e884290260db5fc7b6
The legendary Titanic dataset from [this](https://www.kaggle.com/competitions/titanic/overview) Kaggle competition
phihung/titanic
[ "license:other", "region:us" ]
2022-06-22T15:16:15+00:00
{"license": "other"}
2022-06-22T15:25:32+00:00
[]
[]
TAGS #license-other #region-us
The legendary Titanic dataset from this Kaggle competition
[]
[ "TAGS\n#license-other #region-us \n" ]
31cc015b8ffbafc4168ccef186e3045b181deaf8
# Dataset Card for Object Detection for Chess Pieces ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/faizankshaikh/chessDetection - **Repository:** https://github.com/faizankshaikh/chessDetection - **Paper:** - - **Leaderboard:** - - **Point of Contact:** [Faizan Shaikh](mailto:[email protected]) ### Dataset Summary The "Object Detection for Chess Pieces" dataset is a toy dataset created (as suggested by the name!) to introduce object detection in a beginner friendly way. It is structured in a one object-one image manner, with the objects being of four classes, namely, Black King, White King, Black Queen and White Queen ### Supported Tasks and Leaderboards - `object-detection`: The dataset can be used to train and evaluate simplistic object detection models ### Languages The text (labels) in the dataset is in English ## Dataset Structure ### Data Instances A data point comprises an image and the corresponding objects in bounding boxes. ``` { 'image': <PIL.PngImagePlugin.PngImageFile image mode=RGB size=224x224 at 0x23557C66160>, 'objects': { "label": [ 0 ], "bbox": [ [ 151, 151, 26, 26 ] ] } } ``` ### Data Fields - `image`: A `PIL.Image.Image` object containing the 224x224 image. - `label`: An integer between 0 and 3 representing the classes with the following mapping: | Label | Description | | --- | --- | | 0 | blackKing | | 1 | blackQueen | | 2 | whiteKing | | 3 | whiteQueen | - `bbox`: A list of integers having sequence [x_center, y_center, width, height] for a particular bounding box ### Data Splits The data is split into training and validation set. The training set contains 204 images and the validation set 52 images. ## Dataset Creation ### Curation Rationale The dataset was created to be a simple benchmark for object detection ### Source Data #### Initial Data Collection and Normalization The data is obtained by machine generating images from "python-chess" library. Please refer [this code](https://github.com/faizankshaikh/chessDetection/blob/main/code/1.1%20create_images_with_labels.ipynb) to understand data generation pipeline #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process The annotations were done manually. #### Who are the annotators? The annotations were done manually. ### Personal and Sensitive Information None ## Considerations for Using the Data ### Social Impact of Dataset The dataset can be considered as a beginner-friendly toy dataset for object detection. It should not be used for benchmarking state of the art object detection models, or be used for a deployed model. ### Discussion of Biases [Needs More Information] ### Other Known Limitations The dataset only contains four classes for simplicity. The complexity can be increased by considering all types of chess pieces, and by making it a multi-object detection problem ## Additional Information ### Dataset Curators The dataset was created by Faizan Shaikh ### Licensing Information The dataset is licensed as CC-BY-SA:2.0 ### Citation Information [Needs More Information]
jalFaizy/detect_chess_pieces
[ "task_categories:object-detection", "annotations_creators:machine-generated", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:n<1K", "language:en", "license:other", "region:us" ]
2022-06-22T16:41:58+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": [], "task_categories": ["object-detection"], "task_ids": [], "pretty_name": "Object Detection for Chess Pieces"}
2022-10-25T09:34:41+00:00
[]
[ "en" ]
TAGS #task_categories-object-detection #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-n<1K #language-English #license-other #region-us
Dataset Card for Object Detection for Chess Pieces ================================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: - * Leaderboard: - * Point of Contact: Faizan Shaikh ### Dataset Summary The "Object Detection for Chess Pieces" dataset is a toy dataset created (as suggested by the name!) to introduce object detection in a beginner friendly way. It is structured in a one object-one image manner, with the objects being of four classes, namely, Black King, White King, Black Queen and White Queen ### Supported Tasks and Leaderboards * 'object-detection': The dataset can be used to train and evaluate simplistic object detection models ### Languages The text (labels) in the dataset is in English Dataset Structure ----------------- ### Data Instances A data point comprises an image and the corresponding objects in bounding boxes. ### Data Fields * 'image': A 'PIL.Image.Image' object containing the 224x224 image. * 'label': An integer between 0 and 3 representing the classes with the following mapping: * 'bbox': A list of integers having sequence [x\_center, y\_center, width, height] for a particular bounding box ### Data Splits The data is split into training and validation set. The training set contains 204 images and the validation set 52 images. Dataset Creation ---------------- ### Curation Rationale The dataset was created to be a simple benchmark for object detection ### Source Data #### Initial Data Collection and Normalization The data is obtained by machine generating images from "python-chess" library. Please refer this code to understand data generation pipeline #### Who are the source language producers? ### Annotations #### Annotation process The annotations were done manually. #### Who are the annotators? The annotations were done manually. ### Personal and Sensitive Information None Considerations for Using the Data --------------------------------- ### Social Impact of Dataset The dataset can be considered as a beginner-friendly toy dataset for object detection. It should not be used for benchmarking state of the art object detection models, or be used for a deployed model. ### Discussion of Biases ### Other Known Limitations The dataset only contains four classes for simplicity. The complexity can be increased by considering all types of chess pieces, and by making it a multi-object detection problem Additional Information ---------------------- ### Dataset Curators The dataset was created by Faizan Shaikh ### Licensing Information The dataset is licensed as CC-BY-SA:2.0
[ "### Dataset Summary\n\n\nThe \"Object Detection for Chess Pieces\" dataset is a toy dataset created (as suggested by the name!) to introduce object detection in a beginner friendly way. It is structured in a one object-one image manner, with the objects being of four classes, namely, Black King, White King, Black Queen and White Queen", "### Supported Tasks and Leaderboards\n\n\n* 'object-detection': The dataset can be used to train and evaluate simplistic object detection models", "### Languages\n\n\nThe text (labels) in the dataset is in English\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA data point comprises an image and the corresponding objects in bounding boxes.", "### Data Fields\n\n\n* 'image': A 'PIL.Image.Image' object containing the 224x224 image.\n* 'label': An integer between 0 and 3 representing the classes with the following mapping:\n* 'bbox': A list of integers having sequence [x\\_center, y\\_center, width, height] for a particular bounding box", "### Data Splits\n\n\nThe data is split into training and validation set. The training set contains 204 images and the validation set 52 images.\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe dataset was created to be a simple benchmark for object detection", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe data is obtained by machine generating images from \"python-chess\" library. Please refer this code to understand data generation pipeline", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nThe annotations were done manually.", "#### Who are the annotators?\n\n\nThe annotations were done manually.", "### Personal and Sensitive Information\n\n\nNone\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe dataset can be considered as a beginner-friendly toy dataset for object detection. It should not be used for benchmarking state of the art object detection models, or be used for a deployed model.", "### Discussion of Biases", "### Other Known Limitations\n\n\nThe dataset only contains four classes for simplicity. The complexity can be increased by considering all types of chess pieces, and by making it a multi-object detection problem\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Faizan Shaikh", "### Licensing Information\n\n\nThe dataset is licensed as CC-BY-SA:2.0" ]
[ "TAGS\n#task_categories-object-detection #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-n<1K #language-English #license-other #region-us \n", "### Dataset Summary\n\n\nThe \"Object Detection for Chess Pieces\" dataset is a toy dataset created (as suggested by the name!) to introduce object detection in a beginner friendly way. It is structured in a one object-one image manner, with the objects being of four classes, namely, Black King, White King, Black Queen and White Queen", "### Supported Tasks and Leaderboards\n\n\n* 'object-detection': The dataset can be used to train and evaluate simplistic object detection models", "### Languages\n\n\nThe text (labels) in the dataset is in English\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA data point comprises an image and the corresponding objects in bounding boxes.", "### Data Fields\n\n\n* 'image': A 'PIL.Image.Image' object containing the 224x224 image.\n* 'label': An integer between 0 and 3 representing the classes with the following mapping:\n* 'bbox': A list of integers having sequence [x\\_center, y\\_center, width, height] for a particular bounding box", "### Data Splits\n\n\nThe data is split into training and validation set. The training set contains 204 images and the validation set 52 images.\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe dataset was created to be a simple benchmark for object detection", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe data is obtained by machine generating images from \"python-chess\" library. Please refer this code to understand data generation pipeline", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nThe annotations were done manually.", "#### Who are the annotators?\n\n\nThe annotations were done manually.", "### Personal and Sensitive Information\n\n\nNone\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe dataset can be considered as a beginner-friendly toy dataset for object detection. It should not be used for benchmarking state of the art object detection models, or be used for a deployed model.", "### Discussion of Biases", "### Other Known Limitations\n\n\nThe dataset only contains four classes for simplicity. The complexity can be increased by considering all types of chess pieces, and by making it a multi-object detection problem\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Faizan Shaikh", "### Licensing Information\n\n\nThe dataset is licensed as CC-BY-SA:2.0" ]
944b156abfdad7627c3221b5ec4f6a6fb060a197
# Dataset Card for PartiPrompts (P2) ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://parti.research.google/ - **Repository:** https://github.com/google-research/parti - **Paper:** https://gweb-research-parti.web.app/parti_paper.pdf ### Dataset Summary PartiPrompts (P2) is a rich set of over 1600 prompts in English that we release as part of this work. P2 can be used to measure model capabilities across various categories and challenge aspects. ![parti-prompts](https://github.com/google-research/parti/blob/main/images/parti-prompts.png?raw=true) P2 prompts can be simple, allowing us to gauge the progress from scaling. They can also be complex, such as the following 67-word description we created for Vincent van Gogh’s *The Starry Night* (1889): *Oil-on-canvas painting of a blue night sky with roiling energy. A fuzzy and bright yellow crescent moon shining at the top. Below the exploding yellow stars and radiating swirls of blue, a distant village sits quietly on the right. Connecting earth and sky is a flame-like cypress tree with curling and swaying branches on the left. A church spire rises as a beacon over rolling blue hills.* ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text descriptions are in English. ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The license for this dataset is the apache-2.0 license. ### Citation Information [More Information Needed] ### Contributions Thanks to [@nateraw](https://github.com/nateraw) for adding this dataset.
nateraw/parti-prompts
[ "license:apache-2.0", "region:us" ]
2022-06-22T16:48:47+00:00
{"license": "apache-2.0"}
2022-06-22T18:17:49+00:00
[]
[]
TAGS #license-apache-2.0 #region-us
# Dataset Card for PartiPrompts (P2) ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: URL ### Dataset Summary PartiPrompts (P2) is a rich set of over 1600 prompts in English that we release as part of this work. P2 can be used to measure model capabilities across various categories and challenge aspects. !parti-prompts P2 prompts can be simple, allowing us to gauge the progress from scaling. They can also be complex, such as the following 67-word description we created for Vincent van Gogh’s *The Starry Night* (1889): *Oil-on-canvas painting of a blue night sky with roiling energy. A fuzzy and bright yellow crescent moon shining at the top. Below the exploding yellow stars and radiating swirls of blue, a distant village sits quietly on the right. Connecting earth and sky is a flame-like cypress tree with curling and swaying branches on the left. A church spire rises as a beacon over rolling blue hills.* ### Supported Tasks and Leaderboards ### Languages The text descriptions are in English. ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information The license for this dataset is the apache-2.0 license. ### Contributions Thanks to @nateraw for adding this dataset.
[ "# Dataset Card for PartiPrompts (P2)", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL", "### Dataset Summary\n\nPartiPrompts (P2) is a rich set of over 1600 prompts in English that we release\nas part of this work. P2 can be used to measure model capabilities across\nvarious categories and challenge aspects.\n\n!parti-prompts\n\nP2 prompts can be simple, allowing us to gauge the progress from scaling. They\ncan also be complex, such as the following 67-word description we created for\nVincent van Gogh’s *The Starry Night* (1889):\n\n*Oil-on-canvas painting of a blue night sky with roiling energy. A fuzzy and\nbright yellow crescent moon shining at the top. Below the exploding yellow stars\nand radiating swirls of blue, a distant village sits quietly on the right.\nConnecting earth and sky is a flame-like cypress tree with curling and swaying\nbranches on the left. A church spire rises as a beacon over rolling blue hills.*", "### Supported Tasks and Leaderboards", "### Languages\n\nThe text descriptions are in English.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe license for this dataset is the apache-2.0 license.", "### Contributions\n\nThanks to @nateraw for adding this dataset." ]
[ "TAGS\n#license-apache-2.0 #region-us \n", "# Dataset Card for PartiPrompts (P2)", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL", "### Dataset Summary\n\nPartiPrompts (P2) is a rich set of over 1600 prompts in English that we release\nas part of this work. P2 can be used to measure model capabilities across\nvarious categories and challenge aspects.\n\n!parti-prompts\n\nP2 prompts can be simple, allowing us to gauge the progress from scaling. They\ncan also be complex, such as the following 67-word description we created for\nVincent van Gogh’s *The Starry Night* (1889):\n\n*Oil-on-canvas painting of a blue night sky with roiling energy. A fuzzy and\nbright yellow crescent moon shining at the top. Below the exploding yellow stars\nand radiating swirls of blue, a distant village sits quietly on the right.\nConnecting earth and sky is a flame-like cypress tree with curling and swaying\nbranches on the left. A church spire rises as a beacon over rolling blue hills.*", "### Supported Tasks and Leaderboards", "### Languages\n\nThe text descriptions are in English.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe license for this dataset is the apache-2.0 license.", "### Contributions\n\nThanks to @nateraw for adding this dataset." ]
09cc9935c19b3bf29ed1dc1e46766f1d2363bfac
# Dataset Card for "nq" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
iohadrubin/nq
[ "region:us" ]
2022-06-22T18:51:42+00:00
{"configs": [{"config_name": "default", "data_files": [{"split": "validation", "path": "data/validation-*"}, {"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "dataset", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": "string"}, {"name": "positive_ctxs", "sequence": [{"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "score", "dtype": "float32"}, {"name": "title_score", "dtype": "int32"}, {"name": "passage_id", "dtype": "string"}]}, {"name": "negative_ctxs", "sequence": [{"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "score", "dtype": "float32"}, {"name": "title_score", "dtype": "int32"}, {"name": "passage_id", "dtype": "string"}]}, {"name": "hard_negative_ctxs", "sequence": [{"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "score", "dtype": "float32"}, {"name": "title_score", "dtype": "int32"}, {"name": "passage_id", "dtype": "string"}]}], "splits": [{"name": "validation", "num_bytes": 645475524, "num_examples": 6515}, {"name": "train", "num_bytes": 5836111764, "num_examples": 58880}], "download_size": 3923060242, "dataset_size": 6481587288}}
2024-01-04T01:36:06+00:00
[]
[]
TAGS #region-us
# Dataset Card for "nq" More Information needed
[ "# Dataset Card for \"nq\"\n\nMore Information needed" ]
[ "TAGS\n#region-us \n", "# Dataset Card for \"nq\"\n\nMore Information needed" ]
c112309ffcd4fb1a8f1567b2941be69bafd8ce24
# GEM Submission Submission name: This is a test name
GEM-submissions/lewtun__this-is-a-test-name__1655928558
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T19:09:21+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test name", "tags": ["evaluation", "benchmark"]}
2022-06-22T19:09:24+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test name
[ "# GEM Submission\n\nSubmission name: This is a test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test name" ]
fcc14d4bc7b2c7d4270ffe34355a62229dbb0838
Contains all TES3:Morrowind dialogues and journal queries. There are in total 4 labels: Journal, Greeting, Persuasion, Topic (Last one being the usual dialogues). The text is already formatted and does not contain duplicates or NaNs.
martosinc/morrowtext
[ "license:mit", "region:us" ]
2022-06-22T22:10:16+00:00
{"license": "mit"}
2022-06-22T22:17:49+00:00
[]
[]
TAGS #license-mit #region-us
Contains all TES3:Morrowind dialogues and journal queries. There are in total 4 labels: Journal, Greeting, Persuasion, Topic (Last one being the usual dialogues). The text is already formatted and does not contain duplicates or NaNs.
[]
[ "TAGS\n#license-mit #region-us \n" ]
3b988d737cc1358ca694149c628bdabe07275fb2
# AutoTrain Dataset for project: Wikipeida_Article_Classifier_by_Chap ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project Wikipeida_Article_Classifier_by_Chap. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "diffuse actinic keratinocyte dysplasia", "target": 15 }, { "text": "cholesterol atheroembolism", "target": 8 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=20, names=['Certain infectious or parasitic diseases', 'Developmental anaomalies', 'Diseases of the blood or blood forming organs', 'Diseases of the genitourinary system', 'Mental behavioural or neurodevelopmental disorders', 'Neoplasms', 'certain conditions originating in the perinatal period', 'conditions related to sexual health', 'diseases of the circulatroy system', 'diseases of the digestive system', 'diseases of the ear or mastoid process', 'diseases of the immune system', 'diseases of the musculoskeletal system or connective tissue', 'diseases of the nervous system', 'diseases of the respiratory system', 'diseases of the skin', 'diseases of the visual system', 'endocrine nutritional or metabolic diseases', 'pregnanacy childbirth or the puerperium', 'sleep-wake disorders'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 9828 | | valid | 2468 |
justpyschitry/autotrain-data-Wikipeida_Article_Classifier_by_Chap
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-06-23T01:13:39+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:34:57+00:00
[]
[ "en" ]
TAGS #task_categories-text-classification #language-English #region-us
AutoTrain Dataset for project: Wikipeida\_Article\_Classifier\_by\_Chap ======================================================================= Dataset Descritpion ------------------- This dataset has been automatically processed by AutoTrain for project Wikipeida\_Article\_Classifier\_by\_Chap. ### Languages The BCP-47 code for the dataset's language is en. Dataset Structure ----------------- ### Data Instances A sample from this dataset looks as follows: ### Dataset Fields The dataset has the following fields (also called "features"): ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow:
[ "### Languages\n\n\nThe BCP-47 code for the dataset's language is en.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA sample from this dataset looks as follows:", "### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):", "### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:" ]
[ "TAGS\n#task_categories-text-classification #language-English #region-us \n", "### Languages\n\n\nThe BCP-47 code for the dataset's language is en.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA sample from this dataset looks as follows:", "### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):", "### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:" ]
bcb86928d649893705003b9311a1170651e396ce
# GEM Submission Submission name: This is another test name
GEM-submissions/lewtun__this-is-another-test-name__1655982268
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T10:04:31+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is another test name", "tags": ["evaluation", "benchmark"]}
2022-06-23T10:04:35+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is another test name
[ "# GEM Submission\n\nSubmission name: This is another test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is another test name" ]
2e883f2ebd5e3c5178b114c1a7d65376d08f7294
# GEM Submission Submission name: This is another test name
GEM-submissions/lewtun__this-is-another-test-name__1655983106
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T10:18:29+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is another test name", "tags": ["evaluation", "benchmark"]}
2022-06-23T10:18:33+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is another test name
[ "# GEM Submission\n\nSubmission name: This is another test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is another test name" ]
b3d6c03c801f1ccabe8afeb5bd139904cee1b6b5
# GEM Submission Submission name: This is another test name
GEM-submissions/lewtun__this-is-another-test-name__1655983383
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T10:23:07+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is another test name", "tags": ["evaluation", "benchmark"]}
2022-06-23T10:23:10+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is another test name
[ "# GEM Submission\n\nSubmission name: This is another test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is another test name" ]
b2932abe00535d815b067005fe46064c5296fcb3
# GEM Submission Submission name: This is another test name
GEM-submissions/lewtun__this-is-another-test-name__1655985826
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T11:03:47+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is another test name", "tags": ["evaluation", "benchmark"]}
2022-06-23T11:03:51+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is another test name
[ "# GEM Submission\n\nSubmission name: This is another test name" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is another test name" ]
f2cb20374e200823a62809449f27dc2f0bebb289
## A Waiter's Tips The following description was retrieved from Kaggle page. Food servers’ tips in restaurants may be influenced by many factors, including the nature of the restaurant, size of the party, and table locations in the restaurant. Restaurant managers need to know which factors matter when they assign tables to food servers. For the sake of staff morale, they usually want to avoid either the substance or the appearance of unfair treatment of the servers, for whom tips (at least in restaurants in the United States) are a major component of pay. In one restaurant, a food server recorded the following data on all cus- tomers they served during an interval of two and a half months in early 1990. The restaurant, located in a suburban shopping mall, was part of a national chain and served a varied menu. In observance of local law, the restaurant offered to seat in a non-smoking section to patrons who requested it. Each record includes a day and time, and taken together, they show the server’s work schedule. **Acknowledgements** The data was reported in a collection of case studies for business statistics. Bryant, P. G. and Smith, M (1995) Practical Data Analysis: Case Studies in Business Statistics. Homewood, IL: Richard D. Irwin Publishing The dataset is also available through the Python package Seaborn.
scikit-learn/tips
[ "region:us" ]
2022-06-23T11:15:45+00:00
{}
2022-06-23T11:21:40+00:00
[]
[]
TAGS #region-us
## A Waiter's Tips The following description was retrieved from Kaggle page. Food servers’ tips in restaurants may be influenced by many factors, including the nature of the restaurant, size of the party, and table locations in the restaurant. Restaurant managers need to know which factors matter when they assign tables to food servers. For the sake of staff morale, they usually want to avoid either the substance or the appearance of unfair treatment of the servers, for whom tips (at least in restaurants in the United States) are a major component of pay. In one restaurant, a food server recorded the following data on all cus- tomers they served during an interval of two and a half months in early 1990. The restaurant, located in a suburban shopping mall, was part of a national chain and served a varied menu. In observance of local law, the restaurant offered to seat in a non-smoking section to patrons who requested it. Each record includes a day and time, and taken together, they show the server’s work schedule. Acknowledgements The data was reported in a collection of case studies for business statistics. Bryant, P. G. and Smith, M (1995) Practical Data Analysis: Case Studies in Business Statistics. Homewood, IL: Richard D. Irwin Publishing The dataset is also available through the Python package Seaborn.
[]
[ "TAGS\n#region-us \n" ]
ccb4d7c47eb82c25b865fb5052e998789b64d95f
Transparent
HekmatTaherinejad/Transparent
[ "region:us" ]
2022-06-23T12:19:14+00:00
{}
2022-06-24T07:45:10+00:00
[]
[]
TAGS #region-us
Transparent
[]
[ "TAGS\n#region-us \n" ]
c878972daa0a5ec5f0d684354b6c8018f27d1316
This dataset contains the subset of ArXiv papers with the "cs.LG" tag to indicate the paper is about Machine Learning. The core dataset is filtered from the full ArXiv dataset hosted on Kaggle: https://www.kaggle.com/datasets/Cornell-University/arxiv. The original dataset contains roughly 2 million papers. This dataset contains roughly 100,000 papers following the category filtering. The dataset is maintained by with requests to the ArXiv API. The current iteration of the dataset only contains the title and abstract of the paper. The ArXiv dataset contains additional features that we may look to include in future releases. We have highlighted the top two features on the roadmap for integration: <ul> <li> <b>authors</b> </li> <li> <b>update_date</b> </li> <li> Submitter </li> <li> Comments </li> <li> Journal-ref </li> <li> doi </li> <li> report-no </li> <li> categories </li> <li> license </li> <li> versions </li> <li> authors_parsed </li> </ul>
CShorten/ML-ArXiv-Papers
[ "license:afl-3.0", "region:us" ]
2022-06-23T13:31:39+00:00
{"license": "afl-3.0"}
2022-06-27T11:15:11+00:00
[]
[]
TAGS #license-afl-3.0 #region-us
This dataset contains the subset of ArXiv papers with the "cs.LG" tag to indicate the paper is about Machine Learning. The core dataset is filtered from the full ArXiv dataset hosted on Kaggle: URL The original dataset contains roughly 2 million papers. This dataset contains roughly 100,000 papers following the category filtering. The dataset is maintained by with requests to the ArXiv API. The current iteration of the dataset only contains the title and abstract of the paper. The ArXiv dataset contains additional features that we may look to include in future releases. We have highlighted the top two features on the roadmap for integration: <ul> <li> <b>authors</b> </li> <li> <b>update_date</b> </li> <li> Submitter </li> <li> Comments </li> <li> Journal-ref </li> <li> doi </li> <li> report-no </li> <li> categories </li> <li> license </li> <li> versions </li> <li> authors_parsed </li> </ul>
[]
[ "TAGS\n#license-afl-3.0 #region-us \n" ]
eab7d7ac5323841cee450a88e4edfd0d7f229fbb
# LibriS2S This repo contains scripts and alignment data to create a dataset build further upon [librivoxDeEn](https://www.cl.uni-heidelberg.de/statnlpgroup/librivoxdeen/) such that it contains (German audio, German transcription, English audio, English transcription) quadruplets and can be used for Speech-to-Speech translation research. Because of this, the alignments are released under the same [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-nc-sa/4.0/) <div> These alignments were collected by downloading the English audiobooks and using [aeneas](https://github.com/readbeyond/aeneas) to align the book chapters to the transcripts. For more information read the original [paper](https://arxiv.org/abs/2204.10593) (Presented at LREC 2022) ### The data The English/German audio are available in the folder EN/DE respectively and can be downloaded from [this onedrive](https://onedrive.live.com/embed?cid=DCE49ACC2BDA7D8C&resid=DCE49ACC2BDA7D8C%2115663&authkey=ANmUz8gRUoyxmjk). In case there are any problems with the download, feel free to open an issue here or on [GitHub](https://github.com/PedroDKE/LibriS2S). <br/> The repo structure is as follow: - Alignments : Contains all the alignments for each book and chapter - DE : Contains the German audio for each chapter per book. - EN : Contains the English audio for each chapter per book. - Example : contains example files on for the scraping and aligning explanations that were used to build this dataset. - LibrivoxDeEn_alignments : Contains the base alignments from the LibrivoxDeEn dataset. <br/> In case you feel a part of the data is missing, feel free to open an issue! The full zipfile is about 52 GB of size. ### Scraping a book from Librivox To download all chapters from a librivox url the following command can be used: ``` python scrape_audio_from_librivox.py \ --url https://librivox.org/undine-by-friedrich-de-la-motte-fouque/ \ --save_dir ./examples ``` ### Allign a book from Librivox with the text from LibrivoxDeEn To allign the previously downloaded book with the txt files and tsv tables provided by LibrivoxDeEn the following command, based on the example provided with this repo, can be used: ``` python align_text_and_audio.py \ --text_dir ./example/en_text/ \ --audio_path ./example/audio_chapters/ \ --aeneas_path ./example/aeneas/ \ --en_audio_export_path ./example/sentence_level_audio/ \ --total_alignment_path ./example/bi-lingual-alignment/ \ --librivoxdeen_alignment ./example/undine_data.tsv \ --aeneas_head_max 120 \ --aeneas_tail_min 5 \ ``` **note:** the example folder in this repo already contains the first two chapters from [Undine](https://librivox.org/undine-by-friedrich-de-la-motte-fouque/) scraped from librivox and their transcripts and (modified to only contain the first 2 chapters) tsv table retrieved from LibrivoxDeEn. Additional data to align can be scraped by using the same file shown previously and combined with the provided data from LibriVoxDeEn Additionally with this repo the full alignment for the 8 following books with following LibrivoxDeEn id's are also given: [9](https://librivox.org/the-picture-of-dorian-gray-1891-version-by-oscar-wilde/), [10](https://librivox.org/pandoras-box-by-frank-wedekind/), [13](https://librivox.org/survivors-of-the-chancellor-by-jules-verne/), [18](https://librivox.org/undine-by-friedrich-de-la-motte-fouque/), [23](https://librivox.org/around-the-world-in-80-days-by-jules-verne/), [108](https://librivox.org/elective-affinities-by-johann-wolfgang-von-goethe/), [110](https://librivox.org/candide-by-voltaire-3/), [120](https://librivox.org/the-metamorphosis-by-franz-kafka/). Other books such as [11](https://librivox.org/the-castle-of-otranto-by-horace-walpole/), [36](https://librivox.org/the-rider-on-the-white-horse-by-theodor-storm/), [67](https://librivox.org/frankenstein-or-the-modern-prometheus-1818-by-mary-wollstonecraft-shelley/) and [54](https://librivox.org/white-nights-other-stories-by-fyodor-dostoyevsky/) are also inside of the librivoxDeEn dataset but the chapters do not correspond in a 1:1 mannner(for example: the German version of book 67 has 27 chapters but the English version has 29 and thus need to be re-aligned before the allignment script in this repo will work). Therefore these alignments are given but might have be different if you scrape them yourselves as the re-alignments might be different for you. ### Metrics on the alignment given in this repo. Using the alignments given in this repo some metrics were collected and quickly displayed here, for this table and the next figure the books which were manually alligned, although provided in the zip, were not accounted for, but the full table can be found in the original paper. | | German | English | | :---: | :-: | :-: | |number of files | 18868 | 18868 | |total time (hh:mm:ss) | 39:11:08 | 40:52:31 | |Speakers | 41 |22 | note: the speakers were counted for each book seperatly so some speakers might be counter more than once. the number of hours for each book aligned in this repo:<br> <img src="https://user-images.githubusercontent.com/43861296/122250648-1f5f7f80-ceca-11eb-84fd-344a2261bf47.png" width="500"> when using this work, please cite the original paper and the LibrivoxDeEn authors ``` @inproceedings{jeuris-niehues-2022-libris2s, title = "{L}ibri{S}2{S}: A {G}erman-{E}nglish Speech-to-Speech Translation Corpus", author = "Jeuris, Pedro and Niehues, Jan", booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference", month = jun, year = "2022", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://aclanthology.org/2022.lrec-1.98", pages = "928--935", abstract = "Recently, we have seen an increasing interest in the area of speech-to-text translation. This has led to astonishing improvements in this area. In contrast, the activities in the area of speech-to-speech translation is still limited, although it is essential to overcome the language barrier. We believe that one of the limiting factors is the availability of appropriate training data. We address this issue by creating LibriS2S, to our knowledge the first publicly available speech-to-speech training corpus between German and English. For this corpus, we used independently created audio for German and English leading to an unbiased pronunciation of the text in both languages. This allows the creation of a new text-to-speech and speech-to-speech translation model that directly learns to generate the speech signal based on the pronunciation of the source language. Using this created corpus, we propose Text-to-Speech models based on the example of the recently proposed FastSpeech 2 model that integrates source language information. We do this by adapting the model to take information such as the pitch, energy or transcript from the source speech as additional input.", } ``` ``` @article{beilharz19, title = {LibriVoxDeEn: A Corpus for German-to-English Speech Translation and Speech Recognition}, author = {Beilharz, Benjamin and Sun, Xin and Karimova, Sariya and Riezler, Stefan}, journal = {Proceedings of the Language Resources and Evaluation Conference}, journal-abbrev = {LREC}, year = {2020}, city = {Marseille, France}, url = {https://arxiv.org/pdf/1910.07924.pdf} } ```
PedroDKE/LibriS2S
[ "task_categories:text-to-speech", "task_categories:automatic-speech-recognition", "task_categories:translation", "multilinguality:multilingual", "size_categories:10K<n<100K", "language:en", "language:de", "license:cc-by-nc-sa-4.0", "LibriS2S", "LibrivoxDeEn", "Speech-to-Speech translation", "LREC2022", "arxiv:2204.10593", "arxiv:1910.07924", "region:us" ]
2022-06-23T13:39:57+00:00
{"annotations_creators": [], "language_creators": [], "language": ["en", "de"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "source_datasets": [], "task_categories": ["text-to-speech", "automatic-speech-recognition", "translation"], "task_ids": [], "pretty_name": "LibriS2S German-English Speech and Text pairs", "tags": ["LibriS2S", "LibrivoxDeEn", "Speech-to-Speech translation", "LREC2022"]}
2023-03-23T13:28:39+00:00
[ "2204.10593", "1910.07924" ]
[ "en", "de" ]
TAGS #task_categories-text-to-speech #task_categories-automatic-speech-recognition #task_categories-translation #multilinguality-multilingual #size_categories-10K<n<100K #language-English #language-German #license-cc-by-nc-sa-4.0 #LibriS2S #LibrivoxDeEn #Speech-to-Speech translation #LREC2022 #arxiv-2204.10593 #arxiv-1910.07924 #region-us
LibriS2S ======== This repo contains scripts and alignment data to create a dataset build further upon librivoxDeEn such that it contains (German audio, German transcription, English audio, English transcription) quadruplets and can be used for Speech-to-Speech translation research. Because of this, the alignments are released under the same Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License These alignments were collected by downloading the English audiobooks and using aeneas to align the book chapters to the transcripts. For more information read the original paper (Presented at LREC 2022) ### The data The English/German audio are available in the folder EN/DE respectively and can be downloaded from this onedrive. In case there are any problems with the download, feel free to open an issue here or on GitHub. The repo structure is as follow: * Alignments : Contains all the alignments for each book and chapter * DE : Contains the German audio for each chapter per book. * EN : Contains the English audio for each chapter per book. * Example : contains example files on for the scraping and aligning explanations that were used to build this dataset. * LibrivoxDeEn\_alignments : Contains the base alignments from the LibrivoxDeEn dataset. In case you feel a part of the data is missing, feel free to open an issue! The full zipfile is about 52 GB of size. ### Scraping a book from Librivox To download all chapters from a librivox url the following command can be used: ### Allign a book from Librivox with the text from LibrivoxDeEn To allign the previously downloaded book with the txt files and tsv tables provided by LibrivoxDeEn the following command, based on the example provided with this repo, can be used: note: the example folder in this repo already contains the first two chapters from Undine scraped from librivox and their transcripts and (modified to only contain the first 2 chapters) tsv table retrieved from LibrivoxDeEn. Additional data to align can be scraped by using the same file shown previously and combined with the provided data from LibriVoxDeEn Additionally with this repo the full alignment for the 8 following books with following LibrivoxDeEn id's are also given: 9, 10, 13, 18, 23, 108, 110, 120. Other books such as 11, 36, 67 and 54 are also inside of the librivoxDeEn dataset but the chapters do not correspond in a 1:1 mannner(for example: the German version of book 67 has 27 chapters but the English version has 29 and thus need to be re-aligned before the allignment script in this repo will work). Therefore these alignments are given but might have be different if you scrape them yourselves as the re-alignments might be different for you. ### Metrics on the alignment given in this repo. Using the alignments given in this repo some metrics were collected and quickly displayed here, for this table and the next figure the books which were manually alligned, although provided in the zip, were not accounted for, but the full table can be found in the original paper. note: the speakers were counted for each book seperatly so some speakers might be counter more than once. the number of hours for each book aligned in this repo: <img src="URL width="500"> when using this work, please cite the original paper and the LibrivoxDeEn authors
[ "### The data\n\n\nThe English/German audio are available in the folder EN/DE respectively and can be downloaded from this onedrive. In case there are any problems with the download, feel free to open an issue here or on GitHub. \n\nThe repo structure is as follow:\n\n\n* Alignments : Contains all the alignments for each book and chapter\n* DE : Contains the German audio for each chapter per book.\n* EN : Contains the English audio for each chapter per book.\n* Example : contains example files on for the scraping and aligning explanations that were used to build this dataset.\n* LibrivoxDeEn\\_alignments : Contains the base alignments from the LibrivoxDeEn dataset.\n\n\nIn case you feel a part of the data is missing, feel free to open an issue!\nThe full zipfile is about 52 GB of size.", "### Scraping a book from Librivox\n\n\nTo download all chapters from a librivox url the following command can be used:", "### Allign a book from Librivox with the text from LibrivoxDeEn\n\n\nTo allign the previously downloaded book with the txt files and tsv tables provided by LibrivoxDeEn the following command, based on the example provided with this repo, can be used:\n\n\nnote: the example folder in this repo already contains the first two chapters from Undine scraped from librivox and their transcripts and (modified to only contain the first 2 chapters) tsv table retrieved from LibrivoxDeEn.\nAdditional data to align can be scraped by using the same file shown previously and combined with the provided data from LibriVoxDeEn\n\n\nAdditionally with this repo the full alignment for the 8 following books with following LibrivoxDeEn id's are also given:\n9, 10, 13, 18, 23, 108, 110, 120.\n\n\nOther books such as 11, 36, 67 and 54 are also inside of the librivoxDeEn dataset but the chapters do not correspond in a 1:1 mannner(for example: the German version of book 67 has 27 chapters but the English version has 29 and thus need to be re-aligned before the allignment script in this repo will work). Therefore these alignments are given but might have be different if you scrape them yourselves as the re-alignments might be different for you.", "### Metrics on the alignment given in this repo.\n\n\nUsing the alignments given in this repo some metrics were collected and quickly displayed here, for this table and the next figure the books which were manually alligned, although provided in the zip, were not accounted for, but the full table can be found in the original paper.\n\n\n\nnote: the speakers were counted for each book seperatly so some speakers might be counter more than once.\n\n\nthe number of hours for each book aligned in this repo: \n\n<img src=\"URL width=\"500\">\n\n\nwhen using this work, please cite the original paper and the LibrivoxDeEn authors" ]
[ "TAGS\n#task_categories-text-to-speech #task_categories-automatic-speech-recognition #task_categories-translation #multilinguality-multilingual #size_categories-10K<n<100K #language-English #language-German #license-cc-by-nc-sa-4.0 #LibriS2S #LibrivoxDeEn #Speech-to-Speech translation #LREC2022 #arxiv-2204.10593 #arxiv-1910.07924 #region-us \n", "### The data\n\n\nThe English/German audio are available in the folder EN/DE respectively and can be downloaded from this onedrive. In case there are any problems with the download, feel free to open an issue here or on GitHub. \n\nThe repo structure is as follow:\n\n\n* Alignments : Contains all the alignments for each book and chapter\n* DE : Contains the German audio for each chapter per book.\n* EN : Contains the English audio for each chapter per book.\n* Example : contains example files on for the scraping and aligning explanations that were used to build this dataset.\n* LibrivoxDeEn\\_alignments : Contains the base alignments from the LibrivoxDeEn dataset.\n\n\nIn case you feel a part of the data is missing, feel free to open an issue!\nThe full zipfile is about 52 GB of size.", "### Scraping a book from Librivox\n\n\nTo download all chapters from a librivox url the following command can be used:", "### Allign a book from Librivox with the text from LibrivoxDeEn\n\n\nTo allign the previously downloaded book with the txt files and tsv tables provided by LibrivoxDeEn the following command, based on the example provided with this repo, can be used:\n\n\nnote: the example folder in this repo already contains the first two chapters from Undine scraped from librivox and their transcripts and (modified to only contain the first 2 chapters) tsv table retrieved from LibrivoxDeEn.\nAdditional data to align can be scraped by using the same file shown previously and combined with the provided data from LibriVoxDeEn\n\n\nAdditionally with this repo the full alignment for the 8 following books with following LibrivoxDeEn id's are also given:\n9, 10, 13, 18, 23, 108, 110, 120.\n\n\nOther books such as 11, 36, 67 and 54 are also inside of the librivoxDeEn dataset but the chapters do not correspond in a 1:1 mannner(for example: the German version of book 67 has 27 chapters but the English version has 29 and thus need to be re-aligned before the allignment script in this repo will work). Therefore these alignments are given but might have be different if you scrape them yourselves as the re-alignments might be different for you.", "### Metrics on the alignment given in this repo.\n\n\nUsing the alignments given in this repo some metrics were collected and quickly displayed here, for this table and the next figure the books which were manually alligned, although provided in the zip, were not accounted for, but the full table can be found in the original paper.\n\n\n\nnote: the speakers were counted for each book seperatly so some speakers might be counter more than once.\n\n\nthe number of hours for each book aligned in this repo: \n\n<img src=\"URL width=\"500\">\n\n\nwhen using this work, please cite the original paper and the LibrivoxDeEn authors" ]
96a6c960623e1b4ad83b38f5e345c9c5632857f7
# Dataset Card for FEVEROUS ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://fever.ai/dataset/feverous.html - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [FEVEROUS: Fact Extraction and VERification Over Unstructured and Structured information](https://arxiv.org/abs/2106.05707) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary With billions of individual pages on the web providing information on almost every conceivable topic, we should have the ability to collect facts that answer almost every conceivable question. However, only a small fraction of this information is contained in structured sources (Wikidata, Freebase, etc.) – we are therefore limited by our ability to transform free-form text to structured knowledge. There is, however, another problem that has become the focus of a lot of recent research and media coverage: false information coming from unreliable sources. The FEVER workshops are a venue for work in verifiable knowledge extraction and to stimulate progress in this direction. FEVEROUS (Fact Extraction and VERification Over Unstructured and Structured information) is a fact verification dataset which consists of 87,026 verified claims. Each claim is annotated with evidence in the form of sentences and/or cells from tables in Wikipedia, as well as a label indicating whether this evidence supports, refutes, or does not provide enough information to reach a verdict. The dataset also contains annotation metadata such as annotator actions (query keywords, clicks on page, time signatures), and the type of challenge each claim poses. ### Supported Tasks and Leaderboards The task is verification of textual claims against textual sources. When compared to textual entailment (TE)/natural language inference, the key difference is that in these tasks the passage to verify each claim is given, and in recent years it typically consists a single sentence, while in verification systems it is retrieved from a large set of documents in order to form the evidence. ### Languages The dataset is in English (`en`). ## Dataset Structure ### Data Instances - **Size of downloaded dataset files:** 187.82 MB - **Size of the generated dataset:** 123.25 MB - **Total amount of disk used:** 311.07 MB An example of 'wikipedia_pages' looks as follows: ``` {'id': 24435, 'label': 1, 'claim': 'Michael Folivi competed with ten teams from 2016 to 2021, appearing in 54 games and making seven goals in total.', 'evidence': [{'content': ['Michael Folivi_cell_1_2_0', 'Michael Folivi_cell_1_7_0', 'Michael Folivi_cell_1_8_0', 'Michael Folivi_cell_1_9_0', 'Michael Folivi_cell_1_12_0'], 'context': [['Michael Folivi_title', 'Michael Folivi_section_4', 'Michael Folivi_header_cell_1_0_0'], ['Michael Folivi_title', 'Michael Folivi_section_4', 'Michael Folivi_header_cell_1_0_0'], ['Michael Folivi_title', 'Michael Folivi_section_4', 'Michael Folivi_header_cell_1_0_0'], ['Michael Folivi_title', 'Michael Folivi_section_4', 'Michael Folivi_header_cell_1_0_0'], ['Michael Folivi_title', 'Michael Folivi_section_4', 'Michael Folivi_header_cell_1_0_0']]}, {'content': ['Michael Folivi_cell_0_13_1', 'Michael Folivi_cell_0_14_1', 'Michael Folivi_cell_0_15_1', 'Michael Folivi_cell_0_16_1', 'Michael Folivi_cell_0_18_1'], 'context': [['Michael Folivi_title', 'Michael Folivi_header_cell_0_13_0', 'Michael Folivi_header_cell_0_11_0'], ['Michael Folivi_title', 'Michael Folivi_header_cell_0_14_0', 'Michael Folivi_header_cell_0_11_0'], ['Michael Folivi_title', 'Michael Folivi_header_cell_0_15_0', 'Michael Folivi_header_cell_0_11_0'], ['Michael Folivi_title', 'Michael Folivi_header_cell_0_16_0', 'Michael Folivi_header_cell_0_11_0'], ['Michael Folivi_title', 'Michael Folivi_header_cell_0_18_0', 'Michael Folivi_header_cell_0_11_0']]}], 'annotator_operations': [{'operation': 'start', 'value': 'start', 'time': 0.0}, {'operation': 'Now on', 'value': '?search=', 'time': 0.78}, {'operation': 'search', 'value': 'Michael Folivi', 'time': 78.101}, {'operation': 'Now on', 'value': 'Michael Folivi', 'time': 78.822}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_1_2_0', 'time': 96.202}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_1_7_0', 'time': 96.9}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_1_8_0', 'time': 97.429}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_1_9_0', 'time': 97.994}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_1_12_0', 'time': 99.02}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_13_1', 'time': 106.108}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_14_1', 'time': 106.702}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_15_1', 'time': 107.423}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_16_1', 'time': 108.186}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_17_1', 'time': 108.788}, {'operation': 'Highlighting', 'value': 'Michael Folivi_header_cell_0_17_0', 'time': 108.8}, {'operation': 'Highlighting', 'value': 'Michael Folivi_cell_0_18_1', 'time': 109.469}, {'operation': 'Highlighting deleted', 'value': 'Michael Folivi_cell_0_17_1', 'time': 124.28}, {'operation': 'Highlighting deleted', 'value': 'Michael Folivi_header_cell_0_17_0', 'time': 124.293}, {'operation': 'finish', 'value': 'finish', 'time': 141.351}], 'expected_challenge': '', 'challenge': 'Numerical Reasoning'} ``` ### Data Fields The data fields are the same among all splits. - `id` (int): ID of the sample. - `label` (ClassLabel): Annotated label for the claim. Can be one of {"SUPPORTS", "REFUTES", "NOT ENOUGH INFO"}. - `claim` (str): Text of the claim. - `evidence` (list of dict): Evidence sets (at maximum three). Each set consists of dictionaries with two fields: - `content` (list of str): List of element IDs serving as the evidence for the claim. Each element ID is in the format `"[PAGE ID]_[EVIDENCE TYPE]_[NUMBER ID]"`, where `[EVIDENCE TYPE]` can be: `sentence`, `cell`, `header_cell`, `table_caption`, `item`. - `context` (list of list of str): List (for each element ID in `content`) of a list of Wikipedia elements that are automatically associated with that element ID and serve as context. This includes an article's title, relevant sections (the section and sub-section(s) the element is located in), and for cells the closest row and column header (multiple row/column headers if they follow each other). - `annotator_operations` (list of dict): List of operations an annotator used to find the evidence and reach a verdict, given the claim. Each element in the list is a dictionary with the fields: - `operation` (str): Operation name. Any of the following: - `start`, `finish`: Annotation started/finished. The value is the name of the operation. - `search`: Annotator used the Wikipedia search function. The value is the entered search term or the term selected from the automatic suggestions. If the annotator did not select any of the suggestions but instead went into advanced search, the term is prefixed with "contains...". - `hyperlink`: Annotator clicked on a hyperlink in the page. The value is the anchor text of the hyperlink. - `Now on`: The page the annotator has landed after a search or a hyperlink click. The value is the PAGE ID. - `Page search`: Annotator search on a page. The value is the search term. - `page-search-reset`: Annotator cleared the search box. The value is the name of the operation. - `Highlighting`, `Highlighting deleted`: Annotator selected/unselected an element on the page. The value is `ELEMENT ID`. - `back-button-clicked`: Annotator pressed the back button. The value is the name of the operation. - `value` (str): Value associated with the operation. - `time` (float): Time in seconds from the start of the annotation. - `expected_challenge` (str): The challenge the claim generator selected will be faced when verifying the claim, one out of the following: `Numerical Reasoning`, `Multi-hop Reasoning`, `Entity Disambiguation`, `Combining Tables and Text`, `Search terms not in claim`, `Other`. - `challenge` (str): Main challenge to verify the claim, one out of the following: `Numerical Reasoning`, `Multi-hop Reasoning`, `Entity Disambiguation`, `Combining Tables and Text`, `Search terms not in claim`, `Other`. ### Data Splits | | train | validation | test | |--------------------|------:|-----------:|-----:| | Number of examples | 71291 | 7890 | 7845 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information ``` These data annotations incorporate material from Wikipedia, which is licensed pursuant to the Wikipedia Copyright Policy. These annotations are made available under the license terms described on the applicable Wikipedia article pages, or, where Wikipedia license terms are unavailable, under the Creative Commons Attribution-ShareAlike License (version 3.0), available at http://creativecommons.org/licenses/by-sa/3.0/ (collectively, the “License Terms”). You may not use these files except in compliance with the applicable License Terms. ``` ### Citation Information If you use this dataset, please cite: ```bibtex @inproceedings{Aly21Feverous, author = {Aly, Rami and Guo, Zhijiang and Schlichtkrull, Michael Sejr and Thorne, James and Vlachos, Andreas and Christodoulopoulos, Christos and Cocarascu, Oana and Mittal, Arpit}, title = {{FEVEROUS}: Fact Extraction and {VERification} Over Unstructured and Structured information}, eprint={2106.05707}, archivePrefix={arXiv}, primaryClass={cs.CL}, year = {2021} } ``` ### Contributions Thanks to [@albertvillanova](https://github.com/albertvillanova) for adding this dataset.
fever/feverous
[ "task_categories:text-classification", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|wikipedia", "language:en", "license:cc-by-sa-3.0", "knowledge-verification", "arxiv:2106.05707", "region:us" ]
2022-06-23T13:46:02+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|wikipedia"], "task_categories": ["text-classification"], "task_ids": [], "paperswithcode_id": "feverous", "pretty_name": "FEVEROUS", "tags": ["knowledge-verification"]}
2022-10-25T04:50:36+00:00
[ "2106.05707" ]
[ "en" ]
TAGS #task_categories-text-classification #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|wikipedia #language-English #license-cc-by-sa-3.0 #knowledge-verification #arxiv-2106.05707 #region-us
Dataset Card for FEVEROUS ========================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: FEVEROUS: Fact Extraction and VERification Over Unstructured and Structured information * Point of Contact: ### Dataset Summary With billions of individual pages on the web providing information on almost every conceivable topic, we should have the ability to collect facts that answer almost every conceivable question. However, only a small fraction of this information is contained in structured sources (Wikidata, Freebase, etc.) – we are therefore limited by our ability to transform free-form text to structured knowledge. There is, however, another problem that has become the focus of a lot of recent research and media coverage: false information coming from unreliable sources. The FEVER workshops are a venue for work in verifiable knowledge extraction and to stimulate progress in this direction. FEVEROUS (Fact Extraction and VERification Over Unstructured and Structured information) is a fact verification dataset which consists of 87,026 verified claims. Each claim is annotated with evidence in the form of sentences and/or cells from tables in Wikipedia, as well as a label indicating whether this evidence supports, refutes, or does not provide enough information to reach a verdict. The dataset also contains annotation metadata such as annotator actions (query keywords, clicks on page, time signatures), and the type of challenge each claim poses. ### Supported Tasks and Leaderboards The task is verification of textual claims against textual sources. When compared to textual entailment (TE)/natural language inference, the key difference is that in these tasks the passage to verify each claim is given, and in recent years it typically consists a single sentence, while in verification systems it is retrieved from a large set of documents in order to form the evidence. ### Languages The dataset is in English ('en'). Dataset Structure ----------------- ### Data Instances * Size of downloaded dataset files: 187.82 MB * Size of the generated dataset: 123.25 MB * Total amount of disk used: 311.07 MB An example of 'wikipedia\_pages' looks as follows: ### Data Fields The data fields are the same among all splits. * 'id' (int): ID of the sample. * 'label' (ClassLabel): Annotated label for the claim. Can be one of {"SUPPORTS", "REFUTES", "NOT ENOUGH INFO"}. * 'claim' (str): Text of the claim. * 'evidence' (list of dict): Evidence sets (at maximum three). Each set consists of dictionaries with two fields: + 'content' (list of str): List of element IDs serving as the evidence for the claim. Each element ID is in the format '"[PAGE ID]*[EVIDENCE TYPE]*[NUMBER ID]"', where '[EVIDENCE TYPE]' can be: 'sentence', 'cell', 'header\_cell', 'table\_caption', 'item'. + 'context' (list of list of str): List (for each element ID in 'content') of a list of Wikipedia elements that are automatically associated with that element ID and serve as context. This includes an article's title, relevant sections (the section and sub-section(s) the element is located in), and for cells the closest row and column header (multiple row/column headers if they follow each other). * 'annotator\_operations' (list of dict): List of operations an annotator used to find the evidence and reach a verdict, given the claim. Each element in the list is a dictionary with the fields: + 'operation' (str): Operation name. Any of the following: - 'start', 'finish': Annotation started/finished. The value is the name of the operation. - 'search': Annotator used the Wikipedia search function. The value is the entered search term or the term selected from the automatic suggestions. If the annotator did not select any of the suggestions but instead went into advanced search, the term is prefixed with "contains...". - 'hyperlink': Annotator clicked on a hyperlink in the page. The value is the anchor text of the hyperlink. - 'Now on': The page the annotator has landed after a search or a hyperlink click. The value is the PAGE ID. - 'Page search': Annotator search on a page. The value is the search term. - 'page-search-reset': Annotator cleared the search box. The value is the name of the operation. - 'Highlighting', 'Highlighting deleted': Annotator selected/unselected an element on the page. The value is 'ELEMENT ID'. - 'back-button-clicked': Annotator pressed the back button. The value is the name of the operation. + 'value' (str): Value associated with the operation. + 'time' (float): Time in seconds from the start of the annotation. * 'expected\_challenge' (str): The challenge the claim generator selected will be faced when verifying the claim, one out of the following: 'Numerical Reasoning', 'Multi-hop Reasoning', 'Entity Disambiguation', 'Combining Tables and Text', 'Search terms not in claim', 'Other'. * 'challenge' (str): Main challenge to verify the claim, one out of the following: 'Numerical Reasoning', 'Multi-hop Reasoning', 'Entity Disambiguation', 'Combining Tables and Text', 'Search terms not in claim', 'Other'. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information If you use this dataset, please cite: ### Contributions Thanks to @albertvillanova for adding this dataset.
[ "### Dataset Summary\n\n\nWith billions of individual pages on the web providing information on almost every conceivable topic, we should have\nthe ability to collect facts that answer almost every conceivable question. However, only a small fraction of this\ninformation is contained in structured sources (Wikidata, Freebase, etc.) – we are therefore limited by our ability to\ntransform free-form text to structured knowledge. There is, however, another problem that has become the focus of a lot\nof recent research and media coverage: false information coming from unreliable sources.\n\n\nThe FEVER workshops are a venue for work in verifiable knowledge extraction and to stimulate progress in this direction.\n\n\nFEVEROUS (Fact Extraction and VERification Over Unstructured and Structured information) is a fact\nverification dataset which consists of 87,026 verified claims. Each claim is annotated with evidence in the form of\nsentences and/or cells from tables in Wikipedia, as well as a label indicating whether this evidence supports, refutes,\nor does not provide enough information to reach a verdict. The dataset also contains annotation metadata such as\nannotator actions (query keywords, clicks on page, time signatures), and the type of challenge each claim poses.", "### Supported Tasks and Leaderboards\n\n\nThe task is verification of textual claims against textual sources.\n\n\nWhen compared to textual entailment (TE)/natural language inference, the key difference is that in these tasks the\npassage to verify each claim is given, and in recent years it typically consists a single sentence, while in\nverification systems it is retrieved from a large set of documents in order to form the evidence.", "### Languages\n\n\nThe dataset is in English ('en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\n* Size of downloaded dataset files: 187.82 MB\n* Size of the generated dataset: 123.25 MB\n* Total amount of disk used: 311.07 MB\n\n\nAn example of 'wikipedia\\_pages' looks as follows:", "### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'id' (int): ID of the sample.\n* 'label' (ClassLabel): Annotated label for the claim. Can be one of {\"SUPPORTS\", \"REFUTES\", \"NOT ENOUGH INFO\"}.\n* 'claim' (str): Text of the claim.\n* 'evidence' (list of dict): Evidence sets (at maximum three). Each set consists of dictionaries with two fields:\n\t+ 'content' (list of str): List of element IDs serving as the evidence for the claim. Each element ID is in the format\n\t'\"[PAGE ID]*[EVIDENCE TYPE]*[NUMBER ID]\"', where '[EVIDENCE TYPE]' can be: 'sentence', 'cell', 'header\\_cell',\n\t'table\\_caption', 'item'.\n\t+ 'context' (list of list of str): List (for each element ID in 'content') of a list of Wikipedia elements that are\n\tautomatically associated with that element ID and serve as context. This includes an article's title, relevant\n\tsections (the section and sub-section(s) the element is located in), and for cells the closest row and column\n\theader (multiple row/column headers if they follow each other).\n* 'annotator\\_operations' (list of dict): List of operations an annotator used to find the evidence and reach a verdict,\ngiven the claim. Each element in the list is a dictionary with the fields:\n\t+ 'operation' (str): Operation name. Any of the following:\n\t\t- 'start', 'finish': Annotation started/finished. The value is the name of the operation.\n\t\t- 'search': Annotator used the Wikipedia search function. The value is the entered search term or the term selected\n\t\tfrom the automatic suggestions. If the annotator did not select any of the suggestions but instead went into\n\t\tadvanced search, the term is prefixed with \"contains...\".\n\t\t- 'hyperlink': Annotator clicked on a hyperlink in the page. The value is the anchor text of the hyperlink.\n\t\t- 'Now on': The page the annotator has landed after a search or a hyperlink click. The value is the PAGE ID.\n\t\t- 'Page search': Annotator search on a page. The value is the search term.\n\t\t- 'page-search-reset': Annotator cleared the search box. The value is the name of the operation.\n\t\t- 'Highlighting', 'Highlighting deleted': Annotator selected/unselected an element on the page. The value is\n\t\t'ELEMENT ID'.\n\t\t- 'back-button-clicked': Annotator pressed the back button. The value is the name of the operation.\n\t+ 'value' (str): Value associated with the operation.\n\t+ 'time' (float): Time in seconds from the start of the annotation.\n* 'expected\\_challenge' (str): The challenge the claim generator selected will be faced when verifying the claim, one\nout of the following: 'Numerical Reasoning', 'Multi-hop Reasoning', 'Entity Disambiguation',\n'Combining Tables and Text', 'Search terms not in claim', 'Other'.\n* 'challenge' (str): Main challenge to verify the claim, one out of the following: 'Numerical Reasoning',\n'Multi-hop Reasoning', 'Entity Disambiguation', 'Combining Tables and Text', 'Search terms not in claim', 'Other'.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nIf you use this dataset, please cite:", "### Contributions\n\n\nThanks to @albertvillanova for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|wikipedia #language-English #license-cc-by-sa-3.0 #knowledge-verification #arxiv-2106.05707 #region-us \n", "### Dataset Summary\n\n\nWith billions of individual pages on the web providing information on almost every conceivable topic, we should have\nthe ability to collect facts that answer almost every conceivable question. However, only a small fraction of this\ninformation is contained in structured sources (Wikidata, Freebase, etc.) – we are therefore limited by our ability to\ntransform free-form text to structured knowledge. There is, however, another problem that has become the focus of a lot\nof recent research and media coverage: false information coming from unreliable sources.\n\n\nThe FEVER workshops are a venue for work in verifiable knowledge extraction and to stimulate progress in this direction.\n\n\nFEVEROUS (Fact Extraction and VERification Over Unstructured and Structured information) is a fact\nverification dataset which consists of 87,026 verified claims. Each claim is annotated with evidence in the form of\nsentences and/or cells from tables in Wikipedia, as well as a label indicating whether this evidence supports, refutes,\nor does not provide enough information to reach a verdict. The dataset also contains annotation metadata such as\nannotator actions (query keywords, clicks on page, time signatures), and the type of challenge each claim poses.", "### Supported Tasks and Leaderboards\n\n\nThe task is verification of textual claims against textual sources.\n\n\nWhen compared to textual entailment (TE)/natural language inference, the key difference is that in these tasks the\npassage to verify each claim is given, and in recent years it typically consists a single sentence, while in\nverification systems it is retrieved from a large set of documents in order to form the evidence.", "### Languages\n\n\nThe dataset is in English ('en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\n* Size of downloaded dataset files: 187.82 MB\n* Size of the generated dataset: 123.25 MB\n* Total amount of disk used: 311.07 MB\n\n\nAn example of 'wikipedia\\_pages' looks as follows:", "### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'id' (int): ID of the sample.\n* 'label' (ClassLabel): Annotated label for the claim. Can be one of {\"SUPPORTS\", \"REFUTES\", \"NOT ENOUGH INFO\"}.\n* 'claim' (str): Text of the claim.\n* 'evidence' (list of dict): Evidence sets (at maximum three). Each set consists of dictionaries with two fields:\n\t+ 'content' (list of str): List of element IDs serving as the evidence for the claim. Each element ID is in the format\n\t'\"[PAGE ID]*[EVIDENCE TYPE]*[NUMBER ID]\"', where '[EVIDENCE TYPE]' can be: 'sentence', 'cell', 'header\\_cell',\n\t'table\\_caption', 'item'.\n\t+ 'context' (list of list of str): List (for each element ID in 'content') of a list of Wikipedia elements that are\n\tautomatically associated with that element ID and serve as context. This includes an article's title, relevant\n\tsections (the section and sub-section(s) the element is located in), and for cells the closest row and column\n\theader (multiple row/column headers if they follow each other).\n* 'annotator\\_operations' (list of dict): List of operations an annotator used to find the evidence and reach a verdict,\ngiven the claim. Each element in the list is a dictionary with the fields:\n\t+ 'operation' (str): Operation name. Any of the following:\n\t\t- 'start', 'finish': Annotation started/finished. The value is the name of the operation.\n\t\t- 'search': Annotator used the Wikipedia search function. The value is the entered search term or the term selected\n\t\tfrom the automatic suggestions. If the annotator did not select any of the suggestions but instead went into\n\t\tadvanced search, the term is prefixed with \"contains...\".\n\t\t- 'hyperlink': Annotator clicked on a hyperlink in the page. The value is the anchor text of the hyperlink.\n\t\t- 'Now on': The page the annotator has landed after a search or a hyperlink click. The value is the PAGE ID.\n\t\t- 'Page search': Annotator search on a page. The value is the search term.\n\t\t- 'page-search-reset': Annotator cleared the search box. The value is the name of the operation.\n\t\t- 'Highlighting', 'Highlighting deleted': Annotator selected/unselected an element on the page. The value is\n\t\t'ELEMENT ID'.\n\t\t- 'back-button-clicked': Annotator pressed the back button. The value is the name of the operation.\n\t+ 'value' (str): Value associated with the operation.\n\t+ 'time' (float): Time in seconds from the start of the annotation.\n* 'expected\\_challenge' (str): The challenge the claim generator selected will be faced when verifying the claim, one\nout of the following: 'Numerical Reasoning', 'Multi-hop Reasoning', 'Entity Disambiguation',\n'Combining Tables and Text', 'Search terms not in claim', 'Other'.\n* 'challenge' (str): Main challenge to verify the claim, one out of the following: 'Numerical Reasoning',\n'Multi-hop Reasoning', 'Entity Disambiguation', 'Combining Tables and Text', 'Search terms not in claim', 'Other'.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nIf you use this dataset, please cite:", "### Contributions\n\n\nThanks to @albertvillanova for adding this dataset." ]
9fed01dc26550f64d98c348f91c827de4417b5da
ConcurrentQA is a textual multi-hop QA benchmark to require concurrent retrieval over multiple data-distributions (i.e. Wikipedia and email data). This dataset was constructed by researchers at Stanford and FAIR, following the data collection process and schema of HotpotQA. This benchmark can be used to study generalization in retrieval as well as privacy when reasoning across multiple privacy scopes --- i.e. public Wikipedia documents and private emails. This dataset is for the Question-Answering task. The dataset for the Retrieval task can be found here: https://huggingface.co/datasets/simarora/ConcurrentQA-Retrieval The corpora of documents (Wikipedia and Emails) over which a system would need to retrieve information and answer questions can be downloaded using the following commands: ``` cd .. mkdir corpora cd corpora wget https://dl.fbaipublicfiles.com/concurrentqa/corpora/enron_only_corpus.json wget https://dl.fbaipublicfiles.com/concurrentqa/corpora/combined_corpus.json wget https://dl.fbaipublicfiles.com/concurrentqa/corpora/wiki_only_corpus.json wget https://dl.fbaipublicfiles.com/concurrentqa/corpora/title2sent_map.json ``` The repo https://github.com/facebookresearch/concurrentqa contains model training and result analysis code. If you find this resource useful, consider citing the paper: ``` @article{arora2023reasoning, title={Reasoning over Public and Private Data in Retrieval-Based Systems}, author={Simran Arora and Patrick Lewis and Angela Fan and Jacob Kahn and Christopher Ré}, year={2023}, url={https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00556/116046/Aggretriever-A-Simple-Approach-to-Aggregate}, journal={Transactions of the Association for Computational Linguistics}, } ``` Please reach out at ```[email protected]``` with questions or feedback!
simarora/ConcurrentQA
[ "task_categories:question-answering", "language:en", "license:mit", "region:us" ]
2022-06-23T18:21:23+00:00
{"language": ["en"], "license": "mit", "task_categories": ["question-answering"]}
2024-01-12T09:04:14+00:00
[]
[ "en" ]
TAGS #task_categories-question-answering #language-English #license-mit #region-us
ConcurrentQA is a textual multi-hop QA benchmark to require concurrent retrieval over multiple data-distributions (i.e. Wikipedia and email data). This dataset was constructed by researchers at Stanford and FAIR, following the data collection process and schema of HotpotQA. This benchmark can be used to study generalization in retrieval as well as privacy when reasoning across multiple privacy scopes --- i.e. public Wikipedia documents and private emails. This dataset is for the Question-Answering task. The dataset for the Retrieval task can be found here: URL The corpora of documents (Wikipedia and Emails) over which a system would need to retrieve information and answer questions can be downloaded using the following commands: The repo URL contains model training and result analysis code. If you find this resource useful, consider citing the paper: Please reach out at with questions or feedback!
[]
[ "TAGS\n#task_categories-question-answering #language-English #license-mit #region-us \n" ]
4e3152110827c8b80f508b1b02677a043756441a
# GEM Submission Submission name: This is a test submission
GEM-submissions/lewtun__this-is-a-test-submission__1656013291
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T18:41:33+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test submission", "tags": ["evaluation", "benchmark"]}
2022-06-23T18:41:36+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test submission
[ "# GEM Submission\n\nSubmission name: This is a test submission" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test submission" ]
c06f6ad845a32812535e6ecef534efd6342dacfb
# GEM Submission Submission name: This is a test submission 1
GEM-submissions/lewtun__this-is-a-test-submission-1__1656014763
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-23T19:06:05+00:00
{"benchmark": "gem", "type": "prediction", "submission_name": "This is a test submission 1", "tags": ["evaluation", "benchmark"]}
2022-06-23T19:06:09+00:00
[]
[]
TAGS #benchmark-gem #evaluation #benchmark #region-us
# GEM Submission Submission name: This is a test submission 1
[ "# GEM Submission\n\nSubmission name: This is a test submission 1" ]
[ "TAGS\n#benchmark-gem #evaluation #benchmark #region-us \n", "# GEM Submission\n\nSubmission name: This is a test submission 1" ]
64179b8f08613459a2265125c29d5290e41baac1
**Date**: 2022-07-10<br/> **Files**: ner_dataset.csv<br/> **Source**: [Kaggle entity annotated corpus](https://www.kaggle.com/datasets/abhinavwalia95/entity-annotated-corpus)<br/> **notes**: The dataset only contains the tokens and ner tag labels. Labels are uppercase. # About Dataset [**from Kaggle Datasets**](https://www.kaggle.com/datasets/abhinavwalia95/entity-annotated-corpus) ## Context Annotated Corpus for Named Entity Recognition using GMB(Groningen Meaning Bank) corpus for entity classification with enhanced and popular features by Natural Language Processing applied to the data set. Tip: Use Pandas Dataframe to load dataset if using Python for convenience. ## Content This is the extract from GMB corpus which is tagged, annotated and built specifically to train the classifier to predict named entities such as name, location, etc. Number of tagged entities: 'O': 1146068', geo-nam': 58388, 'org-nam': 48034, 'per-nam': 23790, 'gpe-nam': 20680, 'tim-dat': 12786, 'tim-dow': 11404, 'per-tit': 9800, 'per-fam': 8152, 'tim-yoc': 5290, 'tim-moy': 4262, 'per-giv': 2413, 'tim-clo': 891, 'art-nam': 866, 'eve-nam': 602, 'nat-nam': 300, 'tim-nam': 146, 'eve-ord': 107, 'per-ini': 60, 'org-leg': 60, 'per-ord': 38, 'tim-dom': 10, 'per-mid': 1, 'art-add': 1 ## Essential info about entities * geo = Geographical Entity * org = Organization * per = Person * gpe = Geopolitical Entity * tim = Time indicator * art = Artifact * eve = Event * nat = Natural Phenomenon * Total Words Count = 1354149 * Target Data Column: "tag" (ner_tag in this repo) Inspiration: This dataset is getting more interested because of more features added to the recent version of this dataset. Also, it helps to create a broad view of Feature Engineering with respect to this dataset. ## Modifications the ner_dataset.csv was modified to have a similar data Structure as [CoNLL-2003 dataset](https://huggingface.co/datasets/conll2003) ## Licensing information Database: Open Database, Contents: Database Contents.
rjac/kaggle-entity-annotated-corpus-ner-dataset
[ "annotations_creators:Abhinav Walia (Owner)", "language:en", "license:odbl", "region:us" ]
2022-06-23T19:31:55+00:00
{"annotations_creators": ["Abhinav Walia (Owner)"], "language": ["en"], "license": ["odbl"]}
2022-10-25T09:37:24+00:00
[]
[ "en" ]
TAGS #annotations_creators-Abhinav Walia (Owner) #language-English #license-odbl #region-us
Date: 2022-07-10<br/> Files: ner_dataset.csv<br/> Source: Kaggle entity annotated corpus<br/> notes: The dataset only contains the tokens and ner tag labels. Labels are uppercase. # About Dataset from Kaggle Datasets ## Context Annotated Corpus for Named Entity Recognition using GMB(Groningen Meaning Bank) corpus for entity classification with enhanced and popular features by Natural Language Processing applied to the data set. Tip: Use Pandas Dataframe to load dataset if using Python for convenience. ## Content This is the extract from GMB corpus which is tagged, annotated and built specifically to train the classifier to predict named entities such as name, location, etc. Number of tagged entities: 'O': 1146068', geo-nam': 58388, 'org-nam': 48034, 'per-nam': 23790, 'gpe-nam': 20680, 'tim-dat': 12786, 'tim-dow': 11404, 'per-tit': 9800, 'per-fam': 8152, 'tim-yoc': 5290, 'tim-moy': 4262, 'per-giv': 2413, 'tim-clo': 891, 'art-nam': 866, 'eve-nam': 602, 'nat-nam': 300, 'tim-nam': 146, 'eve-ord': 107, 'per-ini': 60, 'org-leg': 60, 'per-ord': 38, 'tim-dom': 10, 'per-mid': 1, 'art-add': 1 ## Essential info about entities * geo = Geographical Entity * org = Organization * per = Person * gpe = Geopolitical Entity * tim = Time indicator * art = Artifact * eve = Event * nat = Natural Phenomenon * Total Words Count = 1354149 * Target Data Column: "tag" (ner_tag in this repo) Inspiration: This dataset is getting more interested because of more features added to the recent version of this dataset. Also, it helps to create a broad view of Feature Engineering with respect to this dataset. ## Modifications the ner_dataset.csv was modified to have a similar data Structure as CoNLL-2003 dataset ## Licensing information Database: Open Database, Contents: Database Contents.
[ "# About Dataset\nfrom Kaggle Datasets", "## Context\n\nAnnotated Corpus for Named Entity Recognition using GMB(Groningen Meaning Bank) corpus for entity classification with enhanced and popular features by Natural Language Processing applied to the data set.\n\nTip: Use Pandas Dataframe to load dataset if using Python for convenience.", "## Content\nThis is the extract from GMB corpus which is tagged, annotated and built specifically to train the classifier to predict named entities such as name, location, etc.\n\nNumber of tagged entities:\n\n'O': 1146068', geo-nam': 58388, 'org-nam': 48034, 'per-nam': 23790, 'gpe-nam': 20680, 'tim-dat': 12786, 'tim-dow': 11404, 'per-tit': 9800, 'per-fam': 8152, 'tim-yoc': 5290, 'tim-moy': 4262, 'per-giv': 2413, 'tim-clo': 891, 'art-nam': 866, 'eve-nam': 602, 'nat-nam': 300, 'tim-nam': 146, 'eve-ord': 107, 'per-ini': 60, 'org-leg': 60, 'per-ord': 38, 'tim-dom': 10, 'per-mid': 1, 'art-add': 1", "## Essential info about entities\n\n* geo = Geographical Entity\n* org = Organization\n* per = Person\n* gpe = Geopolitical Entity\n* tim = Time indicator\n* art = Artifact\n* eve = Event\n* nat = Natural Phenomenon\n* Total Words Count = 1354149\n* Target Data Column: \"tag\" (ner_tag in this repo)\n\nInspiration: This dataset is getting more interested because of more features added to the recent version of this dataset. Also, it helps to create a broad view of Feature Engineering with respect to this dataset.", "## Modifications\n \nthe ner_dataset.csv was modified to have a similar data Structure as CoNLL-2003 dataset", "## Licensing information\n\nDatabase: Open Database, Contents: Database Contents." ]
[ "TAGS\n#annotations_creators-Abhinav Walia (Owner) #language-English #license-odbl #region-us \n", "# About Dataset\nfrom Kaggle Datasets", "## Context\n\nAnnotated Corpus for Named Entity Recognition using GMB(Groningen Meaning Bank) corpus for entity classification with enhanced and popular features by Natural Language Processing applied to the data set.\n\nTip: Use Pandas Dataframe to load dataset if using Python for convenience.", "## Content\nThis is the extract from GMB corpus which is tagged, annotated and built specifically to train the classifier to predict named entities such as name, location, etc.\n\nNumber of tagged entities:\n\n'O': 1146068', geo-nam': 58388, 'org-nam': 48034, 'per-nam': 23790, 'gpe-nam': 20680, 'tim-dat': 12786, 'tim-dow': 11404, 'per-tit': 9800, 'per-fam': 8152, 'tim-yoc': 5290, 'tim-moy': 4262, 'per-giv': 2413, 'tim-clo': 891, 'art-nam': 866, 'eve-nam': 602, 'nat-nam': 300, 'tim-nam': 146, 'eve-ord': 107, 'per-ini': 60, 'org-leg': 60, 'per-ord': 38, 'tim-dom': 10, 'per-mid': 1, 'art-add': 1", "## Essential info about entities\n\n* geo = Geographical Entity\n* org = Organization\n* per = Person\n* gpe = Geopolitical Entity\n* tim = Time indicator\n* art = Artifact\n* eve = Event\n* nat = Natural Phenomenon\n* Total Words Count = 1354149\n* Target Data Column: \"tag\" (ner_tag in this repo)\n\nInspiration: This dataset is getting more interested because of more features added to the recent version of this dataset. Also, it helps to create a broad view of Feature Engineering with respect to this dataset.", "## Modifications\n \nthe ner_dataset.csv was modified to have a similar data Structure as CoNLL-2003 dataset", "## Licensing information\n\nDatabase: Open Database, Contents: Database Contents." ]
cf29923953a1580840b263b22f800a2e4cbd66d9
Collated datasets from 10 sources and preprocessed it to have ["texts", "labels"] columns to train/finetune sequence-to-sequence models such as T5/Blenderbot ... Below are the 10 datasets: 1. blended_skill_talk, 2. conv_ai_2 3. empathetic_dialogues 4. wizard_of_wikipedia 5. meta_woz 6. multi_woz, 7. spolin 8. dailydialog 9. cornell_movie_dialogues 10. taskmaster The data access and preprocessing code is [here](https://github.com/pacman100/accelerate-deepspeed-test/blob/main/src/data_preprocessing/DataPreprocessing.ipynb)
smangrul/MuDoConv
[ "license:cc-by-nc-4.0", "region:us" ]
2022-06-24T05:05:04+00:00
{"license": "cc-by-nc-4.0"}
2022-06-29T05:39:30+00:00
[]
[]
TAGS #license-cc-by-nc-4.0 #region-us
Collated datasets from 10 sources and preprocessed it to have ["texts", "labels"] columns to train/finetune sequence-to-sequence models such as T5/Blenderbot ... Below are the 10 datasets: 1. blended_skill_talk, 2. conv_ai_2 3. empathetic_dialogues 4. wizard_of_wikipedia 5. meta_woz 6. multi_woz, 7. spolin 8. dailydialog 9. cornell_movie_dialogues 10. taskmaster The data access and preprocessing code is here
[]
[ "TAGS\n#license-cc-by-nc-4.0 #region-us \n" ]
76cd1995c3c8251656115f75187e1ceeae407448
German validation dataset from WECHSEL () to evaluate LLM perplexity. JSON-line files (on JSON object per line): - `valid.json.gz`: Gzipped validation set as generated by the paper (163,698 docs) - `valid.random_1636.json.gz`: Random 1% (1636 docs) of the validation set
malteos/wechsel_de
[ "task_categories:text-generation", "task_ids:language-modeling", "task_ids:masked-language-modeling", "size_categories:100k<n<1M", "language:de", "region:us" ]
2022-06-24T07:13:38+00:00
{"language": ["de"], "size_categories": ["100k<n<1M"], "task_categories": ["text-generation"], "task_ids": ["language-modeling", "masked-language-modeling"]}
2022-07-30T17:57:02+00:00
[]
[ "de" ]
TAGS #task_categories-text-generation #task_ids-language-modeling #task_ids-masked-language-modeling #size_categories-100k<n<1M #language-German #region-us
German validation dataset from WECHSEL () to evaluate LLM perplexity. JSON-line files (on JSON object per line): - 'URL': Gzipped validation set as generated by the paper (163,698 docs) - 'valid.random_1636.URL': Random 1% (1636 docs) of the validation set
[]
[ "TAGS\n#task_categories-text-generation #task_ids-language-modeling #task_ids-masked-language-modeling #size_categories-100k<n<1M #language-German #region-us \n" ]
0e417a4b73fec1352fdad25aa009950f74ea943f
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: mrm8488/distilroberta-finetuned-age_news-classification * Dataset: ag_news To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@abhishek](https://huggingface.co/abhishek) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-1c7ef613-7224755
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T07:40:40+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["ag_news"], "eval_info": {"task": "multi_class_classification", "model": "mrm8488/distilroberta-finetuned-age_news-classification", "dataset_name": "ag_news", "dataset_config": "default", "dataset_split": "test", "col_mapping": {"text": "text", "target": "label"}}}
2022-06-24T07:41:24+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Multi-class Text Classification * Model: mrm8488/distilroberta-finetuned-age_news-classification * Dataset: ag_news To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @abhishek for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Multi-class Text Classification\n* Model: mrm8488/distilroberta-finetuned-age_news-classification\n* Dataset: ag_news\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @abhishek for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Multi-class Text Classification\n* Model: mrm8488/distilroberta-finetuned-age_news-classification\n* Dataset: ag_news\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @abhishek for evaluating this model." ]
cd036c57e3d2827cbabd8009bcd2fa182c48279c
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: nateraw/bert-base-uncased-ag-news * Dataset: ag_news To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@abhishek](https://huggingface.co/abhishek) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-1c7ef613-7224756
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T07:40:46+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["ag_news"], "eval_info": {"task": "multi_class_classification", "model": "nateraw/bert-base-uncased-ag-news", "dataset_name": "ag_news", "dataset_config": "default", "dataset_split": "test", "col_mapping": {"text": "text", "target": "label"}}}
2022-06-24T07:41:49+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Multi-class Text Classification * Model: nateraw/bert-base-uncased-ag-news * Dataset: ag_news To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @abhishek for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Multi-class Text Classification\n* Model: nateraw/bert-base-uncased-ag-news\n* Dataset: ag_news\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @abhishek for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Multi-class Text Classification\n* Model: nateraw/bert-base-uncased-ag-news\n* Dataset: ag_news\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @abhishek for evaluating this model." ]
5d34bc138f12780d17ed89c92845e6ee6dfe0eb1
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: transformersbook/xlm-roberta-base-finetuned-panx-de * Dataset: xtreme To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-61110342-7234758
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T07:49:47+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["xtreme"], "eval_info": {"task": "entity_extraction", "model": "transformersbook/xlm-roberta-base-finetuned-panx-de", "dataset_name": "xtreme", "dataset_config": "PAN-X.de", "dataset_split": "validation", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T07:52:40+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: transformersbook/xlm-roberta-base-finetuned-panx-de * Dataset: xtreme To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: transformersbook/xlm-roberta-base-finetuned-panx-de\n* Dataset: xtreme\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: transformersbook/xlm-roberta-base-finetuned-panx-de\n* Dataset: xtreme\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
1ef0e0717148e428e134f4ceb3ebc845f917db63
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: transformersbook/xlm-roberta-base-finetuned-panx-all * Dataset: wikiann To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-6a6944f2-7244759
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T07:56:05+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["wikiann"], "eval_info": {"task": "entity_extraction", "model": "transformersbook/xlm-roberta-base-finetuned-panx-all", "dataset_name": "wikiann", "dataset_config": "en", "dataset_split": "test", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T07:58:55+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: transformersbook/xlm-roberta-base-finetuned-panx-all * Dataset: wikiann To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: transformersbook/xlm-roberta-base-finetuned-panx-all\n* Dataset: wikiann\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: transformersbook/xlm-roberta-base-finetuned-panx-all\n* Dataset: wikiann\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
209c35a42f9d52530a83550cefed4e6ee30cd7e8
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: philschmid/distilroberta-base-ner-wikiann * Dataset: wikiann To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-6a6944f2-7244760
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T07:56:08+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["wikiann"], "eval_info": {"task": "entity_extraction", "model": "philschmid/distilroberta-base-ner-wikiann", "dataset_name": "wikiann", "dataset_config": "en", "dataset_split": "test", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T07:58:21+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: philschmid/distilroberta-base-ner-wikiann * Dataset: wikiann To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: philschmid/distilroberta-base-ner-wikiann\n* Dataset: wikiann\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: philschmid/distilroberta-base-ner-wikiann\n* Dataset: wikiann\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
b3e3cb383d1d26bd35c1ac55dc18c8c572ac9a12
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: elastic/distilbert-base-cased-finetuned-conll03-english * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@douwekiela](https://huggingface.co/douwekiela) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-29af5371-7254761
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:00:57+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["conll2003"], "eval_info": {"task": "entity_extraction", "model": "elastic/distilbert-base-cased-finetuned-conll03-english", "dataset_name": "conll2003", "dataset_config": "conll2003", "dataset_split": "validation", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}, "metrics": []}}
2022-06-30T14:09:04+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: elastic/distilbert-base-cased-finetuned-conll03-english * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @douwekiela for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: elastic/distilbert-base-cased-finetuned-conll03-english\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: elastic/distilbert-base-cased-finetuned-conll03-english\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
a422adecc19262f6b1e0501423e18109664f247a
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: elastic/distilbert-base-uncased-finetuned-conll03-english * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@douwekiela](https://huggingface.co/douwekiela) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-29af5371-7254762
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:01:02+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["conll2003"], "eval_info": {"task": "entity_extraction", "model": "elastic/distilbert-base-uncased-finetuned-conll03-english", "dataset_name": "conll2003", "dataset_config": "conll2003", "dataset_split": "validation", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T08:02:06+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: elastic/distilbert-base-uncased-finetuned-conll03-english * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @douwekiela for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: elastic/distilbert-base-uncased-finetuned-conll03-english\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: elastic/distilbert-base-uncased-finetuned-conll03-english\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
c79ece872cd8e360a115f690aee73394ece734a5
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: huggingface-course/bert-finetuned-ner * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@douwekiela](https://huggingface.co/douwekiela) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-29af5371-7254763
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:01:07+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["conll2003"], "eval_info": {"task": "entity_extraction", "model": "huggingface-course/bert-finetuned-ner", "dataset_name": "conll2003", "dataset_config": "conll2003", "dataset_split": "validation", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T08:02:20+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: huggingface-course/bert-finetuned-ner * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @douwekiela for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: huggingface-course/bert-finetuned-ner\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: huggingface-course/bert-finetuned-ner\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
dc45572a60c24c4d731641aed222ef23f1e02a21
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: philschmid/distilroberta-base-ner-conll2003 * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@douwekiela](https://huggingface.co/douwekiela) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-29af5371-7254765
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:01:18+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["conll2003"], "eval_info": {"task": "entity_extraction", "model": "philschmid/distilroberta-base-ner-conll2003", "dataset_name": "conll2003", "dataset_config": "conll2003", "dataset_split": "validation", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-06-24T08:02:22+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Token Classification * Model: philschmid/distilroberta-base-ner-conll2003 * Dataset: conll2003 To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @douwekiela for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: philschmid/distilroberta-base-ner-conll2003\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Token Classification\n* Model: philschmid/distilroberta-base-ner-conll2003\n* Dataset: conll2003\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @douwekiela for evaluating this model." ]
6f9f190ca006db0fc95cad396463b020b7002e61
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: patrickvonplaten/bert2bert_cnn_daily_mail * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-be45ecbd-7284772
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:18:04+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["cnn_dailymail"], "eval_info": {"task": "summarization", "model": "patrickvonplaten/bert2bert_cnn_daily_mail", "dataset_name": "cnn_dailymail", "dataset_config": "3.0.0", "dataset_split": "test", "col_mapping": {"text": "article", "target": "highlights"}}}
2022-06-24T09:01:24+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: patrickvonplaten/bert2bert_cnn_daily_mail * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: patrickvonplaten/bert2bert_cnn_daily_mail\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: patrickvonplaten/bert2bert_cnn_daily_mail\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
a93e51a0086f1bad502798f81b6d8821f8f1090c
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: echarlaix/bart-base-cnn-r2-19.4-d35-hybrid * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-be45ecbd-7284773
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:18:09+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["cnn_dailymail"], "eval_info": {"task": "summarization", "model": "echarlaix/bart-base-cnn-r2-19.4-d35-hybrid", "dataset_name": "cnn_dailymail", "dataset_config": "3.0.0", "dataset_split": "test", "col_mapping": {"text": "article", "target": "highlights"}}}
2022-06-24T08:27:34+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: echarlaix/bart-base-cnn-r2-19.4-d35-hybrid * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: echarlaix/bart-base-cnn-r2-19.4-d35-hybrid\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: echarlaix/bart-base-cnn-r2-19.4-d35-hybrid\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
865abe187dd995261689af51bd95b20d12fcceca
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: echarlaix/bart-base-cnn-r2-18.7-d23-hybrid * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-be45ecbd-7284774
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:18:15+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["cnn_dailymail"], "eval_info": {"task": "summarization", "model": "echarlaix/bart-base-cnn-r2-18.7-d23-hybrid", "dataset_name": "cnn_dailymail", "dataset_config": "3.0.0", "dataset_split": "test", "col_mapping": {"text": "article", "target": "highlights"}}}
2022-06-24T08:27:22+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: echarlaix/bart-base-cnn-r2-18.7-d23-hybrid * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: echarlaix/bart-base-cnn-r2-18.7-d23-hybrid\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: echarlaix/bart-base-cnn-r2-18.7-d23-hybrid\n* Dataset: cnn_dailymail\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
ddff094ce88bfe41c0b749637146722fcc552ddf
# Dataset Card for MultiLingual LibriSpeech ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [MultiLingual LibriSpeech ASR corpus](http://www.openslr.org/94) - **Repository:** [Needs More Information] - **Paper:** [MLS: A Large-Scale Multilingual Dataset for Speech Research](https://arxiv.org/abs/2012.03411) - **Leaderboard:** [Paperswithcode Leaderboard](https://paperswithcode.com/dataset/multilingual-librispeech) ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href="index#supported-frameworks">table</a> to check if a model has fast tokenizer support.</p></div> Multilingual LibriSpeech (MLS) dataset is a large multilingual corpus suitable for speech research. The dataset is derived from read audiobooks from LibriVox and consists of 8 languages - English, German, Dutch, Spanish, French, Italian, Portuguese, Polish. ### Supported Tasks and Leaderboards - `automatic-speech-recognition`, `audio-speaker-identification`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at https://paperswithcode.com/dataset/multilingual-librispeech and ranks models based on their WER. <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href="index#supported-frameworks">table</a> to check if a model has fast tokenizer support.</p></div> <div class="alert alert-danger d-flex align-items-center" role="alert"> <svg class="bi flex-shrink-0 me-2" width="24" height="24" role="img" aria-label="Danger:"><use xlink:href="#exclamation-triangle-fill"/></svg> <div> An example danger alert with an icon </div> </div> <div class="alert alert-block alert-warning"> ⚠ In general, just avoid the red boxes. </div> <div class="alert alert-block alert-danger"> In general, just avoid the red boxes. </div> <div class="alert alert-danger" role="alert"> In general, just avoid the red boxes. </div> <div class="alert" role="alert"> In general, just avoid the red boxes. </div> <div class="course-tip-orange"> <strong>Error:</strong> </div> <div class="alert alert-danger" role="alert"> <div class="row vertical-align"> <div class="col-xs-1 text-center"> <i class="fa fa-exclamation-triangle fa-2x"></i> </div> <div class="col-xs-11"> <strong>Error:</strong> </div> </div> </div> >[!WARNING] >This is a warning _**Warning:** Be very careful here._ <Deprecated> This is a warning </Deprecated> <Tip warning> This is a warning </Tip> <Tip warning={true}> This is a warning </Tip> > **Warning** > This is a warning
albertvillanova/tmp-mention
[ "license:cc-by-4.0", "zenodo", "arxiv:2012.03411", "region:us" ]
2022-06-24T08:24:51+00:00
{"license": "cc-by-4.0", "tags": ["zenodo"]}
2022-09-22T10:26:20+00:00
[ "2012.03411" ]
[]
TAGS #license-cc-by-4.0 #zenodo #arxiv-2012.03411 #region-us
# Dataset Card for MultiLingual LibriSpeech ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: MultiLingual LibriSpeech ASR corpus - Repository: - Paper: MLS: A Large-Scale Multilingual Dataset for Speech Research - Leaderboard: Paperswithcode Leaderboard ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href="index#supported-frameworks">table</a> to check if a model has fast tokenizer support.</p></div> Multilingual LibriSpeech (MLS) dataset is a large multilingual corpus suitable for speech research. The dataset is derived from read audiobooks from LibriVox and consists of 8 languages - English, German, Dutch, Spanish, French, Italian, Portuguese, Polish. ### Supported Tasks and Leaderboards - 'automatic-speech-recognition', 'audio-speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER. <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href="index#supported-frameworks">table</a> to check if a model has fast tokenizer support.</p></div> <div class="alert alert-danger d-flex align-items-center" role="alert"> <svg class="bi flex-shrink-0 me-2" width="24" height="24" role="img" aria-label="Danger:"><use xlink:href="#exclamation-triangle-fill"/></svg> <div> An example danger alert with an icon </div> </div> <div class="alert alert-block alert-warning"> In general, just avoid the red boxes. </div> <div class="alert alert-block alert-danger"> In general, just avoid the red boxes. </div> <div class="alert alert-danger" role="alert"> In general, just avoid the red boxes. </div> <div class="alert" role="alert"> In general, just avoid the red boxes. </div> <div class="course-tip-orange"> <strong>Error:</strong> </div> <div class="alert alert-danger" role="alert"> <div class="row vertical-align"> <div class="col-xs-1 text-center"> <i class="fa fa-exclamation-triangle fa-2x"></i> </div> <div class="col-xs-11"> <strong>Error:</strong> </div> </div> </div> >[!WARNING] >This is a warning _Warning: Be very careful here._ <Deprecated> This is a warning </Deprecated> <Tip warning> This is a warning </Tip> <Tip warning={true}> This is a warning </Tip> > Warning > This is a warning
[ "# Dataset Card for MultiLingual LibriSpeech", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: MultiLingual LibriSpeech ASR corpus\n- Repository: \n- Paper: MLS: A Large-Scale Multilingual Dataset for Speech Research\n- Leaderboard: Paperswithcode Leaderboard", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href=\"index#supported-frameworks\">table</a> to check if a model has fast tokenizer support.</p></div>\n\nMultilingual LibriSpeech (MLS) dataset is a large multilingual corpus suitable for speech research. The dataset is derived from read audiobooks from LibriVox and consists of 8 languages - English, German, Dutch, Spanish, French, Italian, Portuguese, Polish.", "### Supported Tasks and Leaderboards\n\n- 'automatic-speech-recognition', 'audio-speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER.\n\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href=\"index#supported-frameworks\">table</a> to check if a model has fast tokenizer support.</p></div>\n\n\n<div class=\"alert alert-danger d-flex align-items-center\" role=\"alert\">\n <svg class=\"bi flex-shrink-0 me-2\" width=\"24\" height=\"24\" role=\"img\" aria-label=\"Danger:\"><use xlink:href=\"#exclamation-triangle-fill\"/></svg>\n <div>\n An example danger alert with an icon\n </div>\n</div>\n\n<div class=\"alert alert-block alert-warning\"> In general, just avoid the red boxes. </div>\n<div class=\"alert alert-block alert-danger\"> In general, just avoid the red boxes. </div>\n<div class=\"alert alert-danger\" role=\"alert\"> In general, just avoid the red boxes. </div>\n<div class=\"alert\" role=\"alert\"> In general, just avoid the red boxes. </div>\n\n<div class=\"course-tip-orange\">\n<strong>Error:</strong>\n</div>\n\n<div class=\"alert alert-danger\" role=\"alert\">\n <div class=\"row vertical-align\">\n <div class=\"col-xs-1 text-center\">\n <i class=\"fa fa-exclamation-triangle fa-2x\"></i>\n </div>\n <div class=\"col-xs-11\">\n <strong>Error:</strong> \n </div> \n </div> \n</div>\n\n>[!WARNING]\n>This is a warning\n\n_Warning: Be very careful here._\n\n<Deprecated>\nThis is a warning\n</Deprecated>\n\n<Tip warning>\nThis is a warning\n</Tip>\n\n<Tip warning={true}>\nThis is a warning\n</Tip>\n\n\n> Warning\n> This is a warning" ]
[ "TAGS\n#license-cc-by-4.0 #zenodo #arxiv-2012.03411 #region-us \n", "# Dataset Card for MultiLingual LibriSpeech", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: MultiLingual LibriSpeech ASR corpus\n- Repository: \n- Paper: MLS: A Large-Scale Multilingual Dataset for Speech Research\n- Leaderboard: Paperswithcode Leaderboard", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href=\"index#supported-frameworks\">table</a> to check if a model has fast tokenizer support.</p></div>\n\nMultilingual LibriSpeech (MLS) dataset is a large multilingual corpus suitable for speech research. The dataset is derived from read audiobooks from LibriVox and consists of 8 languages - English, German, Dutch, Spanish, French, Italian, Portuguese, Polish.", "### Supported Tasks and Leaderboards\n\n- 'automatic-speech-recognition', 'audio-speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER.\n\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\"><p><b>Deprecated:</b> Not every model supports a fast tokenizer. Take a look at this <a href=\"index#supported-frameworks\">table</a> to check if a model has fast tokenizer support.</p></div>\n\n\n<div class=\"alert alert-danger d-flex align-items-center\" role=\"alert\">\n <svg class=\"bi flex-shrink-0 me-2\" width=\"24\" height=\"24\" role=\"img\" aria-label=\"Danger:\"><use xlink:href=\"#exclamation-triangle-fill\"/></svg>\n <div>\n An example danger alert with an icon\n </div>\n</div>\n\n<div class=\"alert alert-block alert-warning\"> In general, just avoid the red boxes. </div>\n<div class=\"alert alert-block alert-danger\"> In general, just avoid the red boxes. </div>\n<div class=\"alert alert-danger\" role=\"alert\"> In general, just avoid the red boxes. </div>\n<div class=\"alert\" role=\"alert\"> In general, just avoid the red boxes. </div>\n\n<div class=\"course-tip-orange\">\n<strong>Error:</strong>\n</div>\n\n<div class=\"alert alert-danger\" role=\"alert\">\n <div class=\"row vertical-align\">\n <div class=\"col-xs-1 text-center\">\n <i class=\"fa fa-exclamation-triangle fa-2x\"></i>\n </div>\n <div class=\"col-xs-11\">\n <strong>Error:</strong> \n </div> \n </div> \n</div>\n\n>[!WARNING]\n>This is a warning\n\n_Warning: Be very careful here._\n\n<Deprecated>\nThis is a warning\n</Deprecated>\n\n<Tip warning>\nThis is a warning\n</Tip>\n\n<Tip warning={true}>\nThis is a warning\n</Tip>\n\n\n> Warning\n> This is a warning" ]
c46a7c127048d9a3e7464821c50286437a64360e
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: human-centered-summarization/financial-summarization-pegasus * Dataset: xsum To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-38643302-7294782
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:44:47+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["xsum"], "eval_info": {"task": "summarization", "model": "human-centered-summarization/financial-summarization-pegasus", "dataset_name": "xsum", "dataset_config": "default", "dataset_split": "test", "col_mapping": {"text": "document", "target": "summary"}}}
2022-06-24T09:11:47+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: human-centered-summarization/financial-summarization-pegasus * Dataset: xsum To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: human-centered-summarization/financial-summarization-pegasus\n* Dataset: xsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: human-centered-summarization/financial-summarization-pegasus\n* Dataset: xsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
bfb745c7878dc97e211a4d0369d39fde72b8faef
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: philschmid/bart-base-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-84760c85-7314784
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:50:37+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["samsum"], "eval_info": {"task": "summarization", "model": "philschmid/bart-base-samsum", "dataset_name": "samsum", "dataset_config": "samsum", "dataset_split": "test", "col_mapping": {"text": "dialogue", "target": "summary"}}}
2022-06-24T08:51:31+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: philschmid/bart-base-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/bart-base-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/bart-base-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
d2f15534b513134d94a76bca71c15745fa89c28a
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: philschmid/bart-large-cnn-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-84760c85-7314785
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:50:41+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["samsum"], "eval_info": {"task": "summarization", "model": "philschmid/bart-large-cnn-samsum", "dataset_name": "samsum", "dataset_config": "samsum", "dataset_split": "test", "col_mapping": {"text": "dialogue", "target": "summary"}}}
2022-06-24T08:53:26+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: philschmid/bart-large-cnn-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/bart-large-cnn-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/bart-large-cnn-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
990b496e4d876f019f7d2519dfdaa9a2ea633bcf
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: philschmid/distilbart-cnn-12-6-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic evaluation service](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-84760c85-7314786
[ "autotrain", "evaluation", "region:us" ]
2022-06-24T08:50:47+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["samsum"], "eval_info": {"task": "summarization", "model": "philschmid/distilbart-cnn-12-6-samsum", "dataset_name": "samsum", "dataset_config": "samsum", "dataset_split": "test", "col_mapping": {"text": "dialogue", "target": "summary"}}}
2022-06-24T08:52:53+00:00
[]
[]
TAGS #autotrain #evaluation #region-us
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by AutoTrain for the following task and dataset: * Task: Summarization * Model: philschmid/distilbart-cnn-12-6-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's automatic evaluation service. ## Contributions Thanks to @lewtun for evaluating this model.
[ "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/distilbart-cnn-12-6-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
[ "TAGS\n#autotrain #evaluation #region-us \n", "# Dataset Card for AutoTrain Evaluator\n\nThis repository contains model predictions generated by AutoTrain for the following task and dataset:\n\n* Task: Summarization\n* Model: philschmid/distilbart-cnn-12-6-samsum\n* Dataset: samsum\n\nTo run new evaluation jobs, visit Hugging Face's automatic evaluation service.", "## Contributions\n\nThanks to @lewtun for evaluating this model." ]
91860adfc9a09aabca5cddb5247442109b38e213
## Code snippet to visualise the position of the box ```python import matplotlib.image as img import matplotlib.pyplot as plt from datasets import load_dataset from matplotlib.patches import Rectangle # Load dataset ds_name = "HuggingFaceM4/FGVC-Aircraft" ds = load_dataset(ds_name, use_auth_token=True) # Extract information for the sample we want to show index = 300 sample = ds["train"][index] box_coord = sample["bbox"] xmin = box_coord["xmin"] ymin = box_coord["ymin"] xmax = box_coord["xmax"] ymax = box_coord["ymax"] img_path = sample["image"].filename # Create plot # define Matplotlib figure and axis fig, ax = plt.subplots() # plot figure image = img.imread(img_path) ax.imshow(image) # add rectangle to plot ax.add_patch( Rectangle((xmin, ymin), xmax-xmin, ymax - ymin, fill=None) ) # display plot plt.show() ```
HuggingFaceM4/FGVC-Aircraft
[ "region:us" ]
2022-06-24T11:19:13+00:00
{}
2022-06-24T13:18:11+00:00
[]
[]
TAGS #region-us
## Code snippet to visualise the position of the box
[ "## Code snippet to visualise the position of the box" ]
[ "TAGS\n#region-us \n", "## Code snippet to visualise the position of the box" ]
e937c2db8eab109cafc4f5279a396957d38251c5
# Dataset Card for predicting-brazilian-court-decisions ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** https://github.com/lagefreitas/predicting-brazilian-court-decisions - **Paper:** Lage-Freitas, A., Allende-Cid, H., Santana, O., & Oliveira-Lage, L. (2022). Predicting Brazilian Court Decisions. PeerJ. Computer Science, 8, e904–e904. https://doi.org/10.7717/peerj-cs.904 - **Leaderboard:** - **Point of Contact:** [Joel Niklaus](mailto:[email protected]) ### Dataset Summary The dataset is a collection of 4043 *Ementa* (summary) court decisions and their metadata from the *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). The court decisions are labeled according to 7 categories and whether the decisions were unanimous on the part of the judges or not. The dataset supports the task of Legal Judgment Prediction. ### Supported Tasks and Leaderboards Legal Judgment Prediction ### Languages Brazilian Portuguese ## Dataset Structure ### Data Instances The file format is jsonl and three data splits are present (train, validation and test) for each configuration. ### Data Fields The dataset contains the following fields: - `process_number`: A number assigned to the decision by the court - `orgao_julgador`: Judging Body: one of '1ª Câmara Cível', '2ª Câmara Cível', '3ª Câmara Cível', 'Câmara Criminal', ' Tribunal Pleno', 'Seção Especializada Cível' - `publish_date`: The date, when the decision has been published (14/12/2018 - 03/04/2019). At that time (in 2018-2019), the scraping script was limited and not configurable to get data based on date range. Therefore, only the data from the last months has been scraped. - `judge_relator`: Judicial panel - `ementa_text`: Summary of the court decision - `decision_description`: **Suggested input**. Corresponds to ementa_text - judgment_text - unanimity_text. Basic statistics (number of words): mean: 119, median: 88, min: 12, max: 1400 - `judgment_text`: The text used for determining the judgment label - `judgment_label`: **Primary suggested label**. Labels that can be used to train a model for judgment prediction: - `no`: The appeal was denied - `partial`: For partially favourable decisions - `yes`: For fully favourable decisions - removed labels (present in the original dataset): - `conflito-competencia`: Meta-decision. For example, a decision just to tell that Court A should rule this case and not Court B. - `not-cognized`: The appeal was not accepted to be judged by the court - `prejudicada`: The case could not be judged for any impediment such as the appealer died or gave up on the case for instance. - `unanimity_text`: Portuguese text to describe whether the decision was unanimous or not. - `unanimity_label`: **Secondary suggested label**. Unified labels to describe whether the decision was unanimous or not (in some cases contains ```not_determined```); they can be used for model training as well (Lage-Freitas et al., 2019). ### Data Splits The data has been split randomly into 80% train (3234), 10% validation (404), 10% test (405). There are two tasks possible for this dataset. #### Judgment Label Distribution | judgment | train | validation | test | |:----------|---------:|-----------:|--------:| | no | 1960 | 221 | 234 | | partial | 677 | 96 | 93 | | yes | 597 | 87 | 78 | | **total** | **3234** | **404** | **405** | #### Unanimity In this configuration, all cases that have `not_determined` as `unanimity_label` can be removed. Label Distribution | unanimity_label | train | validation | test | |:-----------------|----------:|---------------:|---------:| | not_determined | 1519 | 193 | 201 | | unanimity | 1681 | 205 | 200 | | not-unanimity | 34 | 6 | 4 | | **total** | **3234** | **404** | **405** | ## Dataset Creation ### Curation Rationale This dataset was created to further the research on developing models for predicting Brazilian court decisions that are also able to predict whether the decision will be unanimous. ### Source Data The data was scraped from *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). #### Initial Data Collection and Normalization *“We developed a Web scraper for collecting data from Brazilian courts. The scraper first searched for the URL that contains the list of court cases […]. Then, the scraper extracted from these HTML files the specific case URLs and downloaded their data […]. Next, it extracted the metadata and the contents of legal cases and stored them in a CSV file format […].”* (Lage-Freitas et al., 2022) #### Who are the source language producers? The source language producer are presumably attorneys, judges, and other legal professionals. ### Annotations #### Annotation process The dataset was not annotated. #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information The court decisions might contain sensitive information about individuals. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Note that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton Matoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset consisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the dataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that, differences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to have a look at the conversion script ```convert_to_hf_dataset.py``` in order to retrace the steps for converting the original dataset into the present jsonl-format. For further information on the original dataset structure, we refer to the bibliographical references and the original Github repositories and/or web pages provided in this dataset card. ## Additional Information Lage-Freitas, A., Allende-Cid, H., Santana Jr, O., & Oliveira-Lage, L. (2019). Predicting Brazilian court decisions: - "In Brazil [...] lower court judges decisions might be appealed to Brazilian courts (*Tribiunais de Justiça*) to be reviewed by second instance court judges. In an appellate court, judges decide together upon a case and their decisions are compiled in Agreement reports named *Acóordãos*." ### Dataset Curators The names of the original dataset curators and creators can be found in references given below, in the section *Citation Information*. Additional changes were made by Joel Niklaus ([Email](mailto:[email protected]) ; [Github](https://github.com/joelniklaus)) and Veton Matoshi ([Email](mailto:[email protected]) ; [Github](https://github.com/kapllan)). ### Licensing Information No licensing information was provided for this dataset. However, please make sure that you use the dataset according to Brazilian law. ### Citation Information ``` @misc{https://doi.org/10.48550/arxiv.1905.10348, author = {Lage-Freitas, Andr{\'{e}} and Allende-Cid, H{\'{e}}ctor and Santana, Orivaldo and de Oliveira-Lage, L{\'{i}}via}, doi = {10.48550/ARXIV.1905.10348}, keywords = {Computation and Language (cs.CL),FOS: Computer and information sciences,Social and Information Networks (cs.SI)}, publisher = {arXiv}, title = {{Predicting Brazilian court decisions}}, url = {https://arxiv.org/abs/1905.10348}, year = {2019} } ``` ``` @article{Lage-Freitas2022, author = {Lage-Freitas, Andr{\'{e}} and Allende-Cid, H{\'{e}}ctor and Santana, Orivaldo and Oliveira-Lage, L{\'{i}}via}, doi = {10.7717/peerj-cs.904}, issn = {2376-5992}, journal = {PeerJ. Computer science}, keywords = {Artificial intelligence,Jurimetrics,Law,Legal,Legal NLP,Legal informatics,Legal outcome forecast,Litigation prediction,Machine learning,NLP,Portuguese,Predictive algorithms,judgement prediction}, language = {eng}, month = {mar}, pages = {e904--e904}, publisher = {PeerJ Inc.}, title = {{Predicting Brazilian Court Decisions}}, url = {https://pubmed.ncbi.nlm.nih.gov/35494851 https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9044329/}, volume = {8}, year = {2022} } ``` ### Contributions Thanks to [@kapllan](https://github.com/kapllan) and [@joelniklaus](https://github.com/joelniklaus) for adding this dataset.
joelniklaus/brazilian_court_decisions
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:pt", "license:other", "arxiv:1905.10348", "region:us" ]
2022-06-24T12:50:02+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["pt"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification"], "pretty_name": "predicting-brazilian-court-decisions"}
2022-09-22T12:43:42+00:00
[ "1905.10348" ]
[ "pt" ]
TAGS #task_categories-text-classification #task_ids-multi-class-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Portuguese #license-other #arxiv-1905.10348 #region-us
Dataset Card for predicting-brazilian-court-decisions ===================================================== Table of Contents ----------------- * Table of Contents * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: * Repository: URL * Paper: Lage-Freitas, A., Allende-Cid, H., Santana, O., & Oliveira-Lage, L. (2022). Predicting Brazilian Court Decisions. PeerJ. Computer Science, 8, e904–e904. URL * Leaderboard: * Point of Contact: Joel Niklaus ### Dataset Summary The dataset is a collection of 4043 *Ementa* (summary) court decisions and their metadata from the *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). The court decisions are labeled according to 7 categories and whether the decisions were unanimous on the part of the judges or not. The dataset supports the task of Legal Judgment Prediction. ### Supported Tasks and Leaderboards Legal Judgment Prediction ### Languages Brazilian Portuguese Dataset Structure ----------------- ### Data Instances The file format is jsonl and three data splits are present (train, validation and test) for each configuration. ### Data Fields The dataset contains the following fields: * 'process\_number': A number assigned to the decision by the court * 'orgao\_julgador': Judging Body: one of '1ª Câmara Cível', '2ª Câmara Cível', '3ª Câmara Cível', 'Câmara Criminal', ' Tribunal Pleno', 'Seção Especializada Cível' * 'publish\_date': The date, when the decision has been published (14/12/2018 - 03/04/2019). At that time (in 2018-2019), the scraping script was limited and not configurable to get data based on date range. Therefore, only the data from the last months has been scraped. * 'judge\_relator': Judicial panel * 'ementa\_text': Summary of the court decision * 'decision\_description': Suggested input. Corresponds to ementa\_text - judgment\_text - unanimity\_text. Basic statistics (number of words): mean: 119, median: 88, min: 12, max: 1400 * 'judgment\_text': The text used for determining the judgment label * 'judgment\_label': Primary suggested label. Labels that can be used to train a model for judgment prediction: + 'no': The appeal was denied + 'partial': For partially favourable decisions + 'yes': For fully favourable decisions + removed labels (present in the original dataset): - 'conflito-competencia': Meta-decision. For example, a decision just to tell that Court A should rule this case and not Court B. - 'not-cognized': The appeal was not accepted to be judged by the court - 'prejudicada': The case could not be judged for any impediment such as the appealer died or gave up on the case for instance. * 'unanimity\_text': Portuguese text to describe whether the decision was unanimous or not. * 'unanimity\_label': Secondary suggested label. Unified labels to describe whether the decision was unanimous or not (in some cases contains ); they can be used for model training as well (Lage-Freitas et al., 2019). ### Data Splits The data has been split randomly into 80% train (3234), 10% validation (404), 10% test (405). There are two tasks possible for this dataset. #### Judgment Label Distribution #### Unanimity In this configuration, all cases that have 'not\_determined' as 'unanimity\_label' can be removed. Label Distribution Dataset Creation ---------------- ### Curation Rationale This dataset was created to further the research on developing models for predicting Brazilian court decisions that are also able to predict whether the decision will be unanimous. ### Source Data The data was scraped from *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). #### Initial Data Collection and Normalization *“We developed a Web scraper for collecting data from Brazilian courts. The scraper first searched for the URL that contains the list of court cases […]. Then, the scraper extracted from these HTML files the specific case URLs and downloaded their data […]. Next, it extracted the metadata and the contents of legal cases and stored them in a CSV file format […].”* (Lage-Freitas et al., 2022) #### Who are the source language producers? The source language producer are presumably attorneys, judges, and other legal professionals. ### Annotations #### Annotation process The dataset was not annotated. #### Who are the annotators? ### Personal and Sensitive Information The court decisions might contain sensitive information about individuals. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Note that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton Matoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset consisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the dataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that, differences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to have a look at the conversion script in order to retrace the steps for converting the original dataset into the present jsonl-format. For further information on the original dataset structure, we refer to the bibliographical references and the original Github repositories and/or web pages provided in this dataset card. Additional Information ---------------------- Lage-Freitas, A., Allende-Cid, H., Santana Jr, O., & Oliveira-Lage, L. (2019). Predicting Brazilian court decisions: * "In Brazil [...] lower court judges decisions might be appealed to Brazilian courts (*Tribiunais de Justiça*) to be reviewed by second instance court judges. In an appellate court, judges decide together upon a case and their decisions are compiled in Agreement reports named *Acóordãos*." ### Dataset Curators The names of the original dataset curators and creators can be found in references given below, in the section *Citation Information*. Additional changes were made by Joel Niklaus (Email ; Github) and Veton Matoshi (Email ; Github). ### Licensing Information No licensing information was provided for this dataset. However, please make sure that you use the dataset according to Brazilian law. ### Contributions Thanks to @kapllan and @joelniklaus for adding this dataset.
[ "### Dataset Summary\n\n\nThe dataset is a collection of 4043 *Ementa* (summary) court decisions and their metadata from\nthe *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). The court decisions are labeled\naccording to 7 categories and whether the decisions were unanimous on the part of the judges or not. The dataset\nsupports the task of Legal Judgment Prediction.", "### Supported Tasks and Leaderboards\n\n\nLegal Judgment Prediction", "### Languages\n\n\nBrazilian Portuguese\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe file format is jsonl and three data splits are present (train, validation and test) for each configuration.", "### Data Fields\n\n\nThe dataset contains the following fields:\n\n\n* 'process\\_number': A number assigned to the decision by the court\n* 'orgao\\_julgador': Judging Body: one of '1ª Câmara Cível', '2ª Câmara Cível', '3ª Câmara Cível', 'Câmara Criminal', '\nTribunal Pleno', 'Seção Especializada Cível'\n* 'publish\\_date': The date, when the decision has been published (14/12/2018 - 03/04/2019). At that time (in 2018-2019),\nthe scraping script was limited and not configurable to get data based on date range. Therefore, only the data from\nthe last months has been scraped.\n* 'judge\\_relator': Judicial panel\n* 'ementa\\_text': Summary of the court decision\n* 'decision\\_description': Suggested input. Corresponds to ementa\\_text - judgment\\_text - unanimity\\_text. Basic\nstatistics (number of words): mean: 119, median: 88, min: 12, max: 1400\n* 'judgment\\_text': The text used for determining the judgment label\n* 'judgment\\_label': Primary suggested label. Labels that can be used to train a model for judgment prediction:\n\t+ 'no': The appeal was denied\n\t+ 'partial': For partially favourable decisions\n\t+ 'yes': For fully favourable decisions\n\t+ removed labels (present in the original dataset):\n\t\t- 'conflito-competencia': Meta-decision. For example, a decision just to tell that Court A should rule this case\n\t\tand not Court B.\n\t\t- 'not-cognized': The appeal was not accepted to be judged by the court\n\t\t- 'prejudicada': The case could not be judged for any impediment such as the appealer died or gave up on the\n\t\tcase for instance.\n* 'unanimity\\_text': Portuguese text to describe whether the decision was unanimous or not.\n* 'unanimity\\_label': Secondary suggested label. Unified labels to describe whether the decision was unanimous or\nnot (in some cases contains ); they can be used for model training as well (Lage-Freitas et al.,\n2019).", "### Data Splits\n\n\nThe data has been split randomly into 80% train (3234), 10% validation (404), 10% test (405).\n\n\nThere are two tasks possible for this dataset.", "#### Judgment\n\n\nLabel Distribution", "#### Unanimity\n\n\nIn this configuration, all cases that have 'not\\_determined' as 'unanimity\\_label' can be removed.\n\n\nLabel Distribution\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThis dataset was created to further the research on developing models for predicting Brazilian court decisions that are\nalso able to predict whether the decision will be unanimous.", "### Source Data\n\n\nThe data was scraped from *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil).", "#### Initial Data Collection and Normalization\n\n\n*“We developed a Web scraper for collecting data from Brazilian courts. The scraper first searched for the URL that\ncontains the list of court cases […]. Then, the scraper extracted from these HTML files the specific case URLs and\ndownloaded their data […]. Next, it extracted the metadata and the contents of legal cases and stored them in a CSV file\nformat […].”* (Lage-Freitas et al., 2022)", "#### Who are the source language producers?\n\n\nThe source language producer are presumably attorneys, judges, and other legal professionals.", "### Annotations", "#### Annotation process\n\n\nThe dataset was not annotated.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThe court decisions might contain sensitive information about individuals.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nNote that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton\nMatoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset\nconsisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the\ndataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that,\ndifferences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to\nhave a look at the conversion script in order to retrace the steps for converting the\noriginal dataset into the present jsonl-format. For further information on the original dataset structure, we refer to\nthe bibliographical references and the original Github repositories and/or web pages provided in this dataset card.\n\n\nAdditional Information\n----------------------\n\n\nLage-Freitas, A., Allende-Cid, H., Santana Jr, O., & Oliveira-Lage, L. (2019). Predicting Brazilian court decisions:\n\n\n* \"In Brazil [...] lower court judges decisions might be appealed to Brazilian courts (*Tribiunais de Justiça*) to be\nreviewed by second instance court judges. In an appellate court, judges decide together upon a case and their\ndecisions are compiled in Agreement reports named *Acóordãos*.\"", "### Dataset Curators\n\n\nThe names of the original dataset curators and creators can be found in references given below, in the section *Citation\nInformation*. Additional changes were made by Joel Niklaus (Email\n; Github) and Veton Matoshi (Email\n; Github).", "### Licensing Information\n\n\nNo licensing information was provided for this dataset. However, please make sure that you use the dataset according to\nBrazilian law.", "### Contributions\n\n\nThanks to @kapllan and @joelniklaus for adding this\ndataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-multi-class-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Portuguese #license-other #arxiv-1905.10348 #region-us \n", "### Dataset Summary\n\n\nThe dataset is a collection of 4043 *Ementa* (summary) court decisions and their metadata from\nthe *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil). The court decisions are labeled\naccording to 7 categories and whether the decisions were unanimous on the part of the judges or not. The dataset\nsupports the task of Legal Judgment Prediction.", "### Supported Tasks and Leaderboards\n\n\nLegal Judgment Prediction", "### Languages\n\n\nBrazilian Portuguese\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe file format is jsonl and three data splits are present (train, validation and test) for each configuration.", "### Data Fields\n\n\nThe dataset contains the following fields:\n\n\n* 'process\\_number': A number assigned to the decision by the court\n* 'orgao\\_julgador': Judging Body: one of '1ª Câmara Cível', '2ª Câmara Cível', '3ª Câmara Cível', 'Câmara Criminal', '\nTribunal Pleno', 'Seção Especializada Cível'\n* 'publish\\_date': The date, when the decision has been published (14/12/2018 - 03/04/2019). At that time (in 2018-2019),\nthe scraping script was limited and not configurable to get data based on date range. Therefore, only the data from\nthe last months has been scraped.\n* 'judge\\_relator': Judicial panel\n* 'ementa\\_text': Summary of the court decision\n* 'decision\\_description': Suggested input. Corresponds to ementa\\_text - judgment\\_text - unanimity\\_text. Basic\nstatistics (number of words): mean: 119, median: 88, min: 12, max: 1400\n* 'judgment\\_text': The text used for determining the judgment label\n* 'judgment\\_label': Primary suggested label. Labels that can be used to train a model for judgment prediction:\n\t+ 'no': The appeal was denied\n\t+ 'partial': For partially favourable decisions\n\t+ 'yes': For fully favourable decisions\n\t+ removed labels (present in the original dataset):\n\t\t- 'conflito-competencia': Meta-decision. For example, a decision just to tell that Court A should rule this case\n\t\tand not Court B.\n\t\t- 'not-cognized': The appeal was not accepted to be judged by the court\n\t\t- 'prejudicada': The case could not be judged for any impediment such as the appealer died or gave up on the\n\t\tcase for instance.\n* 'unanimity\\_text': Portuguese text to describe whether the decision was unanimous or not.\n* 'unanimity\\_label': Secondary suggested label. Unified labels to describe whether the decision was unanimous or\nnot (in some cases contains ); they can be used for model training as well (Lage-Freitas et al.,\n2019).", "### Data Splits\n\n\nThe data has been split randomly into 80% train (3234), 10% validation (404), 10% test (405).\n\n\nThere are two tasks possible for this dataset.", "#### Judgment\n\n\nLabel Distribution", "#### Unanimity\n\n\nIn this configuration, all cases that have 'not\\_determined' as 'unanimity\\_label' can be removed.\n\n\nLabel Distribution\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThis dataset was created to further the research on developing models for predicting Brazilian court decisions that are\nalso able to predict whether the decision will be unanimous.", "### Source Data\n\n\nThe data was scraped from *Tribunal de Justiça de Alagoas* (TJAL, the State Supreme Court of Alagoas (Brazil).", "#### Initial Data Collection and Normalization\n\n\n*“We developed a Web scraper for collecting data from Brazilian courts. The scraper first searched for the URL that\ncontains the list of court cases […]. Then, the scraper extracted from these HTML files the specific case URLs and\ndownloaded their data […]. Next, it extracted the metadata and the contents of legal cases and stored them in a CSV file\nformat […].”* (Lage-Freitas et al., 2022)", "#### Who are the source language producers?\n\n\nThe source language producer are presumably attorneys, judges, and other legal professionals.", "### Annotations", "#### Annotation process\n\n\nThe dataset was not annotated.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThe court decisions might contain sensitive information about individuals.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nNote that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton\nMatoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset\nconsisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the\ndataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that,\ndifferences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to\nhave a look at the conversion script in order to retrace the steps for converting the\noriginal dataset into the present jsonl-format. For further information on the original dataset structure, we refer to\nthe bibliographical references and the original Github repositories and/or web pages provided in this dataset card.\n\n\nAdditional Information\n----------------------\n\n\nLage-Freitas, A., Allende-Cid, H., Santana Jr, O., & Oliveira-Lage, L. (2019). Predicting Brazilian court decisions:\n\n\n* \"In Brazil [...] lower court judges decisions might be appealed to Brazilian courts (*Tribiunais de Justiça*) to be\nreviewed by second instance court judges. In an appellate court, judges decide together upon a case and their\ndecisions are compiled in Agreement reports named *Acóordãos*.\"", "### Dataset Curators\n\n\nThe names of the original dataset curators and creators can be found in references given below, in the section *Citation\nInformation*. Additional changes were made by Joel Niklaus (Email\n; Github) and Veton Matoshi (Email\n; Github).", "### Licensing Information\n\n\nNo licensing information was provided for this dataset. However, please make sure that you use the dataset according to\nBrazilian law.", "### Contributions\n\n\nThanks to @kapllan and @joelniklaus for adding this\ndataset." ]
8265518f6b5127d386a85ab5c380d867ff9ae70b
# Dataset Card for MTG Jamendo Dataset ## Dataset Description - **Repository:** [MTG Jamendo dataset repository](https://github.com/MTG/mtg-jamendo-dataset) ### Dataset Summary MTG-Jamendo Dataset, a new open dataset for music auto-tagging. It is built using music available at Jamendo under Creative Commons licenses and tags provided by content uploaders. The dataset contains over 55,000 full audio tracks with 195 tags from genre, instrument, and mood/theme categories. We provide elaborated data splits for researchers and report the performance of a simple baseline approach on five different sets of tags: genre, instrument, mood/theme, top-50, and overall. ## Dataset structure ### Data Fields - `id`: an integer containing the id of the track - `artist_id`: an integer containing the id of the artist - `album_id`: an integer containing the id of the album - `duration_in_sec`: duration of the track as a float - `genres`: list of strings, describing genres the track is assigned to - `instruments`: list of strings for the main instruments of the track - `moods`: list of strings, describing the moods the track is assigned to - `audio`: audio of the track ### Data Splits This dataset has 2 balanced splits: _train_ (90%) and _validation_ (10%) ### Licensing Information This dataset version 1.0.0 is released under the [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0). ### Citation Information ``` @conference {bogdanov2019mtg, author = "Bogdanov, Dmitry and Won, Minz and Tovstogan, Philip and Porter, Alastair and Serra, Xavier", title = "The MTG-Jamendo Dataset for Automatic Music Tagging", booktitle = "Machine Learning for Music Discovery Workshop, International Conference on Machine Learning (ICML 2019)", year = "2019", address = "Long Beach, CA, United States", url = "http://hdl.handle.net/10230/42015" } ```
rkstgr/mtg-jamendo
[ "size_categories:10K<n<100K", "source_datasets:original", "license:apache-2.0", "region:us" ]
2022-06-24T12:51:38+00:00
{"license": ["apache-2.0"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "pretty_name": "MTG Jamendo"}
2022-07-22T11:56:25+00:00
[]
[]
TAGS #size_categories-10K<n<100K #source_datasets-original #license-apache-2.0 #region-us
# Dataset Card for MTG Jamendo Dataset ## Dataset Description - Repository: MTG Jamendo dataset repository ### Dataset Summary MTG-Jamendo Dataset, a new open dataset for music auto-tagging. It is built using music available at Jamendo under Creative Commons licenses and tags provided by content uploaders. The dataset contains over 55,000 full audio tracks with 195 tags from genre, instrument, and mood/theme categories. We provide elaborated data splits for researchers and report the performance of a simple baseline approach on five different sets of tags: genre, instrument, mood/theme, top-50, and overall. ## Dataset structure ### Data Fields - 'id': an integer containing the id of the track - 'artist_id': an integer containing the id of the artist - 'album_id': an integer containing the id of the album - 'duration_in_sec': duration of the track as a float - 'genres': list of strings, describing genres the track is assigned to - 'instruments': list of strings for the main instruments of the track - 'moods': list of strings, describing the moods the track is assigned to - 'audio': audio of the track ### Data Splits This dataset has 2 balanced splits: _train_ (90%) and _validation_ (10%) ### Licensing Information This dataset version 1.0.0 is released under the Apache-2.0 License.
[ "# Dataset Card for MTG Jamendo Dataset", "## Dataset Description\n- Repository: MTG Jamendo dataset repository", "### Dataset Summary\n\nMTG-Jamendo Dataset, a new open dataset for music auto-tagging. It is built using music available at Jamendo under Creative Commons licenses and tags provided by content uploaders. The dataset contains over 55,000 full audio tracks with 195 tags from genre, instrument, and mood/theme categories. We provide elaborated data splits for researchers and report the performance of a simple baseline approach on five different sets of tags: genre, instrument, mood/theme, top-50, and overall.", "## Dataset structure", "### Data Fields\n\n- 'id': an integer containing the id of the track\n- 'artist_id': an integer containing the id of the artist\n- 'album_id': an integer containing the id of the album\n- 'duration_in_sec': duration of the track as a float\n- 'genres': list of strings, describing genres the track is assigned to\n- 'instruments': list of strings for the main instruments of the track\n- 'moods': list of strings, describing the moods the track is assigned to\n- 'audio': audio of the track", "### Data Splits\n\nThis dataset has 2 balanced splits: _train_ (90%) and _validation_ (10%)", "### Licensing Information\n\nThis dataset version 1.0.0 is released under the Apache-2.0 License." ]
[ "TAGS\n#size_categories-10K<n<100K #source_datasets-original #license-apache-2.0 #region-us \n", "# Dataset Card for MTG Jamendo Dataset", "## Dataset Description\n- Repository: MTG Jamendo dataset repository", "### Dataset Summary\n\nMTG-Jamendo Dataset, a new open dataset for music auto-tagging. It is built using music available at Jamendo under Creative Commons licenses and tags provided by content uploaders. The dataset contains over 55,000 full audio tracks with 195 tags from genre, instrument, and mood/theme categories. We provide elaborated data splits for researchers and report the performance of a simple baseline approach on five different sets of tags: genre, instrument, mood/theme, top-50, and overall.", "## Dataset structure", "### Data Fields\n\n- 'id': an integer containing the id of the track\n- 'artist_id': an integer containing the id of the artist\n- 'album_id': an integer containing the id of the album\n- 'duration_in_sec': duration of the track as a float\n- 'genres': list of strings, describing genres the track is assigned to\n- 'instruments': list of strings for the main instruments of the track\n- 'moods': list of strings, describing the moods the track is assigned to\n- 'audio': audio of the track", "### Data Splits\n\nThis dataset has 2 balanced splits: _train_ (90%) and _validation_ (10%)", "### Licensing Information\n\nThis dataset version 1.0.0 is released under the Apache-2.0 License." ]
8587e5a368f814fd15928af0254ee8d2b19e4471
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed]
israfelsr/img-wikipedia-simple
[ "task_categories:image-to-text", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "language:en", "region:us" ]
2022-06-24T12:59:27+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": [], "multilinguality": ["monolingual"], "size_categories": [], "source_datasets": [], "task_categories": ["image-to-text"], "pretty_name": "image-wikipedia-simple"}
2022-08-26T15:13:05+00:00
[]
[ "en" ]
TAGS #task_categories-image-to-text #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #language-English #region-us
# Dataset Card for [Dataset Name] ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations
[ "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage:\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations" ]
[ "TAGS\n#task_categories-image-to-text #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #language-English #region-us \n", "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage:\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations" ]