iperbole commited on
Commit
c3871b7
·
verified ·
1 Parent(s): 6f775d4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -5,6 +5,8 @@ language:
5
  license: apache-2.0
6
  pipeline_tag: text-generation
7
  library_name: transformers
 
 
8
  ---
9
 
10
  # Mistral-7B-v0.1-Italian-RANDOM
@@ -16,13 +18,13 @@ library_name: transformers
16
 
17
  The **Mistral-7B-v0.1-Adapted** collection of large language models (LLMs), is a collection of adapted generative models in 7B (text in/text out), adapted models from **Mistral-7B-Base-v0.1**.
18
 
19
- *Mistral-v0.1-Italian-RANDOM* is a continual trained mistral model, after tokenizer substitution.
20
 
21
  The tokenizer of this models after adaptation is the same of [Minverva-3B](https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0).
22
 
23
  **Model developer:** SapienzaNLP, ISTI-CNR, ILC-CNR
24
 
25
- **Model Architecture:** Mistral-7B-v0.1-Adapted is an auto-regressive language model that uses an optimized transformer architecture.
26
 
27
  ## Data used for the adaptation
28
 
@@ -34,7 +36,7 @@ The data are extracted to be skewed toward Italian language with a ration of one
34
 
35
  You can run conversational inference using the Transformers pipeline abstraction or by leveraging the Auto classes with the generate() function.
36
 
37
- Make sure to update your transformers installation via pip install --upgrade transformers.
38
 
39
  ```python
40
  import transformers
@@ -49,6 +51,8 @@ pipeline = transformers.pipeline(
49
  pipeline("Cosa si può fare in una bella giornata di sole?")
50
  ```
51
 
 
 
52
  ## Citation
53
 
54
  If you use any part of this work, please consider citing the paper as follows:
 
5
  license: apache-2.0
6
  pipeline_tag: text-generation
7
  library_name: transformers
8
+ base_model:
9
+ - mistralai/Mistral-7B-v0.1
10
  ---
11
 
12
  # Mistral-7B-v0.1-Italian-RANDOM
 
18
 
19
  The **Mistral-7B-v0.1-Adapted** collection of large language models (LLMs), is a collection of adapted generative models in 7B (text in/text out), adapted models from **Mistral-7B-Base-v0.1**.
20
 
21
+ *Mistral-v0.1-Italian-RANDOM* is a continually trained mistral model, after tokenizer substitution.
22
 
23
  The tokenizer of this models after adaptation is the same of [Minverva-3B](https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0).
24
 
25
  **Model developer:** SapienzaNLP, ISTI-CNR, ILC-CNR
26
 
27
+ **Model Architecture:** Mistral-7B-v0.1-Adapted are auto-regressive language models that uses an optimized transformer architecture.
28
 
29
  ## Data used for the adaptation
30
 
 
36
 
37
  You can run conversational inference using the Transformers pipeline abstraction or by leveraging the Auto classes with the generate() function.
38
 
39
+ Make sure to update your transformers installation via `pip install --upgrade transformers`.
40
 
41
  ```python
42
  import transformers
 
51
  pipeline("Cosa si può fare in una bella giornata di sole?")
52
  ```
53
 
54
+ Code: https://github.com/SapienzaNLP/sava
55
+
56
  ## Citation
57
 
58
  If you use any part of this work, please consider citing the paper as follows: