Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,8 @@ language:
|
|
5 |
license: apache-2.0
|
6 |
pipeline_tag: text-generation
|
7 |
library_name: transformers
|
|
|
|
|
8 |
---
|
9 |
|
10 |
# Mistral-7B-v0.1-Italian-RANDOM
|
@@ -16,13 +18,13 @@ library_name: transformers
|
|
16 |
|
17 |
The **Mistral-7B-v0.1-Adapted** collection of large language models (LLMs), is a collection of adapted generative models in 7B (text in/text out), adapted models from **Mistral-7B-Base-v0.1**.
|
18 |
|
19 |
-
*Mistral-v0.1-Italian-RANDOM* is a
|
20 |
|
21 |
The tokenizer of this models after adaptation is the same of [Minverva-3B](https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0).
|
22 |
|
23 |
**Model developer:** SapienzaNLP, ISTI-CNR, ILC-CNR
|
24 |
|
25 |
-
**Model Architecture:** Mistral-7B-v0.1-Adapted
|
26 |
|
27 |
## Data used for the adaptation
|
28 |
|
@@ -34,7 +36,7 @@ The data are extracted to be skewed toward Italian language with a ration of one
|
|
34 |
|
35 |
You can run conversational inference using the Transformers pipeline abstraction or by leveraging the Auto classes with the generate() function.
|
36 |
|
37 |
-
Make sure to update your transformers installation via pip install --upgrade transformers
|
38 |
|
39 |
```python
|
40 |
import transformers
|
@@ -49,6 +51,8 @@ pipeline = transformers.pipeline(
|
|
49 |
pipeline("Cosa si può fare in una bella giornata di sole?")
|
50 |
```
|
51 |
|
|
|
|
|
52 |
## Citation
|
53 |
|
54 |
If you use any part of this work, please consider citing the paper as follows:
|
|
|
5 |
license: apache-2.0
|
6 |
pipeline_tag: text-generation
|
7 |
library_name: transformers
|
8 |
+
base_model:
|
9 |
+
- mistralai/Mistral-7B-v0.1
|
10 |
---
|
11 |
|
12 |
# Mistral-7B-v0.1-Italian-RANDOM
|
|
|
18 |
|
19 |
The **Mistral-7B-v0.1-Adapted** collection of large language models (LLMs), is a collection of adapted generative models in 7B (text in/text out), adapted models from **Mistral-7B-Base-v0.1**.
|
20 |
|
21 |
+
*Mistral-v0.1-Italian-RANDOM* is a continually trained mistral model, after tokenizer substitution.
|
22 |
|
23 |
The tokenizer of this models after adaptation is the same of [Minverva-3B](https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0).
|
24 |
|
25 |
**Model developer:** SapienzaNLP, ISTI-CNR, ILC-CNR
|
26 |
|
27 |
+
**Model Architecture:** Mistral-7B-v0.1-Adapted are auto-regressive language models that uses an optimized transformer architecture.
|
28 |
|
29 |
## Data used for the adaptation
|
30 |
|
|
|
36 |
|
37 |
You can run conversational inference using the Transformers pipeline abstraction or by leveraging the Auto classes with the generate() function.
|
38 |
|
39 |
+
Make sure to update your transformers installation via `pip install --upgrade transformers`.
|
40 |
|
41 |
```python
|
42 |
import transformers
|
|
|
51 |
pipeline("Cosa si può fare in una bella giornata di sole?")
|
52 |
```
|
53 |
|
54 |
+
Code: https://github.com/SapienzaNLP/sava
|
55 |
+
|
56 |
## Citation
|
57 |
|
58 |
If you use any part of this work, please consider citing the paper as follows:
|