Update README.md
Browse files
README.md
CHANGED
@@ -16,6 +16,8 @@ widget:
|
|
16 |
|
17 |
TookaBERT models are a family of encoder models trained on Persian in two sizes base and large. These Models pre-trained on over 500GB of Persian data including a variety of topics such as News, Blogs, Forums, Books, etc. They pre-trained with the MLM (WWM) objective using two context lengths.
|
18 |
|
|
|
|
|
19 |
## How to use
|
20 |
|
21 |
You can use this model directly for Masked Language Modeling using the provided code below.
|
|
|
16 |
|
17 |
TookaBERT models are a family of encoder models trained on Persian in two sizes base and large. These Models pre-trained on over 500GB of Persian data including a variety of topics such as News, Blogs, Forums, Books, etc. They pre-trained with the MLM (WWM) objective using two context lengths.
|
18 |
|
19 |
+
For more information you can read our paper on [arXiv](https://arxiv.org/abs/2407.16382).
|
20 |
+
|
21 |
## How to use
|
22 |
|
23 |
You can use this model directly for Masked Language Modeling using the provided code below.
|