m-nagoudi commited on
Commit
d31959f
1 Parent(s): b5156bf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -1,17 +1,13 @@
1
  # AraT5-base
2
  # AraT5: Text-to-Text Transformers for Arabic Language Generation
3
 
4
- <img src="AraT5_CR_new.png" alt="AraT5" width="55%" height="45%" align="right"/>
5
 
6
  This is the repository accompanying our paper [AraT5: Text-to-Text Transformers for Arabic Language Understanding and Generation](https://arxiv.org/abs/2109.12068). In this is the repository we introduce:
7
  * Introduce **AraT5<sub>MSA</sub>**, **AraT5<sub>Tweet</sub>**, and **AraT5**: three powerful Arabic-specific text-to-text Transformer based models;
8
  * Introduce **ARGEN**: A new benchmark for Arabic language generation and evaluation for four Arabic NLP tasks, namely, ```machine translation```, ```summarization```, ```news title generation```, ```question generation```, , ```paraphrasing```, ```transliteration```, and ```code-switched translation```.
9
  * Evaluate ```AraT5``` models on ```ARGEN``` and compare against available language models.
10
 
11
- Our models establish new state-of-the-art (SOTA) on several publicly available datasets.
12
- Our language models are publicaly available for research (see below).
13
-
14
- The rest of this repository provides more information about our new language models, benchmark, and experiments.
15
 
16
  ---
17
  # How to use AraT5 models
@@ -56,14 +52,16 @@ AraT5 Pytorch and TensorFlow checkpoints are available on the Huggingface websit
56
 
57
  If you use our models (Arat5-base, Arat5-msa-base, Arat5-tweet-base, Arat5-msa-small, or Arat5-tweet-small ) for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows (to be updated):
58
  ```bibtex
59
- @inproceedings{araT5-2021,
60
- title = "{AraT5: Text-to-Text Transformers for Arabic Language Understanding and Generation",
61
  author = "Nagoudi, El Moatez Billah and
62
  Elmadany, AbdelRahim and
63
  Abdul-Mageed, Muhammad",
64
- booktitle = "https://arxiv.org/abs/2109.12068",
65
- month = aug,
66
- year = "2021"}
 
 
67
  ```
68
 
69
  ## Acknowledgments
 
1
  # AraT5-base
2
  # AraT5: Text-to-Text Transformers for Arabic Language Generation
3
 
4
+ <img src="https://huggingface.co/UBC-NLP/AraT5-base/resolve/main/AraT5_CR_new.png" alt="AraT5" width="45%" height="35%" align="right"/>
5
 
6
  This is the repository accompanying our paper [AraT5: Text-to-Text Transformers for Arabic Language Understanding and Generation](https://arxiv.org/abs/2109.12068). In this is the repository we introduce:
7
  * Introduce **AraT5<sub>MSA</sub>**, **AraT5<sub>Tweet</sub>**, and **AraT5**: three powerful Arabic-specific text-to-text Transformer based models;
8
  * Introduce **ARGEN**: A new benchmark for Arabic language generation and evaluation for four Arabic NLP tasks, namely, ```machine translation```, ```summarization```, ```news title generation```, ```question generation```, , ```paraphrasing```, ```transliteration```, and ```code-switched translation```.
9
  * Evaluate ```AraT5``` models on ```ARGEN``` and compare against available language models.
10
 
 
 
 
 
11
 
12
  ---
13
  # How to use AraT5 models
 
52
 
53
  If you use our models (Arat5-base, Arat5-msa-base, Arat5-tweet-base, Arat5-msa-small, or Arat5-tweet-small ) for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows (to be updated):
54
  ```bibtex
55
+ @inproceedings{nagoudi-2022-arat5,
56
+ title = "{AraT5: Text-to-Text Transformers for Arabic Language Generation",
57
  author = "Nagoudi, El Moatez Billah and
58
  Elmadany, AbdelRahim and
59
  Abdul-Mageed, Muhammad",
60
+ booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics",
61
+ month = May,
62
+ year = "2022",
63
+ address = "Online",
64
+ publisher = "Association for Computational Linguistics",
65
  ```
66
 
67
  ## Acknowledgments