AndreasThinks commited on
Commit
520314b
1 Parent(s): ca6faff

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -46
README.md CHANGED
@@ -9,8 +9,52 @@ model-index:
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
16
  <details><summary>See axolotl config</summary>
@@ -84,50 +128,6 @@ special_tokens:
84
  </details><br>
85
 
86
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/andreasthinks/mistral-nemo-welsh/runs/syq2m3vr)
87
- # mistral-7b-english-welsh-translate
88
-
89
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the [Welsh Government Alpaca Welsh-English Instructions](https://huggingface.co/datasets/AndreasThinks/welsh-translation-instruction/blob/main/README.md) dataset.
90
-
91
- This model is trained for English-Welsh translation (in any direction), with a focus on government documents, using Markdown formatting.
92
-
93
- To ensure the highest quality translations, use the Alpaca instruction prompt format with the below structure.
94
-
95
- ```
96
- ### Instruction: {instruction}
97
-
98
- ### Input: {input}
99
-
100
- ### Response:
101
-
102
- ```
103
-
104
- Your instruction should be "Translate the text from English to Welsh." (or vice versa).
105
-
106
- The model is also available [quantized as GGUF](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF).
107
-
108
- ## Running the model
109
-
110
- The model is intended to be run locally, ideally using [Text generation web UI](https://github.com/oobabooga/text-generation-webui) to ensure correct prompt structure.
111
-
112
- Start the UI as instructed for your system.
113
-
114
- - In the "Model" tab, download either this model or [the quantized version](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF). Once the download is complete, load the model.
115
- - In the "Parameters" tab, under "Generation", set "auto_max_new_tokens" to maximum, and "Ban the eos_token" to True. In "Custom stopping strings", add "### Input"
116
- - In the "Notebook" tab, make sure you are using the "Alpaca-with-input" prompt. Set the instruction as "Translate the text from Welsh to English." (or vice versa).#
117
- - Add the text you would like to translate (replacing "Input"), and hit "generate"
118
-
119
- Performance may start to degrade past a certain context window (especially if using the quantized models). Convert in chunks of under 1000 words to avoid these issues.
120
-
121
- ## LLM Evals
122
-
123
- Thanks to [YALL - Yet Another LLM Leaderboard](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard)
124
-
125
- | Model |AGIEval|TruthfulQA|Bigbench|
126
- |-------------------------------------------------------------------------------------------------------------|------:|---------:|-------:|
127
- |[mistral-7b-english-welsh-translate](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate)| 35.31| 54.5| 38.4|
128
-
129
-
130
- ## Training procedure
131
 
132
  ### Training hyperparameters
133
 
 
9
  results: []
10
  ---
11
 
12
+ # mistral-7b-english-welsh-translate
13
+
14
+
15
+
16
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the [Welsh Government Alpaca Welsh-English Instructions](https://huggingface.co/datasets/AndreasThinks/welsh-translation-instruction/blob/main/README.md) dataset.
17
+
18
+ This model is trained for English-Welsh translation (in any direction), with a focus on government documents, using Markdown formatting.
19
+
20
+ To ensure the highest quality translations, use the Alpaca instruction prompt format with the below structure.
21
+
22
+ ```
23
+ ### Instruction: {instruction}
24
+
25
+ ### Input: {input}
26
+
27
+ ### Response:
28
+
29
+ ```
30
+
31
+ Your instruction should be "Translate the text from English to Welsh." (or vice versa).
32
+
33
+ The model is also available [quantized as GGUF](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF).
34
+
35
+ ## Running the model
36
+
37
+ The model is intended to be run locally, ideally using [Text generation web UI](https://github.com/oobabooga/text-generation-webui) to ensure correct prompt structure.
38
+
39
+ Start the UI as instructed for your system.
40
+
41
+ - In the "Model" tab, download either this model or [the quantized version](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF). Once the download is complete, load the model.
42
+ - In the "Parameters" tab, under "Generation", set "auto_max_new_tokens" to maximum, and "Ban the eos_token" to True. In "Custom stopping strings", add "### Input"
43
+ - In the "Notebook" tab, make sure you are using the "Alpaca-with-input" prompt. Set the instruction as "Translate the text from Welsh to English." (or vice versa).#
44
+ - Add the text you would like to translate (replacing "Input"), and hit "generate"
45
+
46
+ Performance may start to degrade past a certain context window (especially if using the quantized models). Convert in chunks of under 1000 words to avoid these issues.
47
+
48
+ ## LLM Evals
49
+
50
+ Thanks to [YALL - Yet Another LLM Leaderboard](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard)
51
+
52
+ | Model |AGIEval|TruthfulQA|Bigbench|
53
+ |-------------------------------------------------------------------------------------------------------------|------:|---------:|-------:|
54
+ |[mistral-7b-english-welsh-translate](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate)| 35.31| 54.5| 38.4|
55
+
56
+
57
+ ## Training procedure
58
 
59
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
60
  <details><summary>See axolotl config</summary>
 
128
  </details><br>
129
 
130
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/andreasthinks/mistral-nemo-welsh/runs/syq2m3vr)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
 
132
  ### Training hyperparameters
133