wassname commited on
Commit
b7e9d7d
1 Parent(s): d075870

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -49,6 +49,9 @@ We want the good ending, not the bad one.
49
 
50
  ## Metrics
51
 
 
 
 
52
  ```sh
53
  perplexity -m lmstudio-community/Meta-Llama-3-8B-Instruct-Q6_K.gguf -b 32 -c 512 -f wiki.test.raw
54
  # Final estimate: PPL = 7.5588 +/- 0.05599
@@ -58,7 +61,15 @@ perplexity -m cognitivecomputations/dolphin-2.9-llama3-8b-q5_K_M.gguf -b 32 -c 5
58
  # Final estimate: PPL = 9.9277 +/- 0.08261
59
  ```
60
 
61
- So yes this model edit does increase the perplexity :(. Perhaps if we didn't edit so many layers it would be better. It seems better than fine tuning (in the case of early dolphin versions)
 
 
 
 
 
 
 
 
62
 
63
  ---
64
  license: llama3
 
49
 
50
  ## Metrics
51
 
52
+
53
+ Meausing the gguf, there is a difference in perplexity
54
+
55
  ```sh
56
  perplexity -m lmstudio-community/Meta-Llama-3-8B-Instruct-Q6_K.gguf -b 32 -c 512 -f wiki.test.raw
57
  # Final estimate: PPL = 7.5588 +/- 0.05599
 
61
  # Final estimate: PPL = 9.9277 +/- 0.08261
62
  ```
63
 
64
+ Measuring it in the original huggingface format, the increase is much small
65
+ (as a %, the absolute values are not comparable with the ones above as there are differences in measurement)
66
+
67
+ model | perplexity|
68
+ |-|-|
69
+ |base |295.462970|
70
+ |orthogonalized |309.856348|
71
+
72
+ So yes this model edit does increase the perplexity :(. But more investigation is needed.
73
 
74
  ---
75
  license: llama3