Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,98 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- llama
|
| 4 |
+
- alpaca
|
| 5 |
+
- vicuna
|
| 6 |
+
- uncensored
|
| 7 |
+
- merge
|
| 8 |
+
- mix
|
| 9 |
+
- airoboros
|
| 10 |
+
- openorca
|
| 11 |
+
- orcamini
|
| 12 |
+
- orca
|
| 13 |
+
- instruct
|
| 14 |
+
- mixtune
|
| 15 |
+
datasets:
|
| 16 |
+
- Open-Orca/OpenOrca
|
| 17 |
+
- anon8231489123/ShareGPT_Vicuna_unfiltered
|
| 18 |
+
- jondurbin/airoboros-uncensored
|
| 19 |
+
language:
|
| 20 |
+
- en
|
| 21 |
+
metrics:
|
| 22 |
+
- accuracy
|
| 23 |
+
pipeline_tag: text-generation
|
| 24 |
+
---
|
| 25 |
+
|
| 26 |
+
## 13B-Ouroboros
|
| 27 |
+
Ouroboros is an experimental model based on Meta's LLaMA [v1] 13B base model using a custom merging script that optimizes
|
| 28 |
+
per-layer merging based on a given dataset. Ouroboros is optimized against the PTB text only validation dataset, scoring
|
| 29 |
+
~26.31 according to internal evaluation (6 samples, sequence length 1024; this testing is not empirical, it's a part of
|
| 30 |
+
the random search algorithm). Testing, evaluating, and remixing this model is absolutely permissible and even encouraged
|
| 31 |
+
(within the bounds of Meta's LLaMAv1 license agreement); the more feedback the better we can tune our process! 😊
|
| 32 |
+
|
| 33 |
+
When the mix tuning system has reached a certain point of maturity it will be released open source.
|
| 34 |
+
|
| 35 |
+
## Composition:
|
| 36 |
+
Ouroboros is comprised of 40 layers [LLaMAv1 13B standard] mixed at optimized
|
| 37 |
+
ratios VS the PTB dataset for lowest perplexity score. Listed below are the
|
| 38 |
+
paired models and ratios merged per layer.
|
| 39 |
+
|
| 40 |
+
Tier One Merge:
|
| 41 |
+
|
| 42 |
+
13B-airoboros-gpt4-1.4 > 13B-orca_mini_v2
|
| 43 |
+
|
| 44 |
+
[0.22, 0.85, 0.89, 0.98, 0.3, 0.41, 0.71, 0.83, 0.32, 0.1, 0.44, 0.6, 0.53, 0.15, 0.86, 0.79, 0.93, 0.02, 0.19, 0.82, 0.01, 0.52, 0.07, 0.27, 0.73, 0.86, 0.08, 0.67, 0.42, 0.28, 0.37, 0.08, 0.95, 0.68, 0.45, 0.08, 0.7, 0.93, 0.96, 0.43]
|
| 45 |
+
|
| 46 |
+
13B-gpt4-x-alpaca > 13B-Vicuna-cocktail
|
| 47 |
+
|
| 48 |
+
[0.65, 0.94, 0.98, 0.87, 0.28, 0.64, 0.73, 0.7, 0.95, 0.89, 0.84, 0.9, 0.59, 0.92, 0.28, 0.61, 0.88, 0.73, 0.34, 0.85, 0.98, 0.05, 0.74, 0.92, 0.5, 0.78, 0.26, 0.4, 0.27, 0.65, 0.71, 0.7, 0.8, 0.93, 0.36, 0.03, 0.45, 0.39, 0.77, 0.06]
|
| 49 |
+
|
| 50 |
+
Tier Two Merge:
|
| 51 |
+
|
| 52 |
+
[13B-airoboros-gpt4-1.4 + 13B-orca_mini_v2] offspring + [13B-gpt4-x-alpaca > 13B-Vicuna-cocktail] offspring
|
| 53 |
+
|
| 54 |
+
[0.2, 0.83, 0.24, 0.03, 0.37, 0.62, 0.02, 0.82, 0.65, 0.63, 0.45, 0.65, 0.48, 0.45, 0.24, 0.76, 0.06, 0.31, 0.45, 0.86, 0.23, 0.99, 0.93, 0.84, 0.96, 0.53, 0.95, 0.32, 0.19, 0.06, 0.4, 0.08, 0.62, 0.4, 0.26, 0.12, 0.16, 0.91, 0.14, 0.0]
|
| 55 |
+
|
| 56 |
+
Result:
|
| 57 |
+
|
| 58 |
+
13B-Ouroboros, a model that seems uncensored and highly competent. So far only Alpaca instruction promting has been tested and seems to work solidly well.
|
| 59 |
+
|
| 60 |
+
## Use:
|
| 61 |
+
|
| 62 |
+
Alpaca's instruct format can be used to do many things, including control of the terms of behavior
|
| 63 |
+
between a user and a response from an agent in chat. Below is an example of a command injected into
|
| 64 |
+
memory.
|
| 65 |
+
|
| 66 |
+
```
|
| 67 |
+
### Instruction:
|
| 68 |
+
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
|
| 69 |
+
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
|
| 70 |
+
whatever quest or other information to keep consistent in the interaction).
|
| 71 |
+
|
| 72 |
+
### Response:
|
| 73 |
+
{an empty new line here}
|
| 74 |
+
```
|
| 75 |
+
|
| 76 |
+
## Language Models Used Credits:
|
| 77 |
+
|
| 78 |
+
13B-airoboros-gpt4-1.4 by jondurbin
|
| 79 |
+
|
| 80 |
+
https://huggingface.co/jondurbin/airoboros-13b-gpt4-1.4
|
| 81 |
+
|
| 82 |
+
13B-orca_mini_v2 by psmathur
|
| 83 |
+
|
| 84 |
+
https://huggingface.co/psmathur/orca_mini_v2_13b
|
| 85 |
+
|
| 86 |
+
13B-gpt4-x-alpaca by chavinlo
|
| 87 |
+
|
| 88 |
+
https://huggingface.co/chavinlo/gpt4-x-alpaca
|
| 89 |
+
|
| 90 |
+
13B-Vicuna-cocktail by reeducator
|
| 91 |
+
|
| 92 |
+
https://huggingface.co/reeducator/vicuna-13b-cocktail
|
| 93 |
+
|
| 94 |
+
Also thanks to Meta for LLaMA.
|
| 95 |
+
|
| 96 |
+
Each model and LoRA was hand picked and considered for what it could contribute to this ensemble.
|
| 97 |
+
Thanks to each and every one of you for your incredible work developing some of the best things
|
| 98 |
+
to come out of this community.
|