This merge required the
enable_fix_mistral_regex_true.md patch for tokenizer stability.
The following YAML configuration was used to produce this model:
# --copy-tokenizer --allow-crimes --out-shard-size 5B --trust-remote-code --lazy-unpickle --random-seed 420 --cuda --fix-mistral-regex
base_model: B:/12B/models--KOOWEEYUS--BlackSheep-RP-12B
architecture: MistralForCausalLM
merge_method: slerp
slices:
- sources:
- model: B:/12B/models--KOOWEEYUS--BlackSheep-RP-12B
layer_range: [0, 40]
- model: B:/12B/models--dphn--dolphin-2.9.3-mistral-nemo-12b
layer_range: [0, 40]
parameters:
t: 0.5
dtype: float32
out_dtype: bfloat16
tokenizer:
source: base
chat_template: auto
name: ๐ฌ BlackDolphin 12B