Avg. Total Time
33.95s
Avg. TTFT
16.76s
Avg. Prefill TPS
1752.22
Avg. Gen TPS
19.11
Context Size
32768
Quantization
r64
Engine
aphrodite
Creation Method
Merge
Model Type
Llama70B
Chat Template
Llama 3
Reasoning
No
Vision
No
Parameters
70B
Added At
4/11/2025
base_model:
~ We are Legion...

My biggest merge yet, consisting of a total of 20 specially curated models. My methodology in approaching this was to create 5 highly specialized models:
These five models went through a series of iterations until I got something I thought worked well and then combined them to make LEGION.
The full list of models used in this merge is below:
Recommended settings:
Temp 1.0
Min P 0.02
Because of the nature of this sort of 'Hyper Multi Model Merge', my recommendation is not to run this on anything lower than a Q5 quant.
If you enjoy my work, please consider supporting me, It helps me make more models like this! Support on KO-FI <3
This is a merge of pre-trained language models created using mergekit.
This model was merged using the DARE TIES merge method using TareksLab/L-BASE-V1 as a base.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
models:
- model: TareksLab/L2-MERGE2a
parameters:
weight: 0.20
density: 0.5
- model: TareksLab/L2-MERGE4
parameters:
weight: 0.20
density: 0.5
- model: TareksLab/L-BASE-V1
parameters:
weight: 0.20
density: 0.5
- model: TareksLab/L2-MERGE3
parameters:
weight: 0.20
density: 0.5
- model: TareksLab/L2-MERGE1
parameters:
weight: 0.20
density: 0.5
merge_method: dare_ties
base_model: TareksLab/L-BASE-V1
parameters:
normalize: false
out_dtype: bfloat16
chat_template: llama3
tokenizer:
source: base