Avg. Total Time
22.58s
Avg. TTFT
12.60s
Avg. Prefill TPS
645.58
Avg. Gen TPS
19.40
Context Size
32768
Quantization
r64
Engine
aphrodite
Creation Method
Merge
Model Type
Llama70B
Chat Template
Llama 3
Reasoning
No
Vision
No
Parameters
70B
Added At
1/18/2025
base_model: [] library_name: transformers tags:
This is a merge of pre-trained language models created using mergekit.
This model was merged using the SLERP merge method.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
models:
- model: Sao10K/70B-L3.3-Cirrus-x1
- model: Doctor-Shotgun/L3.3-70B-Magnum-v4-SE
merge_method: slerp
base_model: Sao10K/70B-L3.3-Cirrus-x1
parameters:
t:
- value: 0.5
dtype: bfloat16