Avg. Total Time
24.34s
Avg. TTFT
9.86s
Avg. Prefill TPS
267.99
Avg. Gen TPS
24.14
Context Size
32768
Quantization
r64
Engine
aphrodite
Creation Method
Merge
Model Type
Llama70B
Chat Template
Llama 3
Reasoning
No
Vision
No
Parameters
70B
Added At
1/16/2025
base_model:
Good Evening.
New model mix because I got frustrated of dealing with wrangling with parameters and I chalked it up to a "Seems like a model issue".
Going back to merge stock since i dont feel like experimenting and want to try something that vibes well out of the box.
ChatML works. Same goes for L3 chat.
This model was merged using the Model Stock merge method using Llama-3.3-70B-Instruct as a base.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
models:
- model: Blackroot/Mirai-3.0-70B
- model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
- model: TheDrummer/Anubis-70B-v1
- model: Sao10K/L3.3-70B-Euryale-v2.3
- model: Sao10K/70B-L3.3-Cirrus-x1
- model: nitky/Llama-3.3-SuperSwallowX-70B-Instruct-v0.1
- model: KaraKaraWitch/Llama-MiraiFanfare-2-3.3-70B
- model: Undi95/Sushi-v1.4
- model: Nohobby/L3.3-Prikol-70B-v0.2
merge_method: model_stock
base_model: Llama-3.3-70B-Instruct
parameters:
normalize: true
dtype: bfloat16