Avg. Total Time
12.97s
Avg. TTFT
6.69s
Avg. Prefill TPS
791.78
Avg. Gen TPS
21.91
Context Size
32768
Quantization
r64
Engine
aphrodite
Creation Method
Merge
Model Type
Llama70B
Chat Template
Llama 3
Reasoning
No
Vision
No
Parameters
70B
Added At
2/2/2025
base_model:
I don't even know anymore

RP-model with good dialogue flow and some creative input.
Tried to make this thing have less slop than the previous iteration. Didn't work out too well, but the NSFW parts are a little more elaborate than before. So yeah, it's an improvement.
Prompt format: Llama3 OR Llama3 Context and ChatML Instruct
Settings: This kinda works but I'm weird
dtype: bfloat16
tokenizer_source: base
merge_method: nuslerp
parameters:
nuslerp_row_wise: true
models:
- model: SicariusSicariiStuff/Negative_LLAMA_70B
parameters:
weight:
- filter: v_proj
value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0]
- filter: o_proj
value: [1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1]
- filter: up_proj
value: [1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1]
- filter: gate_proj
value: [0, 0, 1, 1, 1, 1, 1, 1, 1, 0, 0]
- filter: down_proj
value: [0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0]
- value: [0.2, 0.35, 0.4, 0.35, 0.2]
- model: Nohobby/L3.3-Prikol-70B-v0.2
parameters:
weight:
- filter: v_proj
value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1]
- filter: o_proj
value: [0, 1, 0, 1, 1, 1, 1, 1, 0, 0, 0]
- filter: up_proj
value: [0, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0]
- filter: gate_proj
value: [1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1]
- filter: down_proj
value: [1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 1]
- value: [0.8, 0.65, 0.6, 0.65, 0.8]