Avg. Total Time
69.01s
Avg. TTFT
19.14s
Avg. Prefill TPS
537.42
Avg. Gen TPS
9.38
Context Size
262144
Quantization
r64
Engine
vllm
Creation Method
LoRA
Model Type
Qwen35
Chat Template
Qwen3.5
Reasoning
Yes
Vision
Yes
Parameters
27B
Added At
4/4/2026
license: apache-2.0 datasets:
A writing & roleplay finetune of Qwen3.5 27B. The primary emphasis is on writing quality as it strongly generalizes across both domains. This model is also trained from ConicCat/Qwen3.5-Antirep-27B to mitigate repetition issues.
The basic idea is to use a curriculum learning setup to overcome the lack of high quality roleplay data by first training on lower quality roleplay data, then training on higher quality writing data. Starting from ConicCat/Qwen3.5-Antirep-27B, the model was trained on a roughly equal mixture of instruct / roleplay / writing data for three epochs. The model was then trained for eleven epochs on a smaller dataset of short story anthologies by critically acclaimed authors.
<think>\n\n</think> or {{char}}: prefill. Only non-thinking was trained, but thinking probably still works.0.70.951.05 or a moderate dry setting should suffice.~100k context on 24GB Vram20-24k context with the vulkan backend, although it's pretty tight and may require some fiddling around with open programs e.t.c.