Llama-3.3-70B-Dungeonmaster-V2.2-Expanded

Creative Model

View on Hugging FaceBack to Models

Hourly Usage

Performance Metrics

Avg. Total Time

58.65s

Avg. TTFT

57.97s

Avg. Prefill TPS

4.33

Avg. Gen TPS

20.66

Model Information

Context Size

32768

Quantization

r64

Engine

aphrodite

Creation Method

Merge

Model Type

Llama70B

Chat Template

Llama 3

Reasoning

No

Vision

No

Parameters

70B

Added At

3/18/2025


base_model:

  • TareksLab/TRP-BASE-SCE-V1-70B
  • ArliAI/Llama-3.1-70B-ArliAI-RPMax-v1.3
  • TheDrummer/Fallen-Llama-3.3-R1-70B-v1
  • EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
  • SicariusSicariiStuff/Negative_LLAMA_70B
  • LatitudeGames/Wayfarer-Large-70B-Llama-3.3
  • TheDrummer/Anubis-70B-v1
  • Sao10K/70B-L3.3-mhnnn-x1 library_name: transformers tags:
  • mergekit
  • merge
  • not-for-all-audiences license: llama3.3

image/png

V2.2 of Dungeonmaster (Very good at following prompts and quite unhinged), I decided to move away from the R1 base here, because I feel it the pros dont necessarily outweigh the cons. For the V2.X series I decided to go for a custom uncensored base.

Shoutout to Thana Alt from the Beaver AI discord who thoroughly tested this model and was able to get some interesting results (spoiler below)

⚠️ **Warning: Gratuitous violence**

image/webp

(Thana's Silly Tavern advanced formatting settings are in the models files)

The sweet spot for the important sampler settings seems to be around:

Temp: 0.8
Min P: 0.02

Dungeonmaster is meant to be specifically for creative roleplays with stakes and consequences using the following curated models:

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Linear DELLA merge method using TareksLab/L3.3-TRP-BASE-80-70B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: TheDrummer/Fallen-Llama-3.3-R1-70B-v1
    parameters:
      weight: 0.12
      density: 0.7
  - model: ArliAI/Llama-3.3-70B-ArliAI-RPMax-v1.4
    parameters:
      weight: 0.12
      density: 0.7
  - model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
    parameters:
      weight: 0.12
      density: 0.7
  - model: TheDrummer/Anubis-70B-v1
    parameters:
      weight: 0.12
      density: 0.7
  - model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3
    parameters:
      weight: 0.13
      density: 0.7
  - model: SicariusSicariiStuff/Negative_LLAMA_70B
    parameters:
      weight: 0.13
      density: 0.7
  - model: Sao10K/70B-L3.3-mhnnn-x1
    parameters:
      weight: 0.13
      density: 0.7
  - model: TareksLab/L3.3-TRP-BASE-80-70B
    parameters:
      weight: 0.13
      density: 0.7
merge_method: della_linear
base_model: TareksLab/L3.3-TRP-BASE-80-70B
parameters:
  epsilon: 0.2
  lambda: 1.1
  normalize: false
  int8_mask: true
dtype: bfloat16
chat_template: llama3
tokenizer:
 source: base

Support on KO-FI <3