Llama-3.3-70B-GeneticLemonade-Final

Creative Model

View on Hugging FaceBack to Models

Hourly Usage

Performance Metrics

Avg. Total Time

5.15s

Avg. TTFT

4.41s

Avg. Prefill TPS

8.16

Avg. Gen TPS

14.88

Model Information

Context Size

32768

Quantization

r64

Engine

aphrodite

Creation Method

Merge

Model Type

Llama70B

Chat Template

Llama 3

Reasoning

No

Vision

No

Parameters

70B

Added At

3/16/2025


base_model:

  • LatitudeGames/Wayfarer-Large-70B-Llama-3.3
  • Sao10K/L3.3-70B-Euryale-v2.3
  • crestf411/L3.1-nemotron-sunfall-v0.7.0
  • SicariusSicariiStuff/Negative_LLAMA_70B
  • EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
  • meta-llama/Llama-3.3-70B-Instruct
  • OpenBuddy/openbuddy-llama3.3-70b-v24.1-131k
  • nbeerbower/llama3.1-kartoffeldes-70B
  • tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3 library_name: transformers tags:
  • mergekit
  • merge license: llama3

Genetic Lemonade Final

image/png

Inspired to learn how to merge by the Nevoria series from SteelSkull.

This model is the second result of the Genetic Lemonade series.

Model Comparison

Designed for RP and creative writing, all three models are focused around striking a balance between writing style, creativity and intelligence. The basic differences between the models are below.

VersionStrengthWeakness
UnleashedWell balancedSomewhat censored
FinalFully uncensoredLeast intelligent
SunsetWell balanced, most intelligentGPTisms / weakest writing style

SillyTavern Settings

Llam@ception recommended for sane defaults if unsure, import them to SillyTavern and they're plug n play.

Sampler Settings

  • Temp: 0.9-1.0
  • MinP: 0.03-0.05
  • Dry: 0.8, 1.75, 4

Temperature last, neutralize other samplers. This model natively strikes a balance of creativity & intelligence.

Instruct

Llama-3-Instruct-Names but you will need to uncheck "System same as user".

Quants

GGUF

EXL2

Merge Details

Merge Method

This model was merged using the SCE merge method.

The base aims to build a strong general purpose model using high performing models that are trained on various datasets from different languages / cultures. This is to reduce the chance of the same datasets appearing multiple times to build natural creativity into L3.3 The second merge aims to impart specific RP / creative writing knowledge, again focusing on trying to find high performing models that use or likely use different datasets.

Base_6_v2

models:
  - model: OpenBuddy/openbuddy-llama3.3-70b-v24.1-131k
  - model: nbeerbower/llama3.1-kartoffeldes-70B
  - model: tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
  - model: SicariusSicariiStuff/Negative_LLAMA_70B
select_topk: .15
merge_method: sce
base_model: meta-llama/Llama-3.3-70B-Instruct
out_dtype: bfloat16
dype: float32
tokenizer:
  source: base

Genetic Lemonade Final

models:
  - model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
  - model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3
  - model: crestf411/L3.1-nemotron-sunfall-v0.7.0
  - model: SicariusSicariiStuff/Negative_LLAMA_70B
  - model: Sao10K/L3.3-70B-Euryale-v2.3
merge_method: sce
base_model: ./Base_6_v2
select_topk: 0.15
out_dtype: bfloat16
dype: float32
tokenizer:
  source: union