Llama-3.3+(3.1v3.3)-70B-Daybreak-storywriter-v0.4

Creative Model

View on Hugging FaceBack to Models

Hourly Usage

Performance Metrics

Avg. Total Time

11.44s

Avg. TTFT

5.85s

Avg. Prefill TPS

395.48

Avg. Gen TPS

23.66

Model Information

Context Size

32768

Quantization

r64

Engine

aphrodite

Creation Method

LoRA Finetune

Model Type

Llama70B

Chat Template

Llama 3

Reasoning

No

Vision

No

Parameters

70B

Added At

12/22/2024


tags:

  • not-for-all-audiences

Daybreak (2024 May 24) v0.4 LoRA on top of https://huggingface.co/tdrussell/Llama-3-70B-Instruct-Storywriter

Dataset curation to remove slop-perceived expressions continues.

The below regexes return 0 matches. Bold entries are new since v0.3.

  • 'barely above a whisper',
  • 'barely audible',
  • 'shiver([s]?) down',
  • ' ministration',
  • 'audible (["'"]?)p[l]?op',
  • 'can't help but',
  • 'buck([s]?) my ',
  • 'buck([s]?) h[ei][rs] ',
  • '[Dd]espite h[ie][mr]self',
  • 'slick slit',
  • 'whatever it takes',
  • 'unlike anything (s?)he',
  • 'a mix([a-z]*) of',
  • 'wave after wave',
  • 'reckless abandon',
  • '[Mm]aybe, just maybe',
  • 'eyes gleaming',
  • 'mischievously',
  • "couldn't help but",

From testing so far, it feels like temperature 0.8-0.9 is a good starting point. I have mostly tested with everything neutralized. Please give feedback on which parameters work good for you.

EXL2 quants made by kim512 be found in (scroll to bottom for updated quants) https://huggingface.co/crestf411/L3-70B-daybreak-storywriter-v0.4/discussions/2 including the settings used to make them.