|
|
--- |
|
|
library_name: transformers |
|
|
license: llama3 |
|
|
datasets: |
|
|
- zerofata/Roleplay-Anime-Characters |
|
|
- zerofata/Instruct-Anime-CreativeWriting |
|
|
- zerofata/Summaries-Anime-FandomPages |
|
|
base_model: |
|
|
- zerofata/L3.3-GeneticLemonade-Final-70B |
|
|
--- |
|
|
<!DOCTYPE html> |
|
|
<style> |
|
|
body { |
|
|
font-family: sans-serif; |
|
|
color: #f0f0f0; |
|
|
line-height: 1.6; |
|
|
margin: 0; |
|
|
padding: 0; |
|
|
background-color: #1a0f1a; |
|
|
} |
|
|
|
|
|
.lemonade-text { |
|
|
color: #ff3366; |
|
|
position: relative; |
|
|
z-index: 2; |
|
|
margin-left: 0.2em; |
|
|
text-shadow: 0 0 10px #ff3366; |
|
|
} |
|
|
|
|
|
/* Section styling */ |
|
|
.section-container { |
|
|
background-color: rgba(26, 15, 26, 0.7); |
|
|
margin-bottom: 30px; |
|
|
position: relative; |
|
|
overflow: hidden; |
|
|
border-bottom: 1px solid #ff3366; |
|
|
} |
|
|
|
|
|
.section-header { |
|
|
display: flex; |
|
|
align-items: center; |
|
|
background-color: rgba(255, 51, 102, 0.08); |
|
|
padding: 10px 20px; |
|
|
} |
|
|
|
|
|
.section-indicator { |
|
|
width: 8px; |
|
|
height: 20px; |
|
|
background-color: #ff3366; |
|
|
margin-right: 15px; |
|
|
} |
|
|
|
|
|
.section-title { |
|
|
font-family: 'Orbitron', sans-serif; |
|
|
color: #f0f0f0; |
|
|
font-size: 1.3rem; |
|
|
margin: 0; |
|
|
letter-spacing: 2px; |
|
|
text-transform: uppercase; |
|
|
font-weight: 500; |
|
|
} |
|
|
|
|
|
.section-content { |
|
|
padding: 20px; |
|
|
font-family: sans-serif; |
|
|
color: #f0f0f0; |
|
|
line-height: 1.6; |
|
|
} |
|
|
|
|
|
/* Title styling */ |
|
|
.title-container { |
|
|
background-color: #0a0a0a; |
|
|
position: relative; |
|
|
overflow: hidden; |
|
|
margin-bottom: 40px; |
|
|
border-left: 3px solid #ff3366; |
|
|
} |
|
|
|
|
|
.title-wrapper { |
|
|
position: relative; |
|
|
z-index: 2; |
|
|
padding: 25px 20px 30px 30px; |
|
|
font-family: 'Orbitron', sans-serif; |
|
|
} |
|
|
|
|
|
.title-main { |
|
|
color: #f0f0f0; |
|
|
font-size: 2.5rem; |
|
|
font-weight: 700; |
|
|
margin: 0; |
|
|
letter-spacing: 2px; |
|
|
display: inline-block; |
|
|
position: relative; |
|
|
text-transform: uppercase; |
|
|
} |
|
|
|
|
|
.title-prefix { |
|
|
position: relative; |
|
|
z-index: 2; |
|
|
} |
|
|
|
|
|
.title-subtitle { |
|
|
padding-left: 15px; |
|
|
margin-top: 5px; |
|
|
margin-left: 5px; |
|
|
} |
|
|
|
|
|
.subtitle-text { |
|
|
color: #cc0066; |
|
|
font-size: 1.2rem; |
|
|
font-family: 'Orbitron', sans-serif; |
|
|
font-weight: 300; |
|
|
letter-spacing: 3px; |
|
|
text-transform: uppercase; |
|
|
display: inline-block; |
|
|
} |
|
|
|
|
|
.glitchy-overlay { |
|
|
position: absolute; |
|
|
top: 0; |
|
|
left: 0; |
|
|
width: 100%; |
|
|
height: 100%; |
|
|
background-image: repeating-linear-gradient(0deg, rgba(0,0,0,0) 0, rgba(139, 0, 0, 0.1) 1px, rgba(0,0,0,0) 2px); |
|
|
z-index: 1; |
|
|
} |
|
|
|
|
|
/* Data box styling */ |
|
|
.data-box { |
|
|
background-color: rgba(0, 0, 0, 0.4); |
|
|
padding: 15px; |
|
|
border-left: 2px solid #ff3366; |
|
|
margin-bottom: 20px; |
|
|
} |
|
|
|
|
|
.data-row { |
|
|
display: flex; |
|
|
margin-bottom: 8px; |
|
|
} |
|
|
|
|
|
.data-arrow { |
|
|
color: #ff3366; |
|
|
width: 20px; |
|
|
display: inline-block; |
|
|
} |
|
|
|
|
|
.data-label { |
|
|
color: #cc0066; |
|
|
width: 80px; |
|
|
display: inline-block; |
|
|
} |
|
|
|
|
|
/* Subheading styling */ |
|
|
.subheading { |
|
|
color: #cc0066; |
|
|
font-size: 1.1rem; |
|
|
margin-top: 20px; |
|
|
margin-bottom: 15px; |
|
|
font-weight: 400; |
|
|
border-bottom: 1px dashed rgba(204, 0, 102, 0.4); |
|
|
display: inline-block; |
|
|
text-transform: uppercase; |
|
|
letter-spacing: 1px; |
|
|
font-family: 'Orbitron', sans-serif; |
|
|
} |
|
|
|
|
|
/* Links */ |
|
|
a { |
|
|
color: #cc0066; |
|
|
text-decoration: none; |
|
|
} |
|
|
|
|
|
a:hover { |
|
|
text-decoration: underline; |
|
|
color: #ff6600; |
|
|
} |
|
|
|
|
|
/* Container */ |
|
|
.container { |
|
|
max-width: 1200px; |
|
|
margin: 20px auto; |
|
|
padding: 40px 20px; |
|
|
background-color: #0a0a0a; |
|
|
background-image: |
|
|
linear-gradient(rgba(139, 0, 0, 0.12) 1px, transparent 1px), |
|
|
linear-gradient(90deg, rgba(139, 0, 0, 0.12) 1px, transparent 1px); |
|
|
background-size: 20px 20px; |
|
|
min-height: calc(100vh - 40px); |
|
|
border: 1px solid #ff3366; |
|
|
border-radius: 2px; |
|
|
} |
|
|
|
|
|
</style> |
|
|
<html lang="en"> |
|
|
<head> |
|
|
<meta charset="UTF-8"> |
|
|
<meta name="viewport" content="width=device-width, initial-scale=1.0"> |
|
|
<title>GENETIC LEMONADE FINAL v2</title> |
|
|
<link href="https://fonts.googleapis.com/css2?family=Orbitron:wght@400;500;600;700&family=JetBrains+Mono:wght@100;300;400;700&display=swap" rel="stylesheet"> |
|
|
</head> |
|
|
<body> |
|
|
<div class="cyber-grid-bg"></div> |
|
|
|
|
|
<div class="container"> |
|
|
<div class="title-container"> |
|
|
<!-- Glitchy overlay --> |
|
|
<div class="glitchy-overlay"></div> |
|
|
<!-- Main title --> |
|
|
<div class="title-wrapper"> |
|
|
<h1 class="title-main"> |
|
|
<span class="title-prefix">GENETIC</span> |
|
|
<span class="lemonade-text">LEMONADE</span> <!-- Static text with glow --> |
|
|
</h1> |
|
|
<div class="title-subtitle"> |
|
|
<span class="subtitle-text">FINAL v2</span> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
|
|
|
 |
|
|
|
|
|
|
|
|
<div class="section-container"> |
|
|
<div class="section-header"> |
|
|
<div class="section-indicator"></div> |
|
|
<h2 class="section-title">01 // OVERVIEW</h2> |
|
|
</div> |
|
|
<div class="section-content"> |
|
|
<p>Wasn't intending to release another model (so soon at least), but I was testing out some new dataset ideas and thought this model came out pretty nice.</p> |
|
|
<p><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-70B">zerofata/GeneticLemonade-Final</a> SFT QLora finetune.</p> |
|
|
<p>This is an uncensored creative model intended to excel at character driven RP / ERP.</p> |
|
|
<p>This model is designed to provide longer, narrative heavy responses where characters are portrayed accurately and proactively.</p> |
|
|
<p>Compared to Unleashed v3, this model has significantly reduced positivity bias and arguably a nicer writing style. The tradeoff is it swipe heavy, making a few more logical errors and can be a bit too concise at times.</p> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
<div class="section-container"> |
|
|
<div class="section-header"> |
|
|
<div class="section-indicator"></div> |
|
|
<h2 class="section-title">02 // SILLYTAVERN SETTINGS</h2> |
|
|
</div> |
|
|
<div class="section-content"> |
|
|
<p>Play with these, they are not the 'best' settings just a stable baseline.</p> |
|
|
<h3 class="subheading">Recommended Samplers</h3> |
|
|
<div class="data-box"> |
|
|
<div class="data-row"> |
|
|
<span class="data-arrow">></span> |
|
|
<span class="data-label">Temp:</span> |
|
|
<span>0.9 - 1</span> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span class="data-arrow">></span> |
|
|
<span class="data-label">MinP:</span> |
|
|
<span>0.03 - 0.04</span> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span class="data-arrow">></span> |
|
|
<span class="data-label">TopP:</span> |
|
|
<span>0.9 - 1.0</span> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span class="data-arrow">></span> |
|
|
<span class="data-label">Dry:</span> |
|
|
<span>0.8, 1.75, 4</span> |
|
|
</div> |
|
|
</div> |
|
|
<h3 class="subheading">Instruct</h3> |
|
|
<div class="data-box"> |
|
|
<p style="margin: 0;">Llama-3-Instruct-Names but you will need to uncheck "System same as user".</p> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
<div class="section-container"> |
|
|
<div class="section-header"> |
|
|
<div class="section-indicator"></div> |
|
|
<h2 class="section-title">03 // QUANTIZATIONS</h2> |
|
|
</div> |
|
|
<div class="section-content"> |
|
|
<div style="margin-bottom: 20px;"> |
|
|
<h3 class="subheading">GGUF</h3> |
|
|
<div class="data-box"> |
|
|
<div class="data-row"> |
|
|
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/mradermacher/L3.3-GeneticLemonade-Final-v2-70B-i1-GGUF">iMatrix (mradermacher)</a> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
<div> |
|
|
<h3 class="subheading">EXL2</h3> |
|
|
<div class="data-box"> |
|
|
<div class="data-row"> |
|
|
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4bpw-hb6-exl2">4bpw</a> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4.5bpw-hb6-exl2">4.5bpw</a> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4.65bpw-hb6-exl2">4.65bpw</a> |
|
|
</div> |
|
|
<div class="data-row"> |
|
|
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_6bpw-hb8-exl2">6bpw</a> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
<div class="section-container"> |
|
|
<div class="section-header"> |
|
|
<div class="section-indicator"></div> |
|
|
<h2 class="section-title">04 // TRAINING PROCESS</h2> |
|
|
</div> |
|
|
<div class="section-content"> |
|
|
<p>This model was trained using a dataset of approx 4.3 million tokens, 700 RP conversations, 2000 creative writing / instruct samples and about 400 summaries. The bulk of this data has been made public.</p> |
|
|
<p>This model didn't take well to my existing DPO dataset, so it hasn't been used here.</p> |
|
|
</div> |
|
|
</div> |
|
|
</div> |
|
|
<h3 class="subheading">Axolotl configs</h3> |
|
|
<p>Not optimized for cost / performance efficiency, YMMV.</p> |
|
|
<h3>SFT 1*H200</h3> |
|
|
|
|
|
```yml |
|
|
# ==================== |
|
|
# MODEL CONFIGURATION |
|
|
# ==================== |
|
|
base_model: zerofata/L3.3-GeneticLemonade-Unleashed-70B |
|
|
model_type: AutoModelForCausalLM |
|
|
tokenizer_type: AutoTokenizer |
|
|
special_tokens: |
|
|
pad_token: "<|finetune_right_pad_id|>" |
|
|
chat_template: llama3 |
|
|
|
|
|
# ==================== |
|
|
# DATASET CONFIGURATION |
|
|
# ==================== |
|
|
datasets: |
|
|
- path: ./dataset.jsonl |
|
|
type: chat_template |
|
|
split: train |
|
|
chat_template_strategy: tokenizer |
|
|
field_messages: messages |
|
|
message_property_mappings: |
|
|
role: role |
|
|
content: content |
|
|
roles: |
|
|
user: ["user"] |
|
|
assistant: ["assistant"] |
|
|
system: ["system"] |
|
|
|
|
|
test_datasets: |
|
|
- path: ./validate_dataset.jsonl |
|
|
type: chat_template |
|
|
split: train |
|
|
chat_template_strategy: tokenizer |
|
|
field_messages: messages |
|
|
message_property_mappings: |
|
|
role: role |
|
|
content: content |
|
|
roles: |
|
|
user: ["user"] |
|
|
assistant: ["assistant"] |
|
|
system: ["system"] |
|
|
|
|
|
dataset_prepared_path: |
|
|
train_on_inputs: false # Only train on assistant responses |
|
|
|
|
|
# ==================== |
|
|
# QLORA CONFIGURATION |
|
|
# ==================== |
|
|
adapter: qlora |
|
|
load_in_4bit: true |
|
|
lora_r: 64 |
|
|
lora_alpha: 128 |
|
|
lora_dropout: 0.1 |
|
|
lora_target_linear: true |
|
|
# lora_modules_to_save: # Uncomment only if you added NEW tokens |
|
|
|
|
|
# ==================== |
|
|
# TRAINING PARAMETERS |
|
|
# ==================== |
|
|
num_epochs: 2 |
|
|
micro_batch_size: 4 |
|
|
gradient_accumulation_steps: 2 |
|
|
learning_rate: 1.5e-5 |
|
|
optimizer: paged_adamw_8bit |
|
|
lr_scheduler: rex |
|
|
warmup_ratio: 0.05 |
|
|
weight_decay: 0.01 |
|
|
max_grad_norm: 1.0 |
|
|
|
|
|
# ==================== |
|
|
# SEQUENCE & PACKING |
|
|
# ==================== |
|
|
sequence_len: 8192 |
|
|
sample_packing: true |
|
|
eval_sample_packing: false |
|
|
pad_to_sequence_len: true |
|
|
|
|
|
# ==================== |
|
|
# HARDWARE OPTIMIZATIONS |
|
|
# ==================== |
|
|
bf16: auto |
|
|
flash_attention: true |
|
|
gradient_checkpointing: true |
|
|
|
|
|
# ==================== |
|
|
# EVALUATION & CHECKPOINTING |
|
|
# ==================== |
|
|
evaluation_strategy: steps |
|
|
eval_steps: 5 |
|
|
save_strategy: steps |
|
|
save_steps: 5 |
|
|
save_total_limit: 5 # Keep best + last few checkpoints |
|
|
load_best_model_at_end: true |
|
|
metric_for_best_model: eval_loss |
|
|
greater_is_better: false |
|
|
early_stopping_patience: 5 |
|
|
|
|
|
# ==================== |
|
|
# LOGGING & OUTPUT |
|
|
# ==================== |
|
|
output_dir: ./output_model |
|
|
logging_steps: 2 |
|
|
save_safetensors: true |
|
|
|
|
|
# ==================== |
|
|
# WANDB TRACKING |
|
|
# ==================== |
|
|
wandb_project: project_name |
|
|
# wandb_entity: your_entity |
|
|
# wandb_name: your_run_name |
|
|
|
|
|
``` |
|
|
</body> |
|
|
</html> |