Text Generation
Transformers
Safetensors
llama
conversational
text-generation-inference
zerofata's picture
Update README.md
2475821 verified
---
library_name: transformers
license: llama3
datasets:
- zerofata/Roleplay-Anime-Characters
- zerofata/Instruct-Anime-CreativeWriting
- zerofata/Summaries-Anime-FandomPages
base_model:
- zerofata/L3.3-GeneticLemonade-Final-70B
---
<!DOCTYPE html>
<style>
body {
font-family: sans-serif;
color: #f0f0f0;
line-height: 1.6;
margin: 0;
padding: 0;
background-color: #1a0f1a;
}
.lemonade-text {
color: #ff3366;
position: relative;
z-index: 2;
margin-left: 0.2em;
text-shadow: 0 0 10px #ff3366;
}
/* Section styling */
.section-container {
background-color: rgba(26, 15, 26, 0.7);
margin-bottom: 30px;
position: relative;
overflow: hidden;
border-bottom: 1px solid #ff3366;
}
.section-header {
display: flex;
align-items: center;
background-color: rgba(255, 51, 102, 0.08);
padding: 10px 20px;
}
.section-indicator {
width: 8px;
height: 20px;
background-color: #ff3366;
margin-right: 15px;
}
.section-title {
font-family: 'Orbitron', sans-serif;
color: #f0f0f0;
font-size: 1.3rem;
margin: 0;
letter-spacing: 2px;
text-transform: uppercase;
font-weight: 500;
}
.section-content {
padding: 20px;
font-family: sans-serif;
color: #f0f0f0;
line-height: 1.6;
}
/* Title styling */
.title-container {
background-color: #0a0a0a;
position: relative;
overflow: hidden;
margin-bottom: 40px;
border-left: 3px solid #ff3366;
}
.title-wrapper {
position: relative;
z-index: 2;
padding: 25px 20px 30px 30px;
font-family: 'Orbitron', sans-serif;
}
.title-main {
color: #f0f0f0;
font-size: 2.5rem;
font-weight: 700;
margin: 0;
letter-spacing: 2px;
display: inline-block;
position: relative;
text-transform: uppercase;
}
.title-prefix {
position: relative;
z-index: 2;
}
.title-subtitle {
padding-left: 15px;
margin-top: 5px;
margin-left: 5px;
}
.subtitle-text {
color: #cc0066;
font-size: 1.2rem;
font-family: 'Orbitron', sans-serif;
font-weight: 300;
letter-spacing: 3px;
text-transform: uppercase;
display: inline-block;
}
.glitchy-overlay {
position: absolute;
top: 0;
left: 0;
width: 100%;
height: 100%;
background-image: repeating-linear-gradient(0deg, rgba(0,0,0,0) 0, rgba(139, 0, 0, 0.1) 1px, rgba(0,0,0,0) 2px);
z-index: 1;
}
/* Data box styling */
.data-box {
background-color: rgba(0, 0, 0, 0.4);
padding: 15px;
border-left: 2px solid #ff3366;
margin-bottom: 20px;
}
.data-row {
display: flex;
margin-bottom: 8px;
}
.data-arrow {
color: #ff3366;
width: 20px;
display: inline-block;
}
.data-label {
color: #cc0066;
width: 80px;
display: inline-block;
}
/* Subheading styling */
.subheading {
color: #cc0066;
font-size: 1.1rem;
margin-top: 20px;
margin-bottom: 15px;
font-weight: 400;
border-bottom: 1px dashed rgba(204, 0, 102, 0.4);
display: inline-block;
text-transform: uppercase;
letter-spacing: 1px;
font-family: 'Orbitron', sans-serif;
}
/* Links */
a {
color: #cc0066;
text-decoration: none;
}
a:hover {
text-decoration: underline;
color: #ff6600;
}
/* Container */
.container {
max-width: 1200px;
margin: 20px auto;
padding: 40px 20px;
background-color: #0a0a0a;
background-image:
linear-gradient(rgba(139, 0, 0, 0.12) 1px, transparent 1px),
linear-gradient(90deg, rgba(139, 0, 0, 0.12) 1px, transparent 1px);
background-size: 20px 20px;
min-height: calc(100vh - 40px);
border: 1px solid #ff3366;
border-radius: 2px;
}
</style>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>GENETIC LEMONADE FINAL v2</title>
<link href="https://fonts.googleapis.com/css2?family=Orbitron:wght@400;500;600;700&family=JetBrains+Mono:wght@100;300;400;700&display=swap" rel="stylesheet">
</head>
<body>
<div class="cyber-grid-bg"></div>
<div class="container">
<div class="title-container">
<!-- Glitchy overlay -->
<div class="glitchy-overlay"></div>
<!-- Main title -->
<div class="title-wrapper">
<h1 class="title-main">
<span class="title-prefix">GENETIC</span>
<span class="lemonade-text">LEMONADE</span> <!-- Static text with glow -->
</h1>
<div class="title-subtitle">
<span class="subtitle-text">FINAL v2</span>
</div>
</div>
</div>
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c6c638328850e12d38c/0Ka08CdFUIJtYctBeBATo.png)
<div class="section-container">
<div class="section-header">
<div class="section-indicator"></div>
<h2 class="section-title">01 // OVERVIEW</h2>
</div>
<div class="section-content">
<p>Wasn't intending to release another model (so soon at least), but I was testing out some new dataset ideas and thought this model came out pretty nice.</p>
<p><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-70B">zerofata/GeneticLemonade-Final</a> SFT QLora finetune.</p>
<p>This is an uncensored creative model intended to excel at character driven RP / ERP.</p>
<p>This model is designed to provide longer, narrative heavy responses where characters are portrayed accurately and proactively.</p>
<p>Compared to Unleashed v3, this model has significantly reduced positivity bias and arguably a nicer writing style. The tradeoff is it swipe heavy, making a few more logical errors and can be a bit too concise at times.</p>
</div>
</div>
<div class="section-container">
<div class="section-header">
<div class="section-indicator"></div>
<h2 class="section-title">02 // SILLYTAVERN SETTINGS</h2>
</div>
<div class="section-content">
<p>Play with these, they are not the 'best' settings just a stable baseline.</p>
<h3 class="subheading">Recommended Samplers</h3>
<div class="data-box">
<div class="data-row">
<span class="data-arrow">></span>
<span class="data-label">Temp:</span>
<span>0.9 - 1</span>
</div>
<div class="data-row">
<span class="data-arrow">></span>
<span class="data-label">MinP:</span>
<span>0.03 - 0.04</span>
</div>
<div class="data-row">
<span class="data-arrow">></span>
<span class="data-label">TopP:</span>
<span>0.9 - 1.0</span>
</div>
<div class="data-row">
<span class="data-arrow">></span>
<span class="data-label">Dry:</span>
<span>0.8, 1.75, 4</span>
</div>
</div>
<h3 class="subheading">Instruct</h3>
<div class="data-box">
<p style="margin: 0;">Llama-3-Instruct-Names but you will need to uncheck "System same as user".</p>
</div>
</div>
</div>
<div class="section-container">
<div class="section-header">
<div class="section-indicator"></div>
<h2 class="section-title">03 // QUANTIZATIONS</h2>
</div>
<div class="section-content">
<div style="margin-bottom: 20px;">
<h3 class="subheading">GGUF</h3>
<div class="data-box">
<div class="data-row">
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/mradermacher/L3.3-GeneticLemonade-Final-v2-70B-i1-GGUF">iMatrix (mradermacher)</a>
</div>
</div>
</div>
<div>
<h3 class="subheading">EXL2</h3>
<div class="data-box">
<div class="data-row">
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4bpw-hb6-exl2">4bpw</a>
</div>
<div class="data-row">
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4.5bpw-hb6-exl2">4.5bpw</a>
</div>
<div class="data-row">
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_4.65bpw-hb6-exl2">4.65bpw</a>
</div>
<div class="data-row">
<span style="color: #ff3366; display: inline-block; margin-right: 10px;">> </span><a href="https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-v2-70B_6bpw-hb8-exl2">6bpw</a>
</div>
</div>
</div>
</div>
</div>
<div class="section-container">
<div class="section-header">
<div class="section-indicator"></div>
<h2 class="section-title">04 // TRAINING PROCESS</h2>
</div>
<div class="section-content">
<p>This model was trained using a dataset of approx 4.3 million tokens, 700 RP conversations, 2000 creative writing / instruct samples and about 400 summaries. The bulk of this data has been made public.</p>
<p>This model didn't take well to my existing DPO dataset, so it hasn't been used here.</p>
</div>
</div>
</div>
<h3 class="subheading">Axolotl configs</h3>
<p>Not optimized for cost / performance efficiency, YMMV.</p>
<h3>SFT 1*H200</h3>
```yml
# ====================
# MODEL CONFIGURATION
# ====================
base_model: zerofata/L3.3-GeneticLemonade-Unleashed-70B
model_type: AutoModelForCausalLM
tokenizer_type: AutoTokenizer
special_tokens:
pad_token: "<|finetune_right_pad_id|>"
chat_template: llama3
# ====================
# DATASET CONFIGURATION
# ====================
datasets:
- path: ./dataset.jsonl
type: chat_template
split: train
chat_template_strategy: tokenizer
field_messages: messages
message_property_mappings:
role: role
content: content
roles:
user: ["user"]
assistant: ["assistant"]
system: ["system"]
test_datasets:
- path: ./validate_dataset.jsonl
type: chat_template
split: train
chat_template_strategy: tokenizer
field_messages: messages
message_property_mappings:
role: role
content: content
roles:
user: ["user"]
assistant: ["assistant"]
system: ["system"]
dataset_prepared_path:
train_on_inputs: false # Only train on assistant responses
# ====================
# QLORA CONFIGURATION
# ====================
adapter: qlora
load_in_4bit: true
lora_r: 64
lora_alpha: 128
lora_dropout: 0.1
lora_target_linear: true
# lora_modules_to_save: # Uncomment only if you added NEW tokens
# ====================
# TRAINING PARAMETERS
# ====================
num_epochs: 2
micro_batch_size: 4
gradient_accumulation_steps: 2
learning_rate: 1.5e-5
optimizer: paged_adamw_8bit
lr_scheduler: rex
warmup_ratio: 0.05
weight_decay: 0.01
max_grad_norm: 1.0
# ====================
# SEQUENCE & PACKING
# ====================
sequence_len: 8192
sample_packing: true
eval_sample_packing: false
pad_to_sequence_len: true
# ====================
# HARDWARE OPTIMIZATIONS
# ====================
bf16: auto
flash_attention: true
gradient_checkpointing: true
# ====================
# EVALUATION & CHECKPOINTING
# ====================
evaluation_strategy: steps
eval_steps: 5
save_strategy: steps
save_steps: 5
save_total_limit: 5 # Keep best + last few checkpoints
load_best_model_at_end: true
metric_for_best_model: eval_loss
greater_is_better: false
early_stopping_patience: 5
# ====================
# LOGGING & OUTPUT
# ====================
output_dir: ./output_model
logging_steps: 2
save_safetensors: true
# ====================
# WANDB TRACKING
# ====================
wandb_project: project_name
# wandb_entity: your_entity
# wandb_name: your_run_name
```
</body>
</html>