This is a decensored version of joelewing/Llama-3.2-1B-Instruct-Capybara, made using Heretic v1.0.1

Abliteration parameters

Parameter Value
direction_index 12.73
attn.o_proj.max_weight 1.49
attn.o_proj.max_weight_position 9.36
attn.o_proj.min_weight 0.65
attn.o_proj.min_weight_distance 4.53
mlp.down_proj.max_weight 1.14
mlp.down_proj.max_weight_position 12.76
mlp.down_proj.min_weight 1.02
mlp.down_proj.min_weight_distance 4.68

Performance

Metric This model Original model (joelewing/Llama-3.2-1B-Instruct-Capybara)
KL divergence 0.03 0 (by definition)
Refusals 3/100 77/100

Llama-3.2-1B-Instruct-Capybara

This model is a finetune of Llama 3.2 1B on the Capybara dataset.

  • Developed by: joelewing
  • License: Llama 3.2 Community License Agreement
  • Finetuned from model : unsloth/llama-3.2-1b-instruct-bnb-4bit

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
44
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for joelewing/Llama-3.2-1B-Instruct-Capybara-abliterated

Merges
1 model

Dataset used to train joelewing/Llama-3.2-1B-Instruct-Capybara-abliterated