--- base_model: intervitens/mini-magnum-12b-v1.1 pipeline_tag: text-generation datasets: - jondurbin/gutenberg-dpo-v0.1 - nbeerbower/gutenberg2-dpo - nbeerbower/gutenberg-moderne-dpo - sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo - nbeerbower/human-writing-dpo license: apache-2.0 language: - en --- # Model Card for Model ID A DPO qLORA finetune of Mistral Nemo 12b on four Gutenberg datasets plus one more dataset, approx ~9k lines. ## Model Details ### Model Description Finetuned for 1 epoch on an A100 through Vast.AI. ## Credits Thank you to Axolotl for making finetuning easier. Thank you to Docker for... existing, I guess. ## Changes from v1 - Base model changed to `intervitens/mini-magnum-12b-v1.1` - Added nbeerbower/human-writing, which was supposed to be in v1 but I forgot to add it. - Adjusted learning rate/other settings to compensate. ## Addendum You know, I am REALLY regretting panic-naming this line of models so ambiguously now. Well, too late now!