--- base_model: - ddh0/Cassiopeia-70B - Sao10K/Llama-3.3-70B-Vulpecula-r1 base_model_relation: merge license: unknown thumbnail: https://huggingface.co/ddh0/Andromeda-70B/resolve/main/andromeda.png --- # Andromeda-70B ![Andromeda-70B](andromeda.png) **Andromeda-70B** is the result of an experimental SLERP merge of [Cassiopeia-70B](https://huggingface.co/ddh0/Cassiopeia-70B) and [Sao10K/Llama-3.3-70B-Vulpecula-r1](https://huggingface.co/Sao10K/Llama-3.3-70B-Vulpecula-r1). It is a coherent, unaligned model intended to be used for creative tasks such as storywriting, brainstorming, interactive roleplay, etc. ## UPDATE After more thorough testing by myself and others, I don't think this model is very good. :( You should use Cassiopeia or Vulpecula instead. ## Merge composition ```yaml models:   - model: /opt/workspace/hf/Cassiopeia-70B   - model: /opt/workspace/hf/Llama-3.3-70B-Vulpecula-r1 merge_method: slerp base_model: /opt/workspace/hf/Cassiopeia-70B parameters:   t: 0.7 dtype: bfloat16 ``` ## Feedback If you like this model, please support [Sao10k](https://sao10k.carrd.co). Feedback on this merge is very welcome, good or bad! Please leave a comment in this discussion with your thoughts: [Andromeda-70B/discussions/1](https://huggingface.co/ddh0/Andromeda-70B/discussions/1)