πŸ‘‹ Connect with me on LinkedIn!

LinkedIn

Manuel Caccone - Actuarial Data Scientist & Open Source Educator

Let's discuss actuarial science, AI, and open source projects!


Model Thumbnail

πŸ€– Gemma-3 ActuaryEnough2: Bringing Actuarial AI to Everyone


🚩 Model Description

Gemma-3-actuaryEnough2 is a fine-tuned Gemma-3-270M model trained on over 11,000 actuarial question-answer pairs, purpose-built to translate simple insurance queries into rigorous actuarial technical language. It powers ActuaryEnough and is released as open source for educational and research use.


✨ Key Features

  • 🎯 Domain-specific: Focused exclusively on actuarial and insurance Q&A.
  • πŸ“š Educational: Makes complex actuarial terminology accessible for all users.
  • πŸš€ Efficient: Fine-tuned with Unsloth for rapid, scalable training.
  • πŸ”“ Open Source: Apache 2.0 License; easy to reuse, adapt, remix.
  • 🌐 Widget & Demo: Integrated as a live demo on ActuaryEnough.

πŸ’‘ Intended Use Cases

  • Education: For students and actuaries in training, or for professionals retraining in actuarial language.
  • Translation: Make practical insurance questions understandable at professional actuarial level.
  • Research: Support for actuarial research, Q&A, and domain adaptation.

Examples

# Premium Calculation Example
Input: "How much should I pay for car insurance? Rephrase:"
Output: "This relates to premium calculation considering risk factors such as exposure units, loss frequency, severity distributions, and loading factors for expenses and profit margins."

πŸ“‚ Training Data

  • Primary Dataset: actuarial-qa-11k - Over 11,000 manually curated actuarial question–answer pairs
  • Specialized Dataset: actuary-enough-qa-dataset - Actuarial question simplification examples
  • Topics: Life and non-life insurance, risk assessment, regulation, reserves, actuarial mathematics, terminology simplification
  • Language: English
  • Format: Instruction-following format optimized for text generation tasks

πŸ“Š Training Statistics

Metric Value / Range Notes
Epochs ~51 Reached at end of training
Global Steps >68,000
Initial Train Loss ~2.2 At start
Final Train Loss ~1.4 At end
Learning Rate 8e-7 β†’ β‰ˆ0 Linear decay throughout training
Gradient Norm 5 – 15 Generally stable with rare spikes
Hardware RTX 3090, 16-core CPU 24GB VRAM, 94GB RAM, CUDA 12.8, Linux 6.1

πŸ› οΈ Dependencies

Python 3.12.11
transformers
torch
unsloth
wandb==0.21.1
pydantic==2.11.7
# ...for full list, check requirements.txt

⚠️ Limitations & Ethics

  • No pricing or decision support: For education and inspiration only, not for real insurance contracts.
  • Not a substitute for an actuary: Always consult professionals for real-world decisions.
  • Coverage: Designed and tested specifically for the insurance/actuarial domain.
  • Training data bias: Outputs may reflect source content.

πŸ’» Usage Example

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("manuelcaccone/gemma-3-actuaryEnough2")
model = AutoModelForCausalLM.from_pretrained("manuelcaccone/gemma-3-actuaryEnough2")

prompt = "Which factors determine life insurance premiums?"
toks = tokenizer(prompt, return_tensors="pt")
output = model.generate(**toks, max_new_tokens=256, temperature=0.7)
print(tokenizer.decode(output[0], skip_special_tokens=True))

🌟 Related Datasets

This model is part of the ActuaryEnough ecosystem and uses multiple specialized datasets:


πŸ‘€ Author & Citation

@model{caccone2025actuaryenough,
  title={Gemma-3 ActuaryEnough2: A Fine-tuned Model for Actuarial Education},
  author={Caccone, Manuel},
  year={2025},
  publisher={Hugging Face},
  url={https://huggingface.co/manuelcaccone/gemma-3-actuaryEnough2},
  note={Educational model for actuarial science and insurance terminology}
}

πŸ“œ License

Apache 2.0 License β€” use, modify, and cite for ethical, research, and educational purposes.


🀝 Want to collaborate or discuss actuarial AI?

LinkedIn


Part of the ActuaryEnough open-source education initiativeβ€”bringing actuarial science closer to everyone!

Downloads last month
25
Safetensors
Model size
0.3B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for manuelcaccone/gemma-3-actuaryEnough2

Finetuned
(309)
this model

Dataset used to train manuelcaccone/gemma-3-actuaryEnough2