Fino1: On the Transferability of Reasoning Enhanced LLMs to Finance
Paper
β’
2502.08127
β’
Published
β’
58
This repository contains Q4_KM and Q5_KM quantized versions of TheFinAI/Fino1-8B, a financial reasoning model based on Llama 3.1 8B Instruct. These quantized variants maintain the model's financial reasoning capabilities while providing significant memory and speed improvements.
Discover our full range of quantized language models by visiting our SandLogic Lexicon HuggingFace. To learn more about our company and services, check out our website at SandLogic.
Both quantized versions maintain the original model's strengths in:
pip install llama-cpp-python
Please refer to the llama-cpp-python documentation to install with GPU support.
from llama_cpp import Llama
llm = Llama(
model_path="model/path/",
verbose=False,
# n_gpu_layers=-1, # Uncomment to use GPU acceleration
# n_ctx=2048, # Uncomment to increase the context window
)
# Example of a reasoning task
output = llm(
"""Q: A company's revenue grew from $100,000 to $150,000 in one year.
Calculate the percentage growth rate. A: """,
max_tokens=256,
stop=["Q:", "\n\n"],
echo=False
)
print(output["choices"][0]["text"])
4-bit
5-bit
Base model
meta-llama/Llama-3.1-8B