🧠 Llama-3.1-Tulu-3-70B-Ko-Reasoning

A large-scale Korean reasoning model fine-tuned from allenai/Llama-3.1-Tulu-3-70B, designed to excel in logical and multi-hop reasoning tasks in Korean.


πŸ“Œ Overview

Llama-3.1-Tulu-3-70B-Ko-Reasoning is a fine-tuned version of allenai/Llama-3.1-Tulu-3-70B, specifically optimized for logical reasoning in Korean. This model is part of a broader research initiative to explore:

  • The transition from multilingual reasoning LLMs to Korean-specialized reasoning models
  • The enhancement of non-reasoning Korean language models into reasoning-capable variants
  • The development of open-access models that rival proprietary alternatives in complex reasoning tasks

This model was fine-tuned using a large-scale Korean-English instruction dataset containing diverse multi-hop questions, symbolic logic tasks, and human-crafted reasoning steps.


πŸ§‘β€πŸ’» Usage

Install Transformers >= 4.50:

pip install -U transformers

Basic example:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "DimensionSTP/Llama-3.1-Tulu-3-70B-Ko-Reasoning"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "μ„œμšΈκ³Ό λΆ€μ‚° 쀑 μ–΄λ””κ°€ 더 컀?"
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)

model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=32768
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)

🧠 Base Model: allenai/Llama-3.1-Tulu-3-70B

The base model, allenai/Llama-3.1-Tulu-3-70B, is a LLM developed by the Allen AI, fine tuned from Llama 3.1 base. For more technical details, refer to the Tulu Technical Report.


🧱 Model Architecture

Property Value
Architecture LlamaForCausalLM
Parameters 70B
Context Length 131,072 tokens
Tokenizer LLamaTokenizer (BPE)

πŸ“… Release Date

Mar 2025
This model was released in March 2025 as part of the Ko-Reasoning Series, which focuses on pushing the boundaries of open-source reasoning in Korean using modern LLMs.


πŸ“¬ Contact

For questions, collaborations, or deployment inquiries, please contact:


πŸ“¦ Available Checkpoints

  • βœ… main: Final stable version from the last branch
  • βœ… All training artifacts available (tokenizer, config, model weights)
Downloads last month
1
Safetensors
Model size
71B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for DimensionSTP/Llama-3.1-Tulu-3-70B-Ko-Reasoning

Quantizations
2 models

Collection including DimensionSTP/Llama-3.1-Tulu-3-70B-Ko-Reasoning