Model Card for yevvonlim/Llada-8B-Instruct-Kor

yevvonlim/Llada-8B-Instruct-Kor is an instruction-tuned variant of LLADA-8B designed for high-quality conversational responses in both Korean and English. Fine-tuned with supervised data, it excels at understanding and generating context-aware replies for chat applications.

Model Details

Model Description

This model is a supervised fine-tuned (SFT) version of [GSAI-ML/LLaDA-8B-Instruct], developed and shared by Sionic AI. It leverages parameter-efficient fine-tuning (PEFT) with LoRA to adapt the base LLADA-8B model to instruction-following tasks.

  • Developed by: yevvonlim
  • Model type: 8B-parameter encoder-only transformer
  • Language(s): Korean, English
  • License: Apache-2.0
  • Fine-tuned from: GSAI-ML/LLaDA-8B-Instruct

Model Sources

Uses

Direct Use

  • Conversational agents and chatbots in Korean and English
  • Instruction-following and question-answering tasks
  • Assistive tools for writing, translation, and summarization

Out-of-Scope Use

  • Tasks requiring specialized domain knowledge outside the training data
  • Real-time high-stakes decision-making without human oversight

Bias, Risks, and Limitations

  • May produce incorrect or outdated facts
  • Vulnerable to generating biased or stereotypical language present in training data

Recommendations

Users should review and verify model outputs before deployment in critical applications. Implement human-in-the-loop validation for high-stakes use cases.

How to Get Started with the Model

Use the code below to load and generate with the model. Ensure you have defined or imported the generate_stream function provided in the repository.


from transformers import AutoTokenizer, AutoModel


device = "cuda"
model_path = "yevvonlim/Llada-8B-Instruct-Kor"

tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model = AutoModel.from_pretrained(model_path, trust_remote_code=True).to(device).eval()

prompt = "6๋‚˜๋ˆ„๊ธฐ 0์€ ๋ญ์•ผ? let's think step by step."
chat_input = tokenizer.apply_chat_template(
    [{"role": "user", "content": prompt}],
    add_generation_prompt=True,
    tokenize=False,
)
prompt_ids = tokenizer(chat_input, return_tensors="pt").input_ids.to(device)

final_ids = model.generate(prompt_ids)[0, prompt_ids.shape[1]:]
print(tokenizer.decode(final_ids, skip_special_tokens=True))

Contact

For issues or questions, please open an issue on the repo or contact [email protected]

Downloads last month
2
Safetensors
Model size
8B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for yevvonlim/Llada-8B-Instruct-Kor

Finetuned
(6)
this model
Quantizations
2 models