LLM Workshop Hands-on GPT Model

本モデルはチュートリアル「Let's develop LLMs from scratch」から学習した小型GPTモデルです。

特徴

  • vocab_size: 50257
  • emb_dim: 128
  • n_heads: 8
  • n_layers: 4
  • context_length: 128

使い方

from transformers import AutoModelForCausalLM, AutoTokenizer
tok = AutoTokenizer.from_pretrained("gpt2")
model = AutoModelForCausalLM.from_pretrained(
    "itarutomy/llm_workshop_hands_on_gpt-model",
    trust_remote_code=True
)
print(tok.decode(model.generate(**tok("こんにちは", return_tensors="pt"))[0]))
Downloads last month
23
Safetensors
Model size
13.7M params
Tensor type
F32
·
BOOL
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train itarutomy/llm_workshop_hands_on_gpt-model