TinyLlama Instruct Lite v1
π Model Summary
rogersam/tinyllama-instruct-lite-v1 is a LoRA fine-tuned TinyLlama model using Unsloth.
It is designed for instruction-following tasks in English + Tamil, such as:
- General Q&A
- Summarization
- Basic math & reasoning
- English β Tamil translation
This project demonstrates how a lightweight 1B model can be adapted for multiple domains with limited resources.
π Model Details
- Developed by: Roger Samuel J (Hugging Face Profile)
- Model type: Causal LM (decoder-only)
- Languages: English, Tamil
- License: Same as base model (TinyLlama)
- Fine-tuned from:
unsloth/tinyllama-bnb-4bit - Method: LoRA via PEFT + Unsloth
π Model Sources
- Model Repo: rogersam/tinyllama-instruct-lite-v1
- Base Model: unsloth/tinyllama-bnb-4bit
π‘ Uses
Direct Use
- Running lightweight instruction tasks on CPU/GPU
- Translating English β Tamil sentences
- Answering short questions and reasoning queries
- Summarizing small texts
Out-of-Scope
- Sensitive decision-making (finance, healthcare, law)
- Long context generation (>512 tokens)
- Production-grade chatbots
β οΈ Bias, Risks & Limitations
- Small dataset β may hallucinate facts
- Not aligned for safety or toxicity filtering
- Limited Tamil coverage (basic sentences only)
Recommendation: Use for demo & educational purposes only.
π How to Get Started
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
model_id = "rogersam/tinyllama-instruct-lite-v1"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
prompt = "Translate English to Tamil: How are you?"
print(pipe(prompt, max_new_tokens=50)[0]["generated_text"])
- Downloads last month
- 10
Model tree for rogersam/tinyllama-instruct-lite-v1
Base model
unsloth/tinyllama-bnb-4bit