ShallowMind-abeat/blahblahthron-1.1b

NOTE : This version is is not fully trained, I stopped it way earlier because i needed something to experiment with ;)

A custom causal language model implemented as a nn.Module now wrapped into a PreTrainedModel—fully compatible with transformers.from_pretrained(trust_remote_code=True) usage.

Highlights:

Custom GPT-style LM with rotary attention, RMSNorm, SwiGLU, and tied embeddings.

Fully integrated: use from_pretrained(), save_pretrained(), and Transformers pipelines.

Includes QLoRA-compatibility—handles 4-bit finetuning via peft.

Load

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "ShallowMind-abeat/blahblahthron-1.1b",
    trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained("ShallowMind-abeat/blahblahthron-1.1b")

Then use immediately in pipelines or with Trainer:

from transformers import pipeline

pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
Downloads last month
4
Safetensors
Model size
1B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support