Text Generation
PEFT
text-generation-inference
🇪🇺 Region: EU
mrs83's picture
Upload 2 files
e604d0a verified
{
"adapter_path": "adapters",
"batch_size": 2,
"config": "granite_config.yaml",
"data": "mlx-community/WikiSQL",
"fine_tune_type": "dora",
"grad_accumulation_steps": 1,
"grad_checkpoint": true,
"hf_dataset": [
{
"path": "ethicalabs/Kurtis-E1-SFT",
"train_split": "train[:40000]",
"valid_split": "validation[:16]",
"prompt_feature": "question",
"completion_feature": "answer",
"mask_prompt": false
},
{
"path": "ethicalabs/kurtis-v2-sft-mix-tiny",
"train_split": "train[:10000]",
"valid_split": "validation[:16]",
"prompt_feature": "prompt",
"completion_feature": "completion",
"mask_prompt": false
}
],
"iters": 1200,
"learning_rate": 5e-05,
"lora_parameters": {
"rank": 16,
"alpha": 64,
"dropout": 0.1,
"scale": 8.0
},
"lr_schedule": {
"name": "cosine_decay",
"warmup": 120,
"warmup_init": 0,
"arguments": [
5e-05,
1200,
5e-06
]
},
"mask_prompt": false,
"max_seq_length": 2048,
"merge_adapter": true,
"model": "ethicalabs/granite-4.0-1b-MLX",
"num_layers": 3,
"optimizer": "adamw",
"optimizer_config": {
"adamw": {
"betas": [
0.9,
0.98
],
"eps": 1e-07,
"weight_decay": 0.05,
"bias_correction": true
}
},
"output_path": "ethicalabs/kurtis-e1.2-granite-4.0-1b-MLX",
"project_name": null,
"report_to": "wandb",
"resume_adapter_file": null,
"save_every": 300,
"seed": 42,
"steps_per_eval": 200,
"steps_per_report": 10,
"test": false,
"test_batches": 100,
"train": true,
"val_batches": 50,
"wandb": "kurtis-mlx-granite-1b"
}