{ "adapter_path": "adapters", "batch_size": 4, "config": "config.yaml", "data": "data", "fine_tune_type": "lora", "grad_checkpoint": false, "iters": 2079, "learning_rate": 1e-06, "lora_layers": 16, "lora_parameters": { "keys": [ "mlp.gate_proj", "mlp.down_proj", "self_attn.q_proj", "mlp.up_proj", "self_attn.o_proj", "self_attn.v_proj", "self_attn.k_proj" ], "rank": 128, "alpha": 256, "scale": 10.0, "dropout": 0.05 }, "lr_schedule": null, "max_seq_length": 2549, "model": "mlx-community/Llama-3.2-1B-Instruct-4bit", "num_layers": 16, "resume_adapter_file": null, "save_every": 400, "seed": 0, "steps_per_eval": 200, "steps_per_report": 10, "test": false, "test_batches": -1, "train": true, "val_batches": -1 }