jarvisloh's picture
Update README.md
73d8976 verified
---
datasets:
- osunlp/SMolInstruct
language:
- en
base_model:
- mistralai/Mistral-7B-Instruct-v0.3
tags:
- chemistry
library_name: mlx
---
Finetuned on SmolInstruct's property prediction instruction dataset and HoneyBee's instruction dataset.
[LoRA Config Parameters]
train: true,
fine_tune_type: lora,
seed: 0,
num_layers: 8,
batch_size: 2,
iters: 1000,
val_batches: 25,
learning_rate: 1e-5,
steps_per_report: 10,
steps_per_eval: 200,
resume_adapter_file: null,
adapter_path: "adapters",
save_every: 100,
test: false,
test_batches: 100,
max_seq_length: 2048,
grad_checkpoint: false,
lora_parameters:
keys: ["self_attn.q_proj", "self_attn.v_proj"]
rank: 8
alpha: 8
dropout: 0.0
scale: 20.0