Model Card for traclm-v3-7b-instruct
An Army domain finetune of mistralai/Mistral-7B-v0.1 created by finetuning on a merger of domain-specific and general-purpose instruction-tuning datasets.
Model Details
Model Description
This model is a research project aimed at exploring whether a pretrained LLM can acquire tangible domain-specific knowledge about the Army domain.
- Developed by: The Research and Analysis Center, Army Futures Command, U.S. Army
- License: MIT
- Model Type: MistralForCausalLM
- Finetuned from model: mistralai/Mistral-7B-v0.1
Available Quantizations (for running on low-resource hardware):
Downstream Use
This model is instruction-tuned, and is thus more capable of following user instructions than its corresponding base version. However, this model is still capable of extreme hallucination, so all outputs should be verified by end users.
Out-of-Scope Use
The creation of this model constitutes academic research in partnership with the Naval Postgraduate School. The purpose of this research is to inform future DoD experimentation regarding the development and application of domain-specific large language models. Experiments involving direct application of this model to downstream military tasks are encouraged, but extreme caution should be exercised before productionalization.
Prompt Format
This model was fine-tuned with the chatml prompt format. It is highly recommended that you use the same format for any interactions with the model. Failure to do so will degrade performance significantly.
ChatML Format:
<|im_start|>system
Provide some context and/or instructions to the model.
<|im_end|>
<|im_start|>user
The user’s message goes here
<|im_end|>
<|im_start|>assistant
The ChatML format can easily be applied to text you plan to process with the model using the chat_template
included in the tokenizer. Read here for additional information.
Training Details
Training Data
This model was trained on a shuffled merger of the following datasets:
- General Purpose Instruction Tuning: Open-Orca/SlimOrca-Dedup
- Domain Specific Instruction Tuning: TRAC-MTRY/traclm-v3-data (TBP)
Training Procedure
The model was trained using Open Access AI Collective's Axolotl framework and Microsoft's DeepSpeed framework for model/data parallelism.
Training Hardware
Training was conducted on a single compute node 4x NVIDIA A100 GPUs.
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-06
- train_batch_size: 4
- eval_batch_size: 4
- seed: 28
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 16
- total_eval_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 19
- num_epochs: 3
Training results
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.5726 | 0.0 | 1 | 1.6102 |
1.2333 | 0.2 | 510 | 1.2477 |
1.1485 | 0.4 | 1020 | 1.2010 |
1.106 | 0.6 | 1530 | 1.1687 |
1.1772 | 0.8 | 2040 | 1.1419 |
1.1567 | 1.0 | 2550 | 1.1190 |
1.0359 | 1.19 | 3060 | 1.1130 |
0.945 | 1.39 | 3570 | 1.0977 |
0.9365 | 1.59 | 4080 | 1.0831 |
0.9334 | 1.79 | 4590 | 1.0721 |
0.8913 | 1.99 | 5100 | 1.0627 |
0.804 | 2.18 | 5610 | 1.0922 |
0.7892 | 2.38 | 6120 | 1.0888 |
0.7757 | 2.58 | 6630 | 1.0873 |
0.7797 | 2.78 | 7140 | 1.0864 |
Framework versions
- Transformers 4.38.2
- Pytorch 2.1.2+cu121
- Datasets 2.18.0
- Tokenizers 0.15.0
Model Card Contact
MAJ Daniel C. Ruiz ([email protected])
- Downloads last month
- 2