|
--- |
|
license: other |
|
library_name: transformers |
|
tags: |
|
- generated_from_trainer |
|
- medical |
|
- Healthcare & Lifesciences |
|
- BioMed |
|
- chain-of-thought |
|
base_model: meta-llama/Llama-3.2-1B-Instruct |
|
thumbnail: https://collaiborate.com/logo/logo-blue-bg-1.png |
|
model-index: |
|
- name: Bio-Medical-Llama-3-2-1B-CoT-012025 |
|
results: [] |
|
datasets: |
|
- collaiborateorg/BioMedData |
|
--- |
|
|
|
# Bio-Medical-Llama-3-2-1B-CoT-012025 |
|
|
|
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/653f5b93cd52f288490edc83/zPMUugzfOiwTiRw88jm7T.jpeg) |
|
|
|
This model is a fine-tuned version of [Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on our custom "BioMedData" dataset, enhanced with 625,000 examples, including 25,000 chain-of-thought (CoT) instruction samples to strengthen reasoning capabilities. It is specifically optimized for the Healthcare & Lifesciences (HLS) domain. |
|
|
|
## Model details |
|
|
|
**Model Name:** Bio-Medical-Llama-3-2-1B-CoT-012025 |
|
|
|
**Base Model:** Llama-3.2-1B-Instruct |
|
|
|
**Parameter Count:** 1 billion |
|
|
|
**Training Data:** Custom high-quality biomedical dataset with 625,000 examples, including 25,000 CoT instructions. |
|
|
|
**Number of Entries in Dataset:** 625,000 |
|
|
|
**Dataset Composition:** The dataset comprises a mix of synthetic, manually curated, and reasoning-focused entries, ensuring comprehensive coverage of biomedical knowledge and logical reasoning. |
|
|
|
## Model description |
|
|
|
The Bio-Medical-Llama-3-2-1B-CoT-012025 model is a lightweight yet powerful language model tailored for: |
|
|
|
- Generating domain-specific content in healthcare and biomedical fields. |
|
- Answering complex questions requiring step-by-step reasoning using CoT. |
|
- Supporting researchers, clinicians, and students in their respective biomedical endeavors. |
|
|
|
This model is fine-tuned to provide interpretability and improved logical coherence through its enhanced CoT capabilities. |
|
|
|
## Evaluation Metrics |
|
|
|
Bio-Medical-Llama-3-2-1B-CoT-012025 has been evaluated using the Eleuther AI Language Model Evaluation Harness framework on tasks including: |
|
|
|
- medmcqa |
|
- medqa_4options |
|
- mmlu_anatomy |
|
- mmlu_clinical_knowledge |
|
- mmlu_college_biology |
|
- mmlu_college_medicine |
|
- mmlu_medical_genetics |
|
- mmlu_professional_medicine |
|
- pubmedqa |
|
|
|
Results show consistent performance improvements over general-purpose models of similiar size, particularly in tasks requiring reasoning. |
|
|
|
## Intended uses & limitations |
|
|
|
**Intended Uses:** |
|
1. **Research Support:** Assisting researchers with reasoning and data extraction from biomedical texts. |
|
2. **Clinical Decision Support:** Offering logical and evidence-based information to aid decision-making. |
|
3. **Educational Tool:** Serving as a learning resource for understanding complex biomedical concepts. |
|
|
|
**Limitations and Ethical Considerations:** |
|
|
|
- **Biases:** The model may reflect biases from the training data, despite efforts to mitigate them. |
|
- **Accuracy:** Responses should be cross-verified with reliable sources in critical scenarios. |
|
- **Ethical Use:** The model should augment professional expertise and not replace it, especially in high-stakes applications. |
|
|
|
## How to use |
|
|
|
```python |
|
import transformers |
|
import torch |
|
|
|
model_id = "ContactDoctor/Bio-Medical-Llama-3-2-1B-CoT-012025" |
|
|
|
pipeline = transformers.pipeline( |
|
"text-generation", |
|
model=model_id, |
|
model_kwargs={"torch_dtype": torch.bfloat16}, |
|
device_map="auto", |
|
) |
|
|
|
messages = [ |
|
{"role": "system", "content": "You are an expert trained on healthcare and biomedical domain!"}, |
|
{"role": "user", "content": "What are the differential diagnoses for a patient presenting with shortness of breath and chest pain?"}, |
|
] |
|
|
|
prompt = pipeline.tokenizer.apply_chat_template( |
|
messages, |
|
tokenize=False, |
|
add_generation_prompt=True |
|
) |
|
|
|
terminators = [ |
|
pipeline.tokenizer.eos_token_id, |
|
pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>") |
|
] |
|
|
|
outputs = pipeline( |
|
prompt, |
|
max_new_tokens=256, |
|
eos_token_id=terminators, |
|
do_sample=True, |
|
temperature=0.6, |
|
top_p=0.9, |
|
) |
|
print(outputs[0]["generated_text"][len(prompt):]) |
|
``` |
|
|
|
## License |
|
|
|
This model is licensed under the [Bio-Medical-Llama-3-2-1B-CoT-012025 (Non-Commercial Use Only)](./LICENSE). Please review the terms and conditions before using the model. |
|
|
|
### Contact Information |
|
|
|
For further information, inquiries, or issues related to Bio-Medical-Llama-3-2-1B-CoT-012025, please contact: |
|
|
|
Email: [email protected] |
|
|
|
Website: [https://www.contactdoctor.in](https://www.contactdoctor.in) |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- **Learning Rate:** 0.0002 |
|
- **Train Batch Size:** 8 |
|
- **Eval Batch Size:** 4 |
|
- **Seed:** 42 |
|
- **Gradient Accumulation Steps:** 8 |
|
- **Total Train Batch Size:** 32 |
|
- **Optimizer:** Adam with betas=(0.9, 0.999) and epsilon=1e-08 |
|
- **LR Scheduler Type:** Cosine |
|
- **LR Scheduler Warmup Ratio:** 0.03 |
|
- **Training Steps:** 2000 |
|
- **Mixed Precision Training:** Native AMP |
|
|
|
### Framework versions |
|
|
|
- **PEFT:** 0.11.0 |
|
- **Transformers:** 4.40.2 |
|
- **Pytorch:** 2.1.2 |
|
- **Datasets:** 2.19.1 |
|
- **Tokenizers:** 0.19.1 |
|
|
|
### Citation |
|
|
|
If you use Bio-Medical-Llama-3-2-1B-CoT-012025 in your research or applications, please cite it as follows: |
|
|
|
```bibtex |
|
@misc{ContactDoctor_Bio-Medical-Llama-3.2-1B-CoT-012025, |
|
author = {ContactDoctor}, |
|
title = {Bio-Medical-Llama-3-2-1B-CoT-012025: A Reasoning-Enhanced Biomedical Language Model}, |
|
year = {2025}, |
|
howpublished = {https://huggingface.co/ContactDoctor/Bio-Medical-Llama-3-2-1B-CoT-012025}, |
|
} |
|
``` |
|
|
|
|