|
--- |
|
base_model: BioMistral/BioMistral-7B |
|
library_name: peft |
|
license: apache-2.0 |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
tags: |
|
- biology |
|
- medical |
|
--- |
|
|
|
# Model Card for BioMistral-7B-Finetuned |
|
|
|
## Model Summary |
|
|
|
**BioMistral-7B-Finetuned** is a biomedical language model adapted from the BioMistral-7B model. This fine-tuned model is tailored for biomedical question-answering tasks and optimized through LoRA (Low-Rank Adaptation) on a 4-bit quantized base. It is particularly useful for tasks that require understanding and generating biomedical text in English. |
|
|
|
--- |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
This model was fine-tuned for biomedical applications, primarily focusing on enhancing accuracy in question-answering tasks within this domain. |
|
|
|
- **Base Model**: BioMistral-7B |
|
- **License**: apache-2.0 |
|
- **Fine-tuned for Task**: Biomedical Q&A, text generation |
|
- **Quantization**: 4-bit precision with BitsAndBytes for efficient deployment |
|
|
|
## Uses |
|
|
|
### Direct Use |
|
|
|
The model is suitable for biomedical question-answering and other related language generation tasks. |
|
|
|
### Out-of-Scope Use |
|
|
|
Not recommended for general-purpose NLP tasks outside the biomedical domain or for clinical decision-making. |
|
|
|
--- |
|
|
|
## How to Get Started with the Model |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
from peft import PeftModel |
|
|
|
# Load tokenizer and model |
|
tokenizer = AutoTokenizer.from_pretrained("BeastGokul/BioMistral-7B-Finetuned") |
|
model = AutoModelForCausalLM.from_pretrained("BeastGokul/BioMistral-7B-Finetuned") |
|
|
|
# Example usage |
|
input_text = "What are the symptoms of diabetes?" |
|
inputs = tokenizer(input_text, return_tensors="pt") |
|
outputs = model.generate(**inputs) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |
|
--- |
|
## Training Details |
|
### Training Procedure |
|
The model was fine-tuned using the LoRA (Low-Rank Adaptation) method, with a configuration set for biomedical question-answering. |
|
|
|
Training Hyperparameters |
|
Precision: 4-bit quantization with BitsAndBytes |
|
Learning Rate: 2e-5 |
|
Batch Size: Effective batch size of 16 (4 per device, gradient accumulation steps of 4) |
|
Number of Epochs: 3 |
|
|
|
|
|
## Framework versions |
|
PEFT 0.13.2 |