Transformers
Safetensors
English
medical
text-generation-inference
llm
biomistral
Inference Endpoints
siyah1's picture
Update README.md
85fd568 verified
---
library_name: transformers
tags:
- medical
- text-generation-inference
- llm
- biomistral
license: apache-2.0
datasets:
- ruslanmv/ai-medical-chatbot
- lavita/ChatDoctor-HealthCareMagic-100k
language:
- en
metrics:
- rouge
- bleu
---
# Model Card for BioMistral Multi-Turn Doctor Conversation Model
## Model Details
### Model Description
This model is a fine-tuned version of the BioMistral model, specifically tailored for multi-turn doctor-patient conversations. It leverages the powerful language generation capabilities of BioMistral to provide accurate and context-aware responses in medical dialogue scenarios.
- **Developed by:** Siyahul Haque T P
- **Model type:** Text-generation (LLM)
- **Language(s) (NLP):** English (en)
- **License:** Apache-2.0
- **Finetuned from model:** BioMistral
## Uses
### Direct Use
This model can be directly used for generating responses in multi-turn medical conversations, making it useful for applications like virtual health assistants and medical chatbots.
### Downstream Use
This model can be further fine-tuned or integrated into larger healthcare applications, such as patient management systems or automated symptom checkers.
### Out-of-Scope Use
The model is not suitable for use in emergency medical situations, providing final diagnoses, or replacing professional medical advice.
## Bias, Risks, and Limitations
The model may reflect biases present in the training data, including underrepresentation of certain medical conditions or demographic groups. The model should not be used as a sole source of medical information and must be supervised by qualified healthcare professionals.
### Recommendations
Users should be aware of the potential biases and limitations of the model. It is recommended to use the model as a supplementary tool rather than a primary source of medical advice.
## How to Get Started with the Model
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the tokenizer and model from the Hugging Face Hub
tokenizer = AutoTokenizer.from_pretrained("BioMistral/BioMistral-7B")
model = AutoModelForCausalLM.from_pretrained("siyah1/BioMistral-7b-Chat-Doctor")
# Example input: patient describing a symptom
input_text = "Hello, doctor, I have a headache."
# Tokenize the input text
inputs = tokenizer(input_text, return_tensors="pt")
# Generate a response from the model
outputs = model.generate(**inputs, max_length=100, num_return_sequences=1)
# Decode the generated response
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
# Print the model's response
print("Doctor:", response)