Update README.md
Browse files
README.md
CHANGED
@@ -29,4 +29,33 @@ It is built on Nemo 12b and trained on different datasets as well as some layer
|
|
29 |
|
30 |
- **Base model:** mistralai/Mistral-Nemo-Base-2407
|
31 |
- **Parameter count:** ~12 billion
|
32 |
-
- **Architecture specifics:** Transformer-based language model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
|
30 |
- **Base model:** mistralai/Mistral-Nemo-Base-2407
|
31 |
- **Parameter count:** ~12 billion
|
32 |
+
- **Architecture specifics:** Transformer-based language model
|
33 |
+
|
34 |
+
## Intended Use 🎯
|
35 |
+
As an advanced language model for various natural language processing tasks, including but not limited to text generation (excels in chat), question-answering, and analysis.
|
36 |
+
|
37 |
+
## Ethical Considerations 🤔
|
38 |
+
As a model based on multiple sources, Zinakha-12b may inherit biases and limitations from its constituent models. Users should be aware of potential biases in generated content and use the model responsibly.
|
39 |
+
|
40 |
+
## Performance and Evaluation
|
41 |
+
Performance metrics and evaluation results for Zinakha-12b are yet to be determined. Users are encouraged to contribute their findings and benchmarks.
|
42 |
+
|
43 |
+
## Limitations and Biases
|
44 |
+
The model may exhibit biases present in its training data and constituent models. It's crucial to critically evaluate the model's outputs and use them in conjunction with human judgment.
|
45 |
+
|
46 |
+
## Additional Information
|
47 |
+
For more details on the base model and constituent models, please refer to their respective model cards and documentation.
|
48 |
+
|
49 |
+
## How to Use
|
50 |
+
```python
|
51 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
52 |
+
|
53 |
+
model = AutoModelForCausalLM.from_pretrained("aixonlab/Zinakha-12b")
|
54 |
+
tokenizer = AutoTokenizer.from_pretrained("aixonlab/Zinakha-12b")
|
55 |
+
|
56 |
+
prompt = "Once upon a time"
|
57 |
+
input_ids = tokenizer(prompt, return_tensors="pt").input_ids
|
58 |
+
|
59 |
+
generated_ids = model.generate(input_ids, max_length=100)
|
60 |
+
generated_text = tokenizer.decode(generated_ids, skip_special_tokens=True)
|
61 |
+
print(generated_text)
|