abhiAI777 commited on
Commit
bdb4aad
·
verified ·
1 Parent(s): f680bdc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -1
README.md CHANGED
@@ -29,4 +29,33 @@ It is built on Nemo 12b and trained on different datasets as well as some layer
29
 
30
  - **Base model:** mistralai/Mistral-Nemo-Base-2407
31
  - **Parameter count:** ~12 billion
32
- - **Architecture specifics:** Transformer-based language model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
  - **Base model:** mistralai/Mistral-Nemo-Base-2407
31
  - **Parameter count:** ~12 billion
32
+ - **Architecture specifics:** Transformer-based language model
33
+
34
+ ## Intended Use 🎯
35
+ As an advanced language model for various natural language processing tasks, including but not limited to text generation (excels in chat), question-answering, and analysis.
36
+
37
+ ## Ethical Considerations 🤔
38
+ As a model based on multiple sources, Zinakha-12b may inherit biases and limitations from its constituent models. Users should be aware of potential biases in generated content and use the model responsibly.
39
+
40
+ ## Performance and Evaluation
41
+ Performance metrics and evaluation results for Zinakha-12b are yet to be determined. Users are encouraged to contribute their findings and benchmarks.
42
+
43
+ ## Limitations and Biases
44
+ The model may exhibit biases present in its training data and constituent models. It's crucial to critically evaluate the model's outputs and use them in conjunction with human judgment.
45
+
46
+ ## Additional Information
47
+ For more details on the base model and constituent models, please refer to their respective model cards and documentation.
48
+
49
+ ## How to Use
50
+ ```python
51
+ from transformers import AutoTokenizer, AutoModelForCausalLM
52
+
53
+ model = AutoModelForCausalLM.from_pretrained("aixonlab/Zinakha-12b")
54
+ tokenizer = AutoTokenizer.from_pretrained("aixonlab/Zinakha-12b")
55
+
56
+ prompt = "Once upon a time"
57
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids
58
+
59
+ generated_ids = model.generate(input_ids, max_length=100)
60
+ generated_text = tokenizer.decode(generated_ids, skip_special_tokens=True)
61
+ print(generated_text)