prithivMLmods
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -60,4 +60,53 @@ generated_ids = [
|
|
60 |
]
|
61 |
|
62 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
63 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
60 |
]
|
61 |
|
62 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
63 |
+
```
|
64 |
+
# **Intended Use**
|
65 |
+
1. **Reasoning and Context Understanding**:
|
66 |
+
Designed to assist with complex reasoning tasks, contextual understanding, and solving problems requiring logical deduction and critical thinking.
|
67 |
+
|
68 |
+
2. **Mathematical Problem-Solving**:
|
69 |
+
Specialized for performing advanced mathematical reasoning and calculations, making it suitable for educational, scientific, and research-oriented applications.
|
70 |
+
|
71 |
+
3. **Code Generation and Debugging**:
|
72 |
+
Offers robust support for coding tasks, including writing, debugging, and optimizing code in various programming languages, ideal for developers and software engineers.
|
73 |
+
|
74 |
+
4. **Structured Data Analysis**:
|
75 |
+
Excels in processing and analyzing structured data, such as tables and JSON, and generating structured outputs, which is useful for data analysts and automation workflows.
|
76 |
+
|
77 |
+
5. **Multilingual Applications**:
|
78 |
+
Supports over 29 languages, making it versatile for global applications like multilingual chatbots, content generation, and translations.
|
79 |
+
|
80 |
+
6. **Extended Content Generation**:
|
81 |
+
Capable of generating long-form content (over 8K tokens), useful for writing reports, articles, and creating detailed instructional guides.
|
82 |
+
|
83 |
+
7. **Interactive Role-Playing and Chatbots**:
|
84 |
+
Enhanced capabilities for role-playing and condition-setting, making it ideal for interactive chatbots, virtual assistants, and entertainment purposes.
|
85 |
+
|
86 |
+
8. **Large-Context Tasks**:
|
87 |
+
With a context window of up to 128K tokens, it is ideal for analyzing or generating large documents, books, or datasets in a single session.
|
88 |
+
|
89 |
+
# **Limitations**
|
90 |
+
1. **Hardware Requirements**:
|
91 |
+
Due to its 20B parameter size and support for long-context inputs, running the model requires significant computational resources, including high-memory GPUs or TPUs.
|
92 |
+
|
93 |
+
2. **Potential Bias in Multilingual Outputs**:
|
94 |
+
While it supports 29 languages, the quality and accuracy of outputs may vary depending on the language, especially for less-resourced languages.
|
95 |
+
|
96 |
+
3. **Inconsistent Outputs for Creative Tasks**:
|
97 |
+
The model may occasionally produce inconsistent or repetitive results in creative writing, storytelling, or highly subjective tasks.
|
98 |
+
|
99 |
+
4. **Limited Real-World Awareness**:
|
100 |
+
It lacks real-time knowledge of current events beyond its training cutoff, which may limit its ability to respond accurately to the latest information.
|
101 |
+
|
102 |
+
5. **Error Propagation in Long-Text Outputs**:
|
103 |
+
In generating long texts, minor errors in early outputs can sometimes propagate, reducing the overall coherence and accuracy of the response.
|
104 |
+
|
105 |
+
6. **Dependency on High-Quality Prompts**:
|
106 |
+
Performance may depend on the quality and specificity of the input prompt, requiring users to carefully design queries for optimal results.
|
107 |
+
|
108 |
+
7. **Sensitivity to Adversarial Inputs**:
|
109 |
+
The model may struggle with adversarial or ambiguous inputs, leading to incorrect or irrelevant outputs.
|
110 |
+
|
111 |
+
8. **Ethical and Safety Concerns**:
|
112 |
+
Potential misuse in generating misleading, harmful, or offensive content remains a concern, and guardrails must be implemented to ensure responsible use.
|