File size: 3,426 Bytes
f615d22
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f74da0d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
---
license: llama3.2
datasets:
- AI-MO/NuminaMath-CoT
- prithivMLmods/Math-Solve
- amphora/QwQ-LongCoT-130K
- prithivMLmods/Deepthink-Reasoning
language:
- en
base_model:
- meta-llama/Llama-3.2-1B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- Express
- Llama
- Ollama
- v.1
- text-generation-inference
---

# **Llama-Express.1**

Llama-Express.1 is a 1B model based on Llama 3.2 (1B), fine-tuned on long chain-of-thought datasets. This instruction-tuned, text-only model is optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. It outperforms many of the available open-source and closed chat models.

# **Use with transformers**

Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.

Make sure to update your transformers installation via `pip install --upgrade transformers`.

```python
import torch
from transformers import pipeline

model_id = "prithivMLmods/Llama-Express.1"
pipe = pipeline(
    "text-generation",
    model=model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
messages = [
    {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
    {"role": "user", "content": "Who are you?"},
]
outputs = pipe(
    messages,
    max_new_tokens=256,
)
print(outputs[0]["generated_text"][-1])
```

# **Intended Use**  
1. **Multilingual Dialogue**:  
   - Designed for high-quality, multilingual conversations, making it suitable for applications requiring natural, fluid dialogue across languages.  

2. **Agentic Retrieval**:  
   - Optimized for retrieval-based tasks where reasoning and contextual chaining are crucial for extracting and summarizing relevant information.  

3. **Summarization Tasks**:  
   - Effective in generating concise and accurate summaries from complex and lengthy texts, suitable for academic, professional, and casual use cases.  

4. **Instruction-Following Applications**:  
   - Fine-tuned for tasks requiring adherence to user-provided instructions, making it ideal for automation workflows, content creation, and virtual assistant integrations.  

# **Limitations**  
1. **Monomodal Focus**:  
   - As a text-only model, it cannot process multimodal inputs like images, audio, or videos, limiting its versatility in multimedia applications.  

2. **Context Length Constraints**:  
   - While optimized for long chain-of-thought reasoning, extreme cases with very large contexts may still lead to degraded performance or truncation issues.  

3. **Bias and Ethics**:  
   - The model might reflect biases present in the training datasets, potentially resulting in outputs that could be culturally insensitive or inappropriate.  

4. **Performance in Low-Resource Languages**:  
   - While multilingual, its effectiveness may vary across languages, with possible performance drops in underrepresented or low-resource languages.  

5. **Dependency on Input Quality**:  
   - The model's output is heavily influenced by the clarity and specificity of the input instructions. Ambiguous or vague prompts may lead to suboptimal results.  

6. **Lack of Real-Time Internet Access**:  
   - Without real-time retrieval capabilities, it cannot provide up-to-date information or verify facts against the latest data.