File size: 2,970 Bytes
38292f3
 
 
 
 
 
 
 
cdcb54f
 
 
 
 
 
 
38292f3
 
 
f9b501f
 
38292f3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
---
license: llama3
tags:
- moe
language:
- en
---

<br/><br/>
8bpw/h8 exl2 quantization of [xxx777xxxASD/L3-ChaoticSoliloquy-v2-4x8B-test](https://huggingface.co/xxx777xxxASD/L3-ChaoticSoliloquy-v2-4x8B-test) using default exllamav2 calibration dataset.

---

**ORIGINAL CARD:**

![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f5e51289c121cb864ba464/K3jwX1gq_nohz7Fe78xev.png)
(Maybe i'll change the waifu picture later.)

TRY 1.5 or 1.0 INSTEAD AND CHECK IF THEY WORK BETTER, I DIDN'T TEST THIS VERSION BEFORE PUBLISHING

> [!IMPORTANT]  
> [GGUF quants](https://huggingface.co/collections/xxx777xxxASD/chaoticsoliloquy-v2-4x8b-test-66377d4b9ff6e77b00fa791f)

Experimental RP-oriented MoE, the idea was to get a model that would be equal to or better than the Mixtral 8x7B and it's finetunes in RP/ERP tasks.

The model has totally 25B parameters, of which ~13B are active.

Please feedback me if it's more stable than the [previous version](https://huggingface.co/xxx777xxxASD/L3-ChaoticSoliloquy-v1.5-4x8B)

### Llama 3 ChaoticSoliloquy-v2-4x8B test
```
base_model: L3_ChaosMaid_8B
gate_mode: random
dtype: bfloat16
experts_per_token: 2
experts:
  - source_model: ChaoticNeutrals_Poppy_Porpoise-0.72-L3-8B
  - source_model: L3_ChaosMaid_8B
  - source_model: openlynn_Llama-3-Soliloquy-8B-v2
  - source_model: Sao10K_L3-Solana-8B-v1
```


## Models used

- [ChaoticNeutrals/Poppy_Porpoise-0.72-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-0.72-L3-8B)
- [jeiku/Chaos_RP_l3_8B](https://huggingface.co/jeiku/Chaos_RP_l3_8B)
- [NeverSleep/Llama-3-Lumimaid-8B-v0.1](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1)
- [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
- [Sao10K/L3-Solana-8B-v1](https://huggingface.co/Sao10K/L3-Solana-8B-v1)


## Difference

- Update from [ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B) to [ChaoticNeutrals/Poppy_Porpoise-0.72-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-0.72-L3-8B)
- Update from [openlynn/Llama-3-Soliloquy-8B](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B) to [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
- Change - [NeverSleep/Llama-3-Lumimaid-8B-v0.1](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1) to L3-ChaosMaid-8B

## L3 ChaosMaid-8B
```
models:
  - model: jeiku_Chaos_RP_l3_8B
    # No parameters necessary for base model
  - model: NeverSleep_Llama-3-Lumimaid-8B-v0.1
    parameters:
      density: 0.5
      weight: 0.5
merge_method: dare_ties
base_model: jeiku_Chaos_RP_l3_8B
parameters:
  int8_mask: true
dtype: bfloat16
```

## Vision

[llama3_mmproj](https://huggingface.co/ChaoticNeutrals/LLaVA-Llama-3-8B-mmproj-Updated)

![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f5e51289c121cb864ba464/yv4C6NalqORLjvY3KKZk8.png)


## Prompt format: Llama 3