gelukuMLG's picture
Update README.md
3e0ec1c verified
|
raw
history blame
432 Bytes
metadata
license: llama3

This is an experimental 2x8B moe with random gates, using the following 2 models

Important

Make sure to add </s> a stop sequence as it uses llama-3-cat-8B-instruct-V1 as the base model.