Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,34 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
tags:
|
4 |
-
- unsloth
|
5 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- unsloth
|
5 |
+
---
|
6 |
+
|
7 |
+
# THIS IS A PREVIEW LORA
|
8 |
+
|
9 |
+
|
10 |
+
This modelfile is the LoRA adapter for [Andy-3.5-reasoning](https://huggingface.co/Sweaterdog/Andy-3.5)
|
11 |
+
|
12 |
+
# Why this exists
|
13 |
+
|
14 |
+
This Repo exists because I wanted to make Andy-3.5, as well as its derivitives, such as Andy-3.5-reasoning, fully open-source. Via Unsloth, you are able to continue fine tuning where I left off, so if you made your own dataset, you can continue tuning Andy-3.5 for your exact use case.
|
15 |
+
|
16 |
+
# What if I fine tune off of Andy-3.5?
|
17 |
+
|
18 |
+
If you fine tune Andy-3.5 on your dataset, my dataset, **or any other dataset**, you **have** to provide credit to me for making the base model, which is Andy-3.5, if you wish, you may call the model Andy-3.5-base
|
19 |
+
|
20 |
+
# Why would I want to fine tune off of Andy-3.5?
|
21 |
+
|
22 |
+
Andy-3.5 has a significant amount of knowledge regarding Minecraft and MindCraft, but not unlimited. Andy-3.5 can be trained further on Minecraft knowledge to make the model better, and if you strive for maximum efficiency, it would be best to continue fine-tuning a model based on similar data to help it.
|
23 |
+
|
24 |
+
# What should I call my model if I do tune it?
|
25 |
+
|
26 |
+
You may name it whatever you'd like, but if I may suggest, I would recommend a name that clearly references the fact it originated from Andy-3.5.
|
27 |
+
|
28 |
+
If you'd like an example, if I trained Andy-3.5 on speedrunning tactics, I would call the model **Andy-3.5-Speedrun** or something similar.
|
29 |
+
|
30 |
+
# Important notes:
|
31 |
+
|
32 |
+
* I do not suggest fine tuning off of this model for anything besides reasoning
|
33 |
+
* I do not suggest fine tuning off of this model **currently** as it is a preview model
|
34 |
+
* I do not suggest fine tuning this model with **any dataset** for reasoning that **does not** use the DeepSeek-R1 method of thinking.
|