Sweaterdog commited on
Commit
5583e0e
·
verified ·
1 Parent(s): 82fdbf2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -5
README.md CHANGED
@@ -1,5 +1,34 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - unsloth
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - unsloth
5
+ ---
6
+
7
+ # THIS IS A PREVIEW LORA
8
+
9
+
10
+ This modelfile is the LoRA adapter for [Andy-3.5-reasoning](https://huggingface.co/Sweaterdog/Andy-3.5)
11
+
12
+ # Why this exists
13
+
14
+ This Repo exists because I wanted to make Andy-3.5, as well as its derivitives, such as Andy-3.5-reasoning, fully open-source. Via Unsloth, you are able to continue fine tuning where I left off, so if you made your own dataset, you can continue tuning Andy-3.5 for your exact use case.
15
+
16
+ # What if I fine tune off of Andy-3.5?
17
+
18
+ If you fine tune Andy-3.5 on your dataset, my dataset, **or any other dataset**, you **have** to provide credit to me for making the base model, which is Andy-3.5, if you wish, you may call the model Andy-3.5-base
19
+
20
+ # Why would I want to fine tune off of Andy-3.5?
21
+
22
+ Andy-3.5 has a significant amount of knowledge regarding Minecraft and MindCraft, but not unlimited. Andy-3.5 can be trained further on Minecraft knowledge to make the model better, and if you strive for maximum efficiency, it would be best to continue fine-tuning a model based on similar data to help it.
23
+
24
+ # What should I call my model if I do tune it?
25
+
26
+ You may name it whatever you'd like, but if I may suggest, I would recommend a name that clearly references the fact it originated from Andy-3.5.
27
+
28
+ If you'd like an example, if I trained Andy-3.5 on speedrunning tactics, I would call the model **Andy-3.5-Speedrun** or something similar.
29
+
30
+ # Important notes:
31
+
32
+ * I do not suggest fine tuning off of this model for anything besides reasoning
33
+ * I do not suggest fine tuning off of this model **currently** as it is a preview model
34
+ * I do not suggest fine tuning this model with **any dataset** for reasoning that **does not** use the DeepSeek-R1 method of thinking.