--- license: apache-2.0 datasets: - Sweaterdog/Andy-3.5 language: - en base_model: - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B tags: - Minecraft --- # Welcome to a new generation of Minecraft ## Andy 3.5 is a collection of LOCAL LLM's designed for playing Minecraft Andy 3.5 is designed to be used with MindCraft ### How to Install 1. Select the model you would like to use *(Larger is better)* 2. Download a Modelfile *(One is directly for the tuned model, while the other is for the base model) 3. Once downloaded, open Modelfile in a text editor, and change the path to the download location of the gguf file 4. When changed, save the file, and open command terminal 5. *(Optional if CMD is opened via file explorer)* Navigate to the correct directory using "cd" 6. Run the command ollama create Andy-3.5 -f Modelfile 7. Go to a profile in MindCraft 8. Change the model to be Andy-3.5 9. Enjoy playing with an AI ## How was model trained? The model was trained on a dataset of ~10,000 messages coming directly from MindCraft, ensuring quality data ## What are capabilities and Limitations? The smaller model *(The preview ones at least)* had 1/3 of the parameters tuned, the larger preview model has not been released yet. Andy-3.5 was trained on EVERYTHING regarding Minecraft and MindCraft, it knows how to use commands natively without a system prompt. Andy-3.5 also knows how to build / use !newAction to perform commands, it was trained on lots of building, as well as, using !newAction to do tasks like manually making something or strip mining. ****Know this is a PREVIEW model, it is NOT finished!**** ## Why a preview model? Andy-3.5-preview was made to test the intelligence of a Minecraft Ai with the current dataset, it was meant to see the progress of the training and what area's are needed for the future DO NOT expect this model to be able to do everything perfectly, it only knows as much as the dataset told it, as well as the other 2/3 of the untouched parameters allow. The model *may* experience bugs, such as not saying your name, getting previous messages confused, or other small things. ## Important notes and considerations The preview model of Andy-3.5-mini *(Andy-3.5-mini-preview)* was trained on a context length of 4096, this was meant to speed up training and VRAM usage. The Base model of Andy-3.5-mini-preview was a distilled version of Deepseek-R1, which was a tuned model of Qwen-2.5-1.5b Since a context window of 4096 is not nearly enough for MindCraft, you can go higher, Qwen-2.5-1.5b was trained on a context length of 64,000, the distilled version of Deepseek-R1 was trained on a length of 128,000, the usable length may be closer to 32,000 tokens for Andy-3.5-mini-preview When the full versions of Andy-3.5 and Andy-3.5-preview release, they will both be trained on a context length of 128,000 to ensure proper usage during playing.