Edit model card
YAML Metadata Warning: The pipeline tag "conversational" is not in the official list: text-classification, token-classification, table-question-answering, question-answering, zero-shot-classification, translation, summarization, feature-extraction, text-generation, text2text-generation, fill-mask, sentence-similarity, text-to-speech, text-to-audio, automatic-speech-recognition, audio-to-audio, audio-classification, voice-activity-detection, depth-estimation, image-classification, object-detection, image-segmentation, text-to-image, image-to-text, image-to-image, image-to-video, unconditional-image-generation, video-classification, reinforcement-learning, robotics, tabular-classification, tabular-regression, tabular-to-text, table-to-text, multiple-choice, text-retrieval, time-series-forecasting, text-to-video, image-text-to-text, visual-question-answering, document-question-answering, zero-shot-image-classification, graph-ml, mask-generation, zero-shot-object-detection, text-to-3d, image-to-3d, image-feature-extraction, video-text-to-text, keypoint-detection, any-to-any, other

Giant Macaroni 120b

img

An auto-regressive causal LM created by combining 3x finetuned models into one via passthrough merging slices in a stacked order.

I'll add more later but it is a combination of :

AIDC-ai-business/Marcoroni-70B-v1 garage-bAInd/Platypus2-70B-instruct fangloveskari/ORCA_LLaMA_70B_QLoRA

This is an attempt to create a larger model capable of handling logic and reason well. All three of these models score well in these categories and I used bertviz to find good splice points.

I hope to find out how much $$$ it is going to take to settle this merge with some fine tuning. I think it will do very well once it is settled. These three models could not mix with others I tried because of some changes in their layers involving the addition of a rotary embedding. This seemed to mess with the tokens and could not be mixed with other models that didn't have it. However, all models that do well in logic and reasoning seem to have this addition.

Prompting Format

Both Vicuna and Alpaca will work, but due the final layers belonging primarily to Marcoroni.

Benchmarks

Coming soon.

Acknowledgements

@chargoddard - mergekit. @migtissera - for Tess-XL which inspired me to believe that open models can compete on logic tasks with the big commercial models. @alpindale - for Goliath-120B that started this crazy endeavor for us all @nsfwthrowitaway69 - for sharing the merge config for Venus-120B and getting me off the starting block with some questions on mergekit and tokenizers

Keep it open and keep sharing everyone! With Mixtral and MOE changes to mergekit coupled with these larger merged models? I think the sky is the limit for us all. I can only imagine what will happen if we took a group of these 120 models, fin tuned them each a bit and applied the MOE Mixtral merge method to them? I would also point out that if a clever VC came along and funded that work? You have the people you need right here on huggingface and all they need is the equipment to do it on.

Downloads last month
15
Safetensors
Model size
120B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ibivibiv/giant-macaroni-120b

Quantizations
2 models

Collection including ibivibiv/giant-macaroni-120b