metadata
language:
- en
license: apache-2.0
library_name: transformers
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
- code
- 'medical '
- farmer
- doctor
- Mega-Series
- Cyber-Series
- Role-Play
- Self-Rag
- ThinkingBot
- milestone
- mega-series
- SpydazWebAI
base_model: LeroyDyer/Mixtral_AI_CyberTron_Ultra
pipeline_tag: text-generation
inference: false
metrics:
- accuracy
- bertscore
- bleu
- brier_score
- cer
- character
- charcut_mt
- chrf
- code_eval
datasets:
- gretelai/synthetic_text_to_sql
- HuggingFaceTB/cosmopedia
- teknium/OpenHermes-2.5
- Open-Orca/SlimOrca
- Open-Orca/OpenOrca
- cognitivecomputations/dolphin-coder
- databricks/databricks-dolly-15k
- yahma/alpaca-cleaned
- uonlp/CulturaX
- mwitiderrick/SwahiliPlatypus
- swahili
- Rogendo/English-Swahili-Sentence-Pairs
- ise-uiuc/Magicoder-Evol-Instruct-110K
- meta-math/MetaMathQA
quantized_by: Suparious
LeroyDyer/Mixtral_AI_CyberTron_Ultra AWQ
- Model creator: LeroyDyer
- Original model: Mixtral_AI_CyberTron_Ultra
Model Summary
What does he NOT KNOW ! that is the question!
MOTTO FOR MODEL!
Models are the same as loras , take them with light weight they are like tablets of knowledge!
Exactly ! ( models / loras ? is there a difference ? only mega merges make a true difference ! the small merges are just applying an adapter lol - Its in there somewhere?)