--- license: other license_name: deepnight-responsible-ai license_link: LICENSE language: - en library_name: transformers pipeline_tag: text-generation tags: - 600B - Python - Code - Logical Understanding - Relation Establishment - Translation - ai1 - DEEPNIGHT ---
# DEEPNIGHT ai1 The 600 Billion+ Parameter Model. Yes! We did this! The second largest model in the world, right after GPT-4. --- We at [DEEPNIGHT](https://deepnight.tech) have been working on this for quite some time. We have successfully built the second largest model called ai1 which comes with 600 Billion+ parameters. `ai1` can perform as good as GPT-4 and has a context-window of 8k tokens. ai1 was trained with a new approach where after training the model on a corpus of text from various sources including but not limited to: - RefinedWeb - Opensource code from GitHub - Common Crawl we fine-tuned the model on a huge dataset (generated manually and with automation) for logical understanding and reasoning. We also trained the model for function calling capabilities. --- ## What is special about ai1? ai1 works on a chaining methodology which is built-in. When it receives an input from the user, it tries to understand the input before acting on generation. It generates an instruction-based prompt internally and then works on generation of the response. Benefit of this? We'll just say the jobs of Prompt Engineering are over. Unlike ChatGPT, GPT-4, Llama, and other models, ai1 doesn't require heavy prompt engineering to provide answers. The understanding-development phase in the model takes care of that. What else? - performs as good as GPT-4 - excels in automation tasks - can predict emotions of the user by the conversation (while understanding the input in Phase-1) resulting in better and curated generations. - has an understanding towards human-emotions which helps the model curate the content accordingly - excels in roleplays - excels in writing code - the model has a few global memory units which are used to store data away from the context-window. These memory units are mostly used to store the function schemas but in the end the model decides itself what to store in them. - if we consider how much would it cost, well, on an average $0.005 per 1000 tokens. --- ## Future goals We don't discuss that. Specially after seeing how SOME AI COMPANY ON THEIR DEV DAY just used the opensource research and publications to profit themselves... Hah. --- ## Are we going to allow access? Not for some time. We are still running evaluations and have a lot to learn about how this model can be made better. --- Feel free to reach out to us at research@deepnight.tech - Team [DEEPNIGHT](https://deepnight.tech)