|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
library_name: transformers |
|
tags: |
|
- nlp |
|
- llm |
|
--- |
|
# K2 - Deciphering Llama 2 70B |
|
|
|
K2 is a fully transparent large language model on par with Llama 2 - 70B. |
|
|
|
[insert table with comparison to llama 2 on open llm leaderboard metrics] |
|
|
|
[insert about the data mix] |
|
|
|
|
|
The first 10 checkpoints are available for download now. |
|
|
|
- Checkpoint 360[link] |
|
- Checkpoint 359[link] |
|
- Checkpoint 358[link] |
|
- Checkpoint 357[link] |
|
- Checkpoint 356[link] |
|
- Checkpoint 355[link] |
|
- Checkpoint 354[link] |
|
- Checkpoint 353[link] |
|
- Checkpoint 352[link] |
|
- Checkpoint 351[link] |
|
|
|
## First 10 Checkpoints |
|
| Checkpoints | | |
|
| ----------- | ----------- | |
|
| Checkpoint 360[link] | Checkpoint 355[link] | |
|
| Checkpoint 359[link] | Checkpoint 354[link] | |
|
| Checkpoint 358[link] | Checkpoint 353[link] | |
|
| Checkpoint 357[link] | Checkpoint 352[link] | |
|
| Checkpoint 356[link] | Checkpoint 351[link] | |
|
|
|
|
|
We are working on release caliber artifacts for the dataset, code, and analysis which will be released over the next few weeks. |
|
|
|
|
|
## 🟠 Model Description |
|
|
|
- **Model type:** Language model with the same architecture as LLaMA-7B |
|
- **Language(s) (NLP):** English |
|
- **License:** Apache 2.0 |
|
- **Resources for more information:** |
|
- [Training Code] |
|
- [Data Preparation] |
|
- [Metrics] |
|
- [Fully processed Amber pretraining data] |
|
|
|
|
|
## About LLM360 |
|
LLM360 is an initiative for comprehensive and fully open-sourced LLMs, |
|
where all training details, model checkpoints, intermediate results, and |
|
additional analyses are made available to the community. Our goal is to advance |
|
the field by inviting the community to deepen the understanding of LLMs |
|
together. As the first step of the project LLM360, we release all intermediate |
|
model checkpoints, our fully-prepared pre-training dataset, all source code and |
|
configurations, and training details. We are |
|
committed to continually pushing the boundaries of LLMs through this open-source |
|
effort. |
|
|
|
[Visit us](https://www.llm360.ai/) |
|
|