LDJnr commited on
Commit
47fcafd
·
1 Parent(s): 822dbef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -4
README.md CHANGED
@@ -11,7 +11,6 @@ GGML 4bit Quantization of Nous Research's Puffin Preview 1 Model: https://huggin
11
 
12
  *Thank you to Eachadea for making this quantization possible immediately upon launch*
13
 
14
-
15
 
16
  ![puffin](https://i.imgur.com/R2xTHMb.png)
17
 
@@ -30,7 +29,7 @@ Notable mentions for assisting in some of the training issues goes to: Caseus an
30
 
31
  ## Model Training
32
 
33
- Redmond-Puffin-13B is a new model trained for multiple epochs on a dataset of 3,000 carefully curated GPT-4 examples, most of which are long context conversations between a real human and GPT-4.
34
 
35
  Additional data came from carefully curated subsections of datasets such as CamelAI's Physics, Chemistry, Biology and Math.
36
 
@@ -44,6 +43,12 @@ The model follows the Vicuna ShareGPT prompt format:
44
  ### gpt:
45
  ```
46
 
 
 
 
 
 
 
47
  ## Notable Features:
48
 
49
  - The first Llama-2 based fine-tuned model released by Nous Research.
@@ -66,9 +71,13 @@ We plan to have these solved in an updated Puffin model in the very near future,
66
 
67
  This is a relatively early build amongst the grand plans for the future of Puffin!
68
 
69
- Current limitations: Some token mismatch problems and formatting issues have been idenitifed, these may very possibly effect the current output quality, we plan to have these solved in an updated Puffin model in the near future.
 
 
 
 
70
 
71
- In the near future we plan on releasing an improved version of the model with the help of domain specific expert volunteers, which will help eliminate any wrong data from this curation and improve the further ones.
72
 
73
  ## Benchmarks coming soon
74
 
 
11
 
12
  *Thank you to Eachadea for making this quantization possible immediately upon launch*
13
 
 
14
 
15
  ![puffin](https://i.imgur.com/R2xTHMb.png)
16
 
 
29
 
30
  ## Model Training
31
 
32
+ Redmond-Puffin-13B-V1.3 is a new model trained for multiple epochs on a dataset of 3,000 carefully curated GPT-4 examples, most of which are long context conversations between a real human and GPT-4.
33
 
34
  Additional data came from carefully curated subsections of datasets such as CamelAI's Physics, Chemistry, Biology and Math.
35
 
 
43
  ### gpt:
44
  ```
45
 
46
+ ## Improvements over previous version:
47
+
48
+ The original Puffin model was loved by many, however it was quickly discovered to have dataset errors in a significant amount of the conversations.
49
+ Puffin-V1.3 dataset solves this issue and the resulting fixed model has now fully finished training!
50
+
51
+
52
  ## Notable Features:
53
 
54
  - The first Llama-2 based fine-tuned model released by Nous Research.
 
71
 
72
  This is a relatively early build amongst the grand plans for the future of Puffin!
73
 
74
+ Current limitations: Some token mismatch problems have been identified, these may effect the current output quality, we plan to have this solved in Puffin V2 along with other improvements.
75
+
76
+ ## How you can help!
77
+
78
+ In the near future we plan on leveraging the help of domain specific expert volunteers to eliminate any mathematically/verifiably incorrect answers from our training curations.
79
 
80
+ If you have at-least a bachelors in mathematics, physics, biology or chemistry and would like to volunteer even just 30 minutes of your expertise time, please contact ldj on discord!
81
 
82
  ## Benchmarks coming soon
83