LumenscopeAI commited on
Commit
e52ee89
·
verified ·
1 Parent(s): af81931

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -3
README.md CHANGED
@@ -65,11 +65,18 @@ model-index:
65
 
66
  # BrainTransformers: SNN-LLM
67
 
68
- Based on BrainTransformers, BrainGPTForCausalLM is a Large Language Model (LLM) implemented using Spiking Neural Networks (SNN). Our technical report will be uploaded to arXiv as soon as possible. We plan to further optimize the model at the operator level and adapt it for hardware compatibility, enabling BrainGPTForCausalLM to be deployed on more energy-efficient SNN hardware devices.
 
 
 
 
 
 
 
 
69
 
70
  ## Model Availability
71
 
72
- - The current pre-trained model parameters have been published on ModelScope: [DataLinguistic/BrainTransformers-3B-Chat](https://www.modelscope.cn/models/DataLinguistic/BrainTransformers-3B-Chat)
73
  - The current pre-trained model parameters have been published on Hugging Face.[LumenscopeAI/BrainTransformers-3B-Chat](https://huggingface.co/LumenscopeAI/BrainTransformers-3B-Chat)
74
 
75
  ## Repository
@@ -148,4 +155,10 @@ messages = [
148
  ]
149
  response = generate_text(messages)
150
  print(response)
151
- ```
 
 
 
 
 
 
 
65
 
66
  # BrainTransformers: SNN-LLM
67
 
68
+ Based on BrainTransformers, BrainGPTForCausalLM is a Large Language Model (LLM) implemented using Spiking Neural Networks (SNN). We are excited to announce that an initial version of our technical report is now available in our GitHub repository. This early release allows the community to access our findings while the full report undergoes the arXiv review process.
69
+
70
+ Our comprehensive technical report has been submitted to arXiv and is currently in the "on hold" status, pending review. We will be releasing our findings in stages, with updates and more detailed analyses to follow. The full report will be available on arXiv as soon as the review process is completed.
71
+
72
+ We plan to further optimize the model at the operator level and adapt it for hardware compatibility, enabling BrainGPTForCausalLM to be deployed on more energy-efficient SNN hardware devices.
73
+
74
+ The current open-source version retains some floating-point calculations to ensure computational efficiency. We will continue to optimize this. Some detailed explanations are provided in the comments within the source code.
75
+
76
+ Stay tuned for updates as we continue to refine and expand our research findings.
77
 
78
  ## Model Availability
79
 
 
80
  - The current pre-trained model parameters have been published on Hugging Face.[LumenscopeAI/BrainTransformers-3B-Chat](https://huggingface.co/LumenscopeAI/BrainTransformers-3B-Chat)
81
 
82
  ## Repository
 
155
  ]
156
  response = generate_text(messages)
157
  print(response)
158
+ ```
159
+
160
+ ## Acknowledgments
161
+
162
+ The model was trained using ANN-Base-Qwen2, with a total of three training stages, including SNN-specific neuron synaptic plasticity training. The technical report is still being prepared. Please note that SNN models do not support ANN fine-tuning techniques. We are currently developing specialized fine-tuning code tools for SNN models. Our open-source model has achieved leading SOTA results, and we welcome your stars.
163
+
164
+ This repository includes a complete transformers package, which can directly replace the transformers package in your development environment. This allows compatibility with our SNN-Base-LLM without affecting existing usage.