nouamanetazi HF staff commited on
Commit
0a8f3ab
·
verified ·
1 Parent(s): ce06f49
assets/images/predict_memory_tool.png ADDED

Git LFS Details

  • SHA256: b079e5968c1ddfff6f0f663db43f6fb9715240e92dd455875194575eb4c98313
  • Pointer size: 130 Bytes
  • Size of remote file: 94.3 kB
assets/images/profile_trace_annotated.png CHANGED

Git LFS Details

  • SHA256: e1806f717e427febe26bfa45135d45d76adc9808c8a92553f7f7e0bb9faa80ae
  • Pointer size: 131 Bytes
  • Size of remote file: 995 kB

Git LFS Details

  • SHA256: 7359ca99eff4eaa53952bfba0dd562ab6bb9b109033f283d296aef2471e642bc
  • Pointer size: 131 Bytes
  • Size of remote file: 995 kB
dist/assets/images/predict_memory_tool.png ADDED

Git LFS Details

  • SHA256: a61828c60b0e39e57c3d050474889dc51c87f5fdaa6d5afa4ae7b55e329678b2
  • Pointer size: 130 Bytes
  • Size of remote file: 26.5 kB
dist/index.html CHANGED
@@ -226,19 +226,12 @@
226
  </div>
227
  <p>(Don't worry if you have no idea what's happening in this widget. That's why we're here.)</p>
228
 
229
- <p>While this widget gives a theoretical breakdown the following tool can be used to predict the memory usage:</p>
230
- <ul>
231
- <li>
232
- <p>
233
- <a href="https://huggingface.co/spaces/nanotron/predict_memory">predict_memory</a>
234
- </p>
235
- </li>
236
- <li>
237
- <p>
238
- <a href="https://pytorch.org/docs/stable/torch_cuda_memory.html">torch_cuda_memory</a>
239
- </p>
240
- </li>
241
- </ul>
242
 
243
  <p><strong>Clear code implementations:</strong> theory is one thing, but we discover all kinds of edge cases and important details when we implement something. That’s why we link to implementation references where possible. Depending on the case, we’ll use two code references:</p>
244
 
 
226
  </div>
227
  <p>(Don't worry if you have no idea what's happening in this widget. That's why we're here.)</p>
228
 
229
+ <p>While this widget gives a theoretical breakdown we also made the <a href="https://huggingface.co/spaces/nanotron/predict_memory">following tool</a> that can be used to predict the memory usage during a training run:</p>
230
+
231
+ <a href="https://huggingface.co/spaces/nanotron/predict_memory">
232
+ <img src="/assets/images/predict_memory_tool.png" alt="Predict Memory Tool" />
233
+ </a>
234
+
 
 
 
 
 
 
 
235
 
236
  <p><strong>Clear code implementations:</strong> theory is one thing, but we discover all kinds of edge cases and important details when we implement something. That’s why we link to implementation references where possible. Depending on the case, we’ll use two code references:</p>
237
 
src/index.html CHANGED
@@ -226,19 +226,12 @@
226
  </div>
227
  <p>(Don't worry if you have no idea what's happening in this widget. That's why we're here.)</p>
228
 
229
- <p>While this widget gives a theoretical breakdown the following tool can be used to predict the memory usage:</p>
230
- <ul>
231
- <li>
232
- <p>
233
- <a href="https://huggingface.co/spaces/nanotron/predict_memory">predict_memory</a>
234
- </p>
235
- </li>
236
- <li>
237
- <p>
238
- <a href="https://pytorch.org/docs/stable/torch_cuda_memory.html">torch_cuda_memory</a>
239
- </p>
240
- </li>
241
- </ul>
242
 
243
  <p><strong>Clear code implementations:</strong> theory is one thing, but we discover all kinds of edge cases and important details when we implement something. That’s why we link to implementation references where possible. Depending on the case, we’ll use two code references:</p>
244
 
 
226
  </div>
227
  <p>(Don't worry if you have no idea what's happening in this widget. That's why we're here.)</p>
228
 
229
+ <p>While this widget gives a theoretical breakdown we also made the <a href="https://huggingface.co/spaces/nanotron/predict_memory">following tool</a> that can be used to predict the memory usage during a training run:</p>
230
+
231
+ <a href="https://huggingface.co/spaces/nanotron/predict_memory">
232
+ <img src="/assets/images/predict_memory_tool.png" alt="Predict Memory Tool" />
233
+ </a>
234
+
 
 
 
 
 
 
 
235
 
236
  <p><strong>Clear code implementations:</strong> theory is one thing, but we discover all kinds of edge cases and important details when we implement something. That’s why we link to implementation references where possible. Depending on the case, we’ll use two code references:</p>
237