Molbap HF Staff commited on
Commit
cd385a2
Β·
1 Parent(s): 0b2b42c
Files changed (1) hide show
  1. dist/index.html +1 -1
dist/index.html CHANGED
@@ -464,7 +464,7 @@ machinery is the <code>attention mask</code>, cause of confusion. Thankfully, we
464
  <h3>πŸš€ CUDA Warmup Efficiency Benchmark</h3>
465
  </div>
466
  <div class=demo-content>
467
- <iframe src=https://molbap-cuda-warmup-transformers.hf.space width=100% height=600px frameborder=0 style="border-radius: 8px; background: white;"></iframe>
468
  </div>
469
  <div class=demo-footer>
470
  Real CUDA warmup benchmarking with actual Transformers models. Measure the performance impact of the <code>caching_allocator_warmup</code> function at <code>transformers/src/transformers/modeling_utils.py:6186</code>. This interactive tool loads models twice - once with warmup disabled and once with warmup enabled - to demonstrate the significant loading time improvements.
 
464
  <h3>πŸš€ CUDA Warmup Efficiency Benchmark</h3>
465
  </div>
466
  <div class=demo-content>
467
+ <iframe src=https://molbap-cuda-warmup-transformers.hf.space width=100% height=800px frameborder=0 style="border-radius: 8px; background: white;"></iframe>
468
  </div>
469
  <div class=demo-footer>
470
  Real CUDA warmup benchmarking with actual Transformers models. Measure the performance impact of the <code>caching_allocator_warmup</code> function at <code>transformers/src/transformers/modeling_utils.py:6186</code>. This interactive tool loads models twice - once with warmup disabled and once with warmup enabled - to demonstrate the significant loading time improvements.