update
Browse files- dist/index.html +1 -1
dist/index.html
CHANGED
|
@@ -464,7 +464,7 @@ machinery is the <code>attention mask</code>, cause of confusion. Thankfully, we
|
|
| 464 |
<h3>π CUDA Warmup Efficiency Benchmark</h3>
|
| 465 |
</div>
|
| 466 |
<div class=demo-content>
|
| 467 |
-
<iframe src=https://molbap-cuda-warmup-transformers.hf.space width=100% height=
|
| 468 |
</div>
|
| 469 |
<div class=demo-footer>
|
| 470 |
Real CUDA warmup benchmarking with actual Transformers models. Measure the performance impact of the <code>caching_allocator_warmup</code> function at <code>transformers/src/transformers/modeling_utils.py:6186</code>. This interactive tool loads models twice - once with warmup disabled and once with warmup enabled - to demonstrate the significant loading time improvements.
|
|
|
|
| 464 |
<h3>π CUDA Warmup Efficiency Benchmark</h3>
|
| 465 |
</div>
|
| 466 |
<div class=demo-content>
|
| 467 |
+
<iframe src=https://molbap-cuda-warmup-transformers.hf.space width=100% height=800px frameborder=0 style="border-radius: 8px; background: white;"></iframe>
|
| 468 |
</div>
|
| 469 |
<div class=demo-footer>
|
| 470 |
Real CUDA warmup benchmarking with actual Transformers models. Measure the performance impact of the <code>caching_allocator_warmup</code> function at <code>transformers/src/transformers/modeling_utils.py:6186</code>. This interactive tool loads models twice - once with warmup disabled and once with warmup enabled - to demonstrate the significant loading time improvements.
|