ui: update cta buttons
Browse files- dist/distill.bundle.js +1 -1
- dist/distill.bundle.js.map +0 -0
- dist/index.html +5 -6
- dist/style.css +1 -1
- src/distill.js +1 -12
- src/index.html +5 -6
- src/style.css +1 -1
dist/distill.bundle.js
CHANGED
@@ -2146,7 +2146,7 @@ function _arrayWithHoles(r) { if (Array.isArray(r)) return r; }
|
|
2146 |
function bylineTemplate(frontMatter) {
|
2147 |
return "\n <div class=\"byline grid\">\n <div>\n <h3>Authors</h3>\n <div>\n ".concat(frontMatter.authors.map(function (author, i) {
|
2148 |
return "\n <span class=\"author\">\n ".concat(author.personalURL ? "\n <a class=\"name\" href=\"".concat(author.personalURL, "\">").concat(author.name) + (i + 1 < frontMatter.authors.length ? "," : "") + "</a>" : "\n <span class=\"name\">".concat(author.name) + (i + 1 < frontMatter.authors.length ? "," : "") + "</span>", "\n </span>\n ");
|
2149 |
-
}).join(''), "\n </div>\n </div>\n <div >\n <h3>Affiliation</h3>\n <div><a href=\"https://huggingface.co/\">Hugging Face</a>\n </div>\n </div>\n <div >\n <h3>Published</h3>\n <div>Feb 19, 2025</div>\n </div>\n </div>\n
|
2150 |
}
|
2151 |
var Byline = /*#__PURE__*/function (_HTMLElement4) {
|
2152 |
function Byline() {
|
|
|
2146 |
function bylineTemplate(frontMatter) {
|
2147 |
return "\n <div class=\"byline grid\">\n <div>\n <h3>Authors</h3>\n <div>\n ".concat(frontMatter.authors.map(function (author, i) {
|
2148 |
return "\n <span class=\"author\">\n ".concat(author.personalURL ? "\n <a class=\"name\" href=\"".concat(author.personalURL, "\">").concat(author.name) + (i + 1 < frontMatter.authors.length ? "," : "") + "</a>" : "\n <span class=\"name\">".concat(author.name) + (i + 1 < frontMatter.authors.length ? "," : "") + "</span>", "\n </span>\n ");
|
2149 |
+
}).join(''), "\n </div>\n </div>\n <div >\n <h3>Affiliation</h3>\n <div><a href=\"https://huggingface.co/\">Hugging Face</a>\n </div>\n </div>\n <div >\n <h3>Published</h3>\n <div>Feb 19, 2025</div>\n </div>\n </div>\n\n");
|
2150 |
}
|
2151 |
var Byline = /*#__PURE__*/function (_HTMLElement4) {
|
2152 |
function Byline() {
|
dist/distill.bundle.js.map
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
dist/index.html
CHANGED
@@ -65,7 +65,10 @@
|
|
65 |
</div>
|
66 |
<p style="text-align: cekter; font-style: italic; margin-top: 10px; max-width: 900px; margin-left: auto; margin-right: auto;">We ran over 4,000 scaling experiments on up to 512 GPUs and measured throughput (size of markers) and GPU utilization (color of markers). Note that both are normalized per model size in this visualization.</p>
|
67 |
|
68 |
-
<
|
|
|
|
|
|
|
69 |
</div>
|
70 |
</d-title>
|
71 |
<d-byline></d-byline>
|
@@ -77,11 +80,7 @@
|
|
77 |
Thousands of GPUs humming in perfect harmony. That's what it takes to train today's most powerful AI models β a symphony of computing power that until recently was the exclusive domain of elite research labs. Open source has transformed this landscape, but not completely. Yes, you can download the latest <a href="https://huggingface.co/meta-llama">Llama</a> or <a href="https://huggingface.co/deepseek-ai">DeepSeek</a> models. Yes, you can read their <a href="https://ai.meta.com/research/publications/the-llama-3-herd-of-models/">technical</a> and <a href="https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf">experiment</a> reports. But the most challenging part β the training code, the knowledge and techniques necessary to coordinate GPUs to train these massive systems β remains shrouded in complexity and spread around in a series of disconnected papers and often private codebases.
|
78 |
</p>
|
79 |
<aside>Reading time: 2-4 days. <br>For the best reading experience, we recommend not using a mobile phone.</aside>
|
80 |
-
|
81 |
-
<button class="order-button" onclick="window.open('https://www.lulu.com/shop/nouamane-tazi-and-ferdinand-mom-and-haojun-zhao-and-phuc-nguyen/the-ultra-scale-playbook/paperback/product-45yk9dj.html?page=1&pageSize=4', '_blank')">
|
82 |
-
Order Book Here
|
83 |
-
</button>
|
84 |
-
</div>
|
85 |
<p>
|
86 |
This open source book is here to change that. Starting from the basics, we'll walk you through the knowledge necessary to scale the training of large language models (LLMs) from one GPU to tens, hundreds, and even thousands of GPUs, illustrating theory with practical code examples and reproducible benchmarks.
|
87 |
</p>
|
|
|
65 |
</div>
|
66 |
<p style="text-align: cekter; font-style: italic; margin-top: 10px; max-width: 900px; margin-left: auto; margin-right: auto;">We ran over 4,000 scaling experiments on up to 512 GPUs and measured throughput (size of markers) and GPU utilization (color of markers). Note that both are normalized per model size in this visualization.</p>
|
67 |
|
68 |
+
<div class="order-button-container">
|
69 |
+
<button class="order-button" style="margin: 0 8px;" onclick="window.open('https://www.lulu.com/shop/nouamane-tazi-and-ferdinand-mom-and-haojun-zhao-and-phuc-nguyen/the-ultra-scale-playbook/paperback/product-45yk9dj.html?page=1&pageSize=4', '_blank')">Order Book</button>
|
70 |
+
<button class="order-button" style="margin: 0 8px;" onclick="window.open('https://huggingface.co/nanotron', '_blank')">Get PDF</button>
|
71 |
+
</div>
|
72 |
</div>
|
73 |
</d-title>
|
74 |
<d-byline></d-byline>
|
|
|
80 |
Thousands of GPUs humming in perfect harmony. That's what it takes to train today's most powerful AI models β a symphony of computing power that until recently was the exclusive domain of elite research labs. Open source has transformed this landscape, but not completely. Yes, you can download the latest <a href="https://huggingface.co/meta-llama">Llama</a> or <a href="https://huggingface.co/deepseek-ai">DeepSeek</a> models. Yes, you can read their <a href="https://ai.meta.com/research/publications/the-llama-3-herd-of-models/">technical</a> and <a href="https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf">experiment</a> reports. But the most challenging part β the training code, the knowledge and techniques necessary to coordinate GPUs to train these massive systems β remains shrouded in complexity and spread around in a series of disconnected papers and often private codebases.
|
81 |
</p>
|
82 |
<aside>Reading time: 2-4 days. <br>For the best reading experience, we recommend not using a mobile phone.</aside>
|
83 |
+
|
|
|
|
|
|
|
|
|
84 |
<p>
|
85 |
This open source book is here to change that. Starting from the basics, we'll walk you through the knowledge necessary to scale the training of large language models (LLMs) from one GPU to tens, hundreds, and even thousands of GPUs, illustrating theory with practical code examples and reproducible benchmarks.
|
86 |
</p>
|
dist/style.css
CHANGED
@@ -688,5 +688,5 @@ select[name="presets"] {
|
|
688 |
.order-button-container {
|
689 |
display: flex;
|
690 |
justify-content: center;
|
691 |
-
margin: 40px 0;
|
692 |
}
|
|
|
688 |
.order-button-container {
|
689 |
display: flex;
|
690 |
justify-content: center;
|
691 |
+
margin: 40px 0 20px 0;
|
692 |
}
|
src/distill.js
CHANGED
@@ -2105,18 +2105,7 @@ d-appendix > distill-appendix {
|
|
2105 |
<div>Feb 19, 2025</div>
|
2106 |
</div>
|
2107 |
</div>
|
2108 |
-
|
2109 |
-
<button class="order-button-second" onclick="window.open('https://www.lulu.com/shop/nouamane-tazi-and-ferdinand-mom-and-haojun-zhao-and-phuc-nguyen/the-ultra-scale-playbook/paperback/product-45yk9dj.html?page=1&pageSize=4', '_blank')">
|
2110 |
-
Book
|
2111 |
-
</button>
|
2112 |
-
</div>
|
2113 |
-
|
2114 |
-
<!-- <div class="side pdf-download">
|
2115 |
-
<a href="https://huggingface.co/spaces/nanotron/ultrascale-playbook/resolve/main/The_Ultra-Scale_Playbook_Training_LLMs_on_GPU_Clusters.pdf">Download PDF
|
2116 |
-
<br>
|
2117 |
-
<img style="width: 32px;" src="../assets/images/256px-PDF.png" alt="PDF"></a>
|
2118 |
-
|
2119 |
-
</div> -->
|
2120 |
`;
|
2121 |
}
|
2122 |
|
|
|
2105 |
<div>Feb 19, 2025</div>
|
2106 |
</div>
|
2107 |
</div>
|
2108 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2109 |
`;
|
2110 |
}
|
2111 |
|
src/index.html
CHANGED
@@ -65,7 +65,10 @@
|
|
65 |
</div>
|
66 |
<p style="text-align: cekter; font-style: italic; margin-top: 10px; max-width: 900px; margin-left: auto; margin-right: auto;">We ran over 4,000 scaling experiments on up to 512 GPUs and measured throughput (size of markers) and GPU utilization (color of markers). Note that both are normalized per model size in this visualization.</p>
|
67 |
|
68 |
-
<
|
|
|
|
|
|
|
69 |
</div>
|
70 |
</d-title>
|
71 |
<d-byline></d-byline>
|
@@ -77,11 +80,7 @@
|
|
77 |
Thousands of GPUs humming in perfect harmony. That's what it takes to train today's most powerful AI models β a symphony of computing power that until recently was the exclusive domain of elite research labs. Open source has transformed this landscape, but not completely. Yes, you can download the latest <a href="https://huggingface.co/meta-llama">Llama</a> or <a href="https://huggingface.co/deepseek-ai">DeepSeek</a> models. Yes, you can read their <a href="https://ai.meta.com/research/publications/the-llama-3-herd-of-models/">technical</a> and <a href="https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf">experiment</a> reports. But the most challenging part β the training code, the knowledge and techniques necessary to coordinate GPUs to train these massive systems β remains shrouded in complexity and spread around in a series of disconnected papers and often private codebases.
|
78 |
</p>
|
79 |
<aside>Reading time: 2-4 days. <br>For the best reading experience, we recommend not using a mobile phone.</aside>
|
80 |
-
|
81 |
-
<button class="order-button" onclick="window.open('https://www.lulu.com/shop/nouamane-tazi-and-ferdinand-mom-and-haojun-zhao-and-phuc-nguyen/the-ultra-scale-playbook/paperback/product-45yk9dj.html?page=1&pageSize=4', '_blank')">
|
82 |
-
Order Book Here
|
83 |
-
</button>
|
84 |
-
</div>
|
85 |
<p>
|
86 |
This open source book is here to change that. Starting from the basics, we'll walk you through the knowledge necessary to scale the training of large language models (LLMs) from one GPU to tens, hundreds, and even thousands of GPUs, illustrating theory with practical code examples and reproducible benchmarks.
|
87 |
</p>
|
|
|
65 |
</div>
|
66 |
<p style="text-align: cekter; font-style: italic; margin-top: 10px; max-width: 900px; margin-left: auto; margin-right: auto;">We ran over 4,000 scaling experiments on up to 512 GPUs and measured throughput (size of markers) and GPU utilization (color of markers). Note that both are normalized per model size in this visualization.</p>
|
67 |
|
68 |
+
<div class="order-button-container">
|
69 |
+
<button class="order-button" style="margin: 0 8px;" onclick="window.open('https://www.lulu.com/shop/nouamane-tazi-and-ferdinand-mom-and-haojun-zhao-and-phuc-nguyen/the-ultra-scale-playbook/paperback/product-45yk9dj.html?page=1&pageSize=4', '_blank')">Order Book</button>
|
70 |
+
<button class="order-button" style="margin: 0 8px;" onclick="window.open('https://huggingface.co/nanotron', '_blank')">Get PDF</button>
|
71 |
+
</div>
|
72 |
</div>
|
73 |
</d-title>
|
74 |
<d-byline></d-byline>
|
|
|
80 |
Thousands of GPUs humming in perfect harmony. That's what it takes to train today's most powerful AI models β a symphony of computing power that until recently was the exclusive domain of elite research labs. Open source has transformed this landscape, but not completely. Yes, you can download the latest <a href="https://huggingface.co/meta-llama">Llama</a> or <a href="https://huggingface.co/deepseek-ai">DeepSeek</a> models. Yes, you can read their <a href="https://ai.meta.com/research/publications/the-llama-3-herd-of-models/">technical</a> and <a href="https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf">experiment</a> reports. But the most challenging part β the training code, the knowledge and techniques necessary to coordinate GPUs to train these massive systems β remains shrouded in complexity and spread around in a series of disconnected papers and often private codebases.
|
81 |
</p>
|
82 |
<aside>Reading time: 2-4 days. <br>For the best reading experience, we recommend not using a mobile phone.</aside>
|
83 |
+
|
|
|
|
|
|
|
|
|
84 |
<p>
|
85 |
This open source book is here to change that. Starting from the basics, we'll walk you through the knowledge necessary to scale the training of large language models (LLMs) from one GPU to tens, hundreds, and even thousands of GPUs, illustrating theory with practical code examples and reproducible benchmarks.
|
86 |
</p>
|
src/style.css
CHANGED
@@ -688,5 +688,5 @@ select[name="presets"] {
|
|
688 |
.order-button-container {
|
689 |
display: flex;
|
690 |
justify-content: center;
|
691 |
-
margin: 40px 0;
|
692 |
}
|
|
|
688 |
.order-button-container {
|
689 |
display: flex;
|
690 |
justify-content: center;
|
691 |
+
margin: 40px 0 20px 0;
|
692 |
}
|