Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ datasets:
|
|
8 |
|
9 |
# KernelLLM
|
10 |

|
11 |
-
|
12 |
## Making Kernel Development more accessible with KernelLLM
|
13 |
|
14 |
We introduce KernelLLM, a large language model based on Llama 3.1, which has been trained specifically for the task of authoring GPU kernels using Triton. KernelLLM translates PyTorch modules into Triton kernels and was evaluated on KernelBench-Triton (see [here](https://github.com/ScalingIntelligence/KernelBench/pull/35)).
|
|
|
8 |
|
9 |
# KernelLLM
|
10 |

|
11 |
+
On KernelBench-Triton Level 1, our 8B parameter model matches GPT-4o in single-shot performance. With multiple inferences, KernelLLM's performance matches DeepSeek R1. This is all from a model with two orders of magnitude fewer parameters than its competitors.
|
12 |
## Making Kernel Development more accessible with KernelLLM
|
13 |
|
14 |
We introduce KernelLLM, a large language model based on Llama 3.1, which has been trained specifically for the task of authoring GPU kernels using Triton. KernelLLM translates PyTorch modules into Triton kernels and was evaluated on KernelBench-Triton (see [here](https://github.com/ScalingIntelligence/KernelBench/pull/35)).
|