Zacharias030 commited on
Commit
428cf76
·
verified ·
1 Parent(s): 19545c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -8,7 +8,7 @@ datasets:
8
 
9
  # KernelLLM
10
  ![scatter performance comparison plot](media/llm_performance_comparison.png)
11
- Caption: On KernelBench-Triton Level 1, our 8B parameter model matches GPT-4o in single-shot performance. With multiple inferences, KernelLLM's performance matches DeepSeek R1. This is all from a model with two orders of magnitude fewer parameters than its competitors.
12
  ## Making Kernel Development more accessible with KernelLLM
13
 
14
  We introduce KernelLLM, a large language model based on Llama 3.1, which has been trained specifically for the task of authoring GPU kernels using Triton. KernelLLM translates PyTorch modules into Triton kernels and was evaluated on KernelBench-Triton (see [here](https://github.com/ScalingIntelligence/KernelBench/pull/35)).
 
8
 
9
  # KernelLLM
10
  ![scatter performance comparison plot](media/llm_performance_comparison.png)
11
+ On KernelBench-Triton Level 1, our 8B parameter model matches GPT-4o in single-shot performance. With multiple inferences, KernelLLM's performance matches DeepSeek R1. This is all from a model with two orders of magnitude fewer parameters than its competitors.
12
  ## Making Kernel Development more accessible with KernelLLM
13
 
14
  We introduce KernelLLM, a large language model based on Llama 3.1, which has been trained specifically for the task of authoring GPU kernels using Triton. KernelLLM translates PyTorch modules into Triton kernels and was evaluated on KernelBench-Triton (see [here](https://github.com/ScalingIntelligence/KernelBench/pull/35)).