PaperToRead Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 17 AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 18
Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 17
AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 18
1bitLLM The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 624
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 624
PaperToRead Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 17 AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 18
Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 17
AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 18
1bitLLM The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 624
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 624