PlayWeBit (BitNet b1.58-2B-4T)
This is an implementation of the microsoft/bitnet-b1.58-2B-4T
model, optimized for text generation. It supports PyTorch and TensorFlow and can run on CPU or GPU. A custom model class (custom_bitnet.py
) is included for compatibility with the official Hugging Face transformers
library.
Model Description
- Architecture: BitNet b1.58-2B with 4-bit quantization
- Parameters: 2 billion
- Training Data: Not disclosed (refer to original Microsoft repository)
- Intended Use: Text generation, chat applications
Installation
Install the Hugging Face transformers
library:
pip install transformers
- Downloads last month
- 34
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support