LWM-Text-Chat-128K Model Card

Model details

Model type: LWM-Text-Chat-128K is an open-source model trained from LLaMA-2 on a subset of Books3 filtered data. It is an auto-regressive language model, based on the transformer architecture.

Model date: LWM-Text-Chat-128K was trained in December 2023.

Paper or resources for more information: https://largeworldmodel.github.io/

License

Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.

Where to send questions or comments about the model: https://github.com/LargeWorldModel/lwm/issues

Training dataset

  • 92K subset of Books3 documents with 100K to 200K tokens
Downloads last month
17
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.