Text Generation
GGUF
Finnish
llama.cpp
finnish
llama
conversational

Improve model card: Add GGUF usage, paper link, and correct metadata

#1
by nielsr HF Staff - opened

This PR improves the model card for the QuantFactory/Ahma-3B-GGUF model by:

  • Adding library_name: llama.cpp to the metadata, which is appropriate for a GGUF artifact and enables the relevant "how to use" section.
  • Removing the incorrect inference: false metadata tag.
  • Adding gguf to the model tags for better discoverability.
  • Prominently linking to the paper "Scaling Data-Constrained Language Models", which influenced the training of the base Ahma-3B model.
  • Providing a clear usage example for the GGUF model using llama-cpp-python.
  • Clarifying the existing transformers usage section to indicate it is for the original (non-GGUF) Finnish-NLP/Ahma-3B model.
  • Adding a direct link to the llama.cpp GitHub repository in the introduction.

These changes help users understand how to best utilize this specific GGUF model and provides clearer context for its development.

munish0838 changed pull request status to merged

Sign up or log in to comment