--- license: apache-2.0 datasets: - prithivMLmods/Song-Catalogue-Long-Thought language: - en base_model: - meta-llama/Llama-3.2-3B-Instruct pipeline_tag: text-generation library_name: transformers tags: - safetensors - Llama3.2 - 3B - Extended-Stream - text-generation-inference - Instruct --- ### **Llama-Song-Stream-3B-Instruct Model Card** The **Llama-Song-Stream-3B-Instruct** is a fine-tuned language model specializing in generating music-related text, such as song lyrics, compositions, and musical thoughts. Built upon the **meta-llama/Llama-3.2-3B-Instruct** base, it has been trained with a custom dataset focused on song lyrics and music compositions to produce context-aware, creative, and stylized music output. | **File Name** | **Size** | **Description** | |---------------------------------|------------|-------------------------------------------------| | `.gitattributes` | 1.57 kB | LFS tracking file to manage large model files. | | `README.md` | 282 Bytes | Documentation with model details and usage. | | `config.json` | 1.03 kB | Model configuration settings. | | `generation_config.json` | 248 Bytes | Generation parameters like max sequence length. | | `pytorch_model-00001-of-00002.bin` | 4.97 GB | Primary weights (part 1 of 2). | | `pytorch_model-00002-of-00002.bin` | 1.46 GB | Primary weights (part 2 of 2). | | `pytorch_model.bin.index.json` | 21.2 kB | Index file mapping the checkpoint layers. | | `special_tokens_map.json` | 477 Bytes | Defines special tokens for tokenization. | | `tokenizer.json` | 17.2 MB | Tokenizer data for text generation. | | `tokenizer_config.json` | 57.4 kB | Configuration settings for tokenization. | ### **Key Features** 1. **Song Generation:** - Generates full song lyrics based on user input, maintaining rhyme, meter, and thematic consistency. 2. **Music Context Understanding:** - Trained on lyrics and song patterns to mimic and generate song-like content. 3. **Fine-tuned Creativity:** - Fine-tuned using *Song-Catalogue-Long-Thought* for coherent lyric generation over extended prompts. 4. **Interactive Text Generation:** - Designed for use cases like generating lyrical ideas, creating drafts for songwriters, or exploring themes musically. --- ### **Training Details** - **Base Model:** [meta-llama/Llama-3.2-3B-Instruct](#) - **Finetuning Dataset:** [prithivMLmods/Song-Catalogue-Long-Thought](#) - This dataset comprises 57.7k examples of lyrical patterns, song fragments, and themes. --- ### **Applications** 1. **Songwriting AI Tools:** - Generate lyrics for genres like pop, rock, rap, classical, and others. 2. **Creative Writing Assistance:** - Assist songwriters by suggesting lyric variations and song drafts. 3. **Storytelling via Music:** - Create song narratives using custom themes and moods. 4. **Entertainment AI Integration:** - Build virtual musicians or interactive lyric-based content generators. --- ### **Example Usage** #### **Setup** First, load the Llama-Song-Stream model: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "prithivMLmods/Llama-Song-Stream-3B-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) ``` --- #### **Generate Lyrics Example** ```python prompt = "Write a song about freedom and the open sky" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_length=100, temperature=0.7, num_return_sequences=1) generated_lyrics = tokenizer.decode(outputs[0], skip_special_tokens=True) print(generated_lyrics) ``` --- ### **Deployment Notes** 1. **Serverless vs. Dedicated Endpoints:** The model currently does not have enough usage for a serverless endpoint. Options include: - **Dedicated inference endpoints** for faster responses. - **Custom integrations via Hugging Face inference tools.** 2. **Resource Requirements:** Ensure sufficient GPU memory and compute for large PyTorch model weights. ---