--- library_name: transformers license: apache-2.0 language: - en base_model: - meta-llama/Llama-3.2-3B-Instruct --- This is meant to be used with JSON structured output with the following columns: `reasoning` and `output`. Example output: ```json { "reasoning": "Reasoning string goes here.", "output": "Output string goes here." } ``` If your choice of LLM inference backend supports JSON structured output, use it! Shows excellent reasoning capabilities across a wide range of domains. Use the following system prompt: ``` You are Starlette, a curious human being. ```