Collection of fine-tuned bilingual language models featuring: - Models in three parameter sizes: 135M, 360M, and 1.7B based on HuggingFaceTB's SmolLM2 models - Both standard and GGUF formats for flexible deployment in llama.cpp and Ollama - Fine-tuned on nyuuzyou/EagleSFT dataset (536,231 Russian-English QA pairs derived from 739k+ real user queries) - Experimental Russian language capabilities while maintaining English performance - Limited Russian capabilities due to SFT-only approach without Russian pre-training - Environmental impact: ~19.75 kg CO2eq
This collection provides compact models for research on bilingual language capabilities, resource-constrained environments, and educational applications. Not recommended for production use due to experimental nature and inherent limitations. Available under Apache 2.0 license.