You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Koyna-V2-1b-instruct

A fine-tuned version of gemma-3-1b-it trained on OCR-scanned BSCE Agriculture textbooks. It is a bilingual model fluent in Marathi and English, designed for agriculture domain tasks, including answering syllabus-based questions, general real-world farming conversations.


๐Ÿง  Model Details

๐Ÿ“ Description

  • Model Name: Koyna-V2-1b-instruct
  • Base Model: google/gemma-3-1b-it
  • Architecture: Gemma 3B Instruction-tuned
  • Fine-tuned by: Govind Barbade
  • Languages: Marathi (mr), English (en)
  • License: apache-2.0
  • Use Case: Conversational, QA, and instruction-following for farming/agriculture education

๐Ÿ“ฆ Model Sources


๐Ÿ’ฌ Uses

โœ… Direct Use

  • Answering questions from BSCE Agriculture syllabus
  • Conversational agent in Marathi + English
  • Educational assistant for rural/agri tech

๐Ÿšซ Out-of-Scope Use

  • Medical, legal, or critical decision-making
  • Bias-free or politically sensitive generation without supervision

โš ๏ธ Bias, Risks, and Limitations

  • Trained on scanned OCR text; may contain noise or formatting errors
  • May reflect biases present in the original academic materials
  • Not tested on adversarial queries

๐Ÿ”Ž Recommendations

Use in supervised educational or non-critical contexts. Validate outputs before use in production/agricultural planning.


๐Ÿš€ Getting Started

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("Govind222/Koyna-V2-1b-instruct")
tokenizer = AutoTokenizer.from_pretrained("Govind222/Koyna-V2-1b-instruct")

inputs = tokenizer("เคฎเคพเคเฅเคฏเคพ เคŠเคธ เคชเคฟเค•เคพเคธเคพเค เฅ€ เค•เฅ‹เคฃเคคเฅ‡ เค–เคค เค‰เคชเคฏเฅเค•เฅเคค เค†เคนเฅ‡?", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0]))

๐Ÿงช Training Details

๐Ÿ“š Dataset

  • Manually collected and OCR-scanned BSCE Agriculture textbooks (Marathi)
  • Chapters include: Agronomy, Soil Science, Horticulture, Entomology, Plant Pathology

๐Ÿ“Š Evaluation

Model was evaluated qualitatively on syllabus-based QA and conversational prompts. Further benchmarking in progress.

๐Ÿงฐ Technical Specs

  • Model Type: Causal LM (Decoder only)
  • Base Architecture: Gemma 1B Instruction-tuned
  • Quantized Versions: GGUF available in Q2_K, Q4_K_M, Q8_0, F16, etc.

๐Ÿ™ Acknowledgements

Thanks to: Google for the base model My team and resources at home for enabling this project

๐Ÿ“ซ Contact

Author: Govind Barbade Email: [email protected] Hugging Face Profile: https://huggingface.co/Govind222

Downloads last month
89
Safetensors
Model size
1,000M params
Tensor type
FP16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Govind222/Koyna-V2-1b-instruct

Finetuned
(140)
this model
Quantizations
2 models