AQUA-1B

Model Information

AQUA-1B is Kurma AIโ€™s compact and efficient 1-billion parameter Small Language Model (SLM), It is the first lightweight Aquaculture domain-specific model purpose-built for real-time aquaculture operations involving IoT sensor data, autonomous systems, and robotic agents.

Designed for edge deployments and low-latency environments, AQUA-1B enables on-device decision-making, real-time alert generation, and agentic task execution. It powers intelligent aquaculture systems for Water quality monitoring, Automated feeding routines, Mobile robotic inspections across ponds, tanks, and recirculating aquaculture systems (RAS)

Learn more about Kurma AI.


Key Features

  • Edge-Ready Intelligence Optimized for low-power, real-time inference on embedded devices like Raspberry Pi, Jetson Nano, and Coral TPU.
  • Agentic Task Execution Supports multi-step agent-based workflows such as sensor checks, feeding triggers, water exchange scheduling, and autonomous health checks using instruction-following prompts.
  • IoT-Aware Reasoning Natively understands and reasons over sensor data inputs (e.g., temperature, pH, TDS, turbidity, DO), enabling rapid decisions in fluctuating aquaculture environments.
  • Robotic Automation Control Designed to interact with robotic systems, including underwater, and mobile pond inspectors.
  • Autonomous Alerting Systems Powers local alert mechanisms (via SMS, Telegram bots, or MQTT) that notify farmers when water parameters exceed safe thresholds or when interventions are required.
  • Field-Deployable Decision Engine Enables fully autonomous operation in remote hatcheries and ponds, ensuring uninterrupted control even in offline conditions or low-connectivity zones.

Training Data Highlights

  • Extension workerโ€“farmer dialogues and field advisory logs
  • FAO, ICAR, NOAA, and peer-reviewed aquaculture research
  • Synthetic Q&A from 5,000+ aquaculture-focused topics
  • Climate-resilient practices, hatchery SOPs, and water quality datasets
  • Carefully curated to support species-specific culture methods
  • Scale: Trained on approximately 3 million real and synthetic Q&A pairs, totaling around 1 billion tokens of high-quality, domain-specific data.

Model Specifications

  • Base Model: Gemma 3 1B (by Google DeepMind)
  • Training Tokens: ~1 Billion
  • Released On 4, July 2025
  • Data Volume: 3M+ expert-verified and synthetic instructions
  • Origin: Made in America by Kurma AI
  • Training Technic Model is trained via Fine-tuning using (LoRA-based) Supervised Fine-Tuning (SFT).
  • Training Infrastructure: Trained using 8 NVIDIA H200 GPU Multi Cluster Special Thanks to Nebius

Quickstart

Transformers (Google Colab/ jupyter)

  • Install dependencies
!pip install transformers accelerate
  • Log in with your Hugging Face access token
from huggingface_hub import login
  • Import model from Huggingface
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "KurmaAI/AQUA-1B"

tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    device_map="auto",             # Automatically uses GPU if available
    torch_dtype=torch.float16,     # Use torch.float32 if no GPU
    trust_remote_code=True
)
  • Test Prompt
prompt = "What are the most common diseases in shrimp farming and how can they be prevented?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256)

response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

๐Ÿ™ Acknowledgements

This project was made possible thanks to:

  • Nebius for providing a compute grant and access to NVIDIA H200 GPU servers, which powered the model training process.
  • Google DeepMind for sharing their open-source language models, which made this project possible.
  • Kurma AI research team: including aquaculture experts, machine learning engineers, data annotators, and advisors who collaborated to curate, verify, and refine the domain-specific dataset used for fine-tuning this model.

โš ๏ธ Disclaimer, Bias & Limitations

  • Domain Bias: The model may reflect inherent biases present in the aquaculture data sources and industry practices on which it was trained.
  • Temporal Data Limitation: Climate and environmental recommendations are based on information available up to 2024. Users should cross-check any climate-related advice against the latest advisories (e.g., IMD or NOAA updates).
  • Potential Hallucinations: Like all large language models, Aqua-1B may occasionally generate inaccurate or misleading responses ("hallucinations").
  • Always validate critical, regulatory, or high-impact decisions with a qualified aquaculture professional.
Downloads last month
135
Safetensors
Model size
1,000M params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for KurmaAI/AQUA-1B

Finetuned
(190)
this model
Quantizations
1 model