Text Generation
Transformers
Safetensors
English
gpt2
language
aquif
text-generation-inference
math
coding
small

aquif-neo-2-345m-c2

This is the second checkpoint of the 'aquif-neo-2-345m' model, a next-generation language model developed by aquif AI. This checkpoint builds upon the foundation established in c1, with enhanced fine-tuning focused on code generation and programming tasks, representing the second stage in a 5-checkpoint training process designed to create a versatile and capable model.

Model Details

Base Model: gpt2-medium
Method: LoRA (Low-Rank Adaptation)
Parameter Count: 355 million params

Training Information

This checkpoint was trained as the second stage of a multi-checkpoint process, specifically focusing on improving coding capabilities. The training was performed using a network-resilient script that includes fallback mechanisms for data loading and model initialization.

Checkpoint Number: 2/5
Focus Area: Code generation and programming tasks
Hardware: Trained on a Google Colab T4 GPU
Training Duration: Approximately 2.5 hours for this checkpoint
Training Framework: PyTorch, Hugging Face Transformers, PEFT, bitsandbytes, TRL
Quantization: 8-bit

Training Datasets:

  • CodeAlpaca-20k: 4,000 samples for code generation tasks
  • Python Code Instructions: 2,000 samples for Python-specific programming
  • GPT-4.5 100x: Complete dataset for enhanced writing and factual accuracy

LoRA Configuration:

  • r=8
  • lora_alpha=16
  • target_modules: ["q_attn", "c_attn", "c_proj", "c_fc", "attn.c_attn", "attn.c_proj", "mlp.c_fc", "mlp.c_proj"]
  • lora_dropout=0.05
  • bias="none"
  • task_type="CAUSAL_LM"

Training Arguments:

  • per_device_train_batch_size=2
  • gradient_accumulation_steps=16
  • num_train_epochs=1 (for this checkpoint)
  • learning_rate=1e-5
  • max_steps=400

Optimized for 8-bit training.

Training Loss Data

The following table shows the training loss recorded during the training of this checkpoint:

Step Training Loss
20 1.9971
40 1.9395
60 1.9527
80 1.8974
100 1.9324
120 1.8383
140 1.8405
160 1.7731
180 1.7920
200 1.6615
220 1.6679
240 1.6748
260 1.6765
280 1.6969
300 1.6478
320 1.6690
340 1.6214
360 1.5797
380 1.6272
400 1.6243

Note: Training loss is a metric that indicates how well the model is learning. A decreasing loss generally suggests improvement. C2 shows significantly lower loss values compared to C1, indicating improved learning on the code-focused dataset.

Training Focus

Checkpoint 2 emphasizes code-related tasks and programming instruction following. The synthetic dataset for this checkpoint includes:

  • Python function writing and implementation
  • List comprehensions and data structures
  • String manipulation techniques
  • Algorithm explanations and code examples

This targeted approach builds upon the conversational foundation from c1 while developing stronger programming capabilities.

Intended Use

This checkpoint is an intermediate model in the development of the full 'aquif-neo-2'. While showing improved coding capabilities compared to c1, it is not intended for production use but serves as a foundation for subsequent fine-tuning checkpoints focusing on additional domains and tasks.

How to Load the Model

You can load this model using the Hugging Face 'transformers' library:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "aquiffoo/aquif-neo-2-345m-c2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

Future Checkpoints

This is the second of 5 planned checkpoints. The remaining checkpoints will focus on:

  • c3: Mathematical reasoning and problem-solving
  • c4: Knowledge and factual understanding
  • c5: Mixed capabilities integration and refinement

Each checkpoint builds upon the previous ones to create a more well-rounded and capable model.

License: Apache 2.0

Downloads last month
44
Safetensors
Model size
355M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train aquiffoo/aquif-neo-2-345m-c2

Collection including aquiffoo/aquif-neo-2-345m-c2