MLX
English
finetune
Eval Results
vishal-adithya's picture
Update README.md
a75c6cc verified
metadata
license: apache-2.0
datasets:
  - vishal-adithya/texthumanizer-preprocessed-data
language:
  - en
metrics:
  - accuracy
base_model:
  - mistralai/Ministral-8B-Instruct-2410
library_name: mlx
tags:
  - mlx
  - finetune
model-index:
  - name: Th-4bit
    results:
      - task:
          type: text-generation
        dataset:
          name: train
          type: train
        metrics:
          - name: Training Loss
            type: loss
            value: 0.171
      - task:
          type: text-generation
        dataset:
          name: validation
          type: validation
        metrics:
          - name: Validation Loss
            type: loss
            value: 0.175
      - task:
          type: text-generation
        metrics:
          - name: Learning Rate
            type: lr
            value: 0.00001
      - task:
          type: text-generation
        metrics:
          - name: Tokens per sec
            type: t/sec
            value: 117.298

TextHumanizer - Fine-tuned Ministral-8B-Instruct:

TextHumanizer is a fine-tuned version of the Ministral-8B-Instruct model. It is designed to transform robotic, overly-formal, or synthetic AI-generated text into fluent, natural, and human-like language.

This model was fine-tuned using Apple's MLX framework on a custom dataset of AI-generated text paired with humanized rewrites.


Model Details:

  • Base model: mistralai/Ministral-8B-Instruct
  • Model size: 8B parameters
  • Fine-tuned using: MLX (Apple Silicon-native training)
  • Fine-tuning method: QLora
  • Precision: float16
  • Hardware used: Apple Silicon M1 Macbook Pro
  • Training time: 10 mins
  • Epochs / Steps: 200
  • Batch size: 4

Training Metrics:

Metric Value
Final Training Loss 0.171
Final Validation Loss 0.175

Dataset:

  • Source: Custom synthetic-to-human text pairs
  • Size: 10k rows
  • Structure: Pairs of (synthetic_input, humanized_output)
  • Preprocessing: Standard MLX tokenization, formatting for instruct tuning
  • License: Public

Capabilities:

TextHumanizer is designed to:

  • Improve fluency and tone of AI outputs
  • Make answers sound more relatable and natural
  • Polish robotic or overly formal language

MLX (Apple Silicon):

from mlx_lm import load, generate

model, tokenizer = load("vishal-adithya/ministral-8B-texthumanizer")
prompt = "Instruction: Make this sound more natural.Input: The individual proceeded to consume nourishment.Response:"
generate(model, tokenizer, prompt)