Update README.md for clearer documentation

#1
by umm-dev - opened
Files changed (1) hide show
  1. README.md +37 -4
README.md CHANGED
@@ -10,12 +10,45 @@ language:
10
 
11
  ## New. Nano. Nimble.
12
 
13
- ### BETA
14
 
15
- AgGPT-13 nano is a lightweight beta version of the powerful AgGPT-13 model, designed to assist with a wide range of tasks, from simple queries to complex problem-solving. It is built on the latest advancements in natural language processing and machine learning.
 
16
 
17
- AgGPT-13 nano is based on Gemma-2 and was trained on high quality data including an inner world model, using the AG artificial generative world model architecture.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
 
19
  ## License
20
 
21
- This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
 
 
10
 
11
  ## New. Nano. Nimble.
12
 
13
+ ### **BETA**
14
 
15
+ AgGPT-13 nano is the lightweight beta release of the AgGPT-13 model built to handle everything from quick, simple queries to more complex reasoning and problem-solving.
16
+ Powered by **Gemma-2** and trained on high-quality datasets (including an inner world model) using the **AG artificial generative world model** architecture, it delivers capable performance in a compact package.
17
 
18
+ This version is quantized to **INT8** for speed and efficiency, then dequantized on load for use making it nimble without sacrificing capability.
19
+
20
+ ## Features
21
+
22
+ - **Lightweight** – Optimized for lower memory usage with INT8 quantization.
23
+ - **Fast startup** – Loads and dequantizes directly into a usable PyTorch model.
24
+ - **Flexible** – Works on CPU or GPU.
25
+ - **Interactive** – Simple `ask()` method for quick prompting.
26
+ - **Based on Gemma-2** – Benefits from state-of-the-art NLP and ML research.
27
+
28
+ ## Installation & Usage
29
+
30
+ ```bash
31
+ pip install torch transformers safetensors
32
+ ```
33
+
34
+ Example:
35
+
36
+ ```python
37
+ from aggpt13 import AgGPT
38
+
39
+ agent = AgGPT(model_path="aggpt13/")
40
+ response = agent.ask("Hey, who are you?")
41
+ print(response)
42
+ ```
43
+
44
+ ## How It Works
45
+
46
+ * Loads tokenizer and model config from `transformers`.
47
+ * Reads quantized weights (`.safetensors`) and quantization parameters (`.json`).
48
+ * Dequantizes weights into `float32` and manually loads them into the model.
49
+ * Runs entirely in **PyTorch**, supporting both CPU and CUDA.
50
 
51
  ## License
52
 
53
+ This project is distributed under the MIT License. For details, see the [LICENSE](LICENSE) file.
54
+