SymbioticLM-14B

Model Type: Hybrid Symbolic–Transformer with Persistent Memory
Base Model: Qwen-14B
Framework: PyTorch + HuggingFace Transformers
Purpose: Full-scale cognitive reasoning model with self-organizing memory and generative symbolic evolution


Overview

SymbioticLM-14B is a state-of-the-art 17.8 billion parameter symbolic–transformer hybrid model that tightly couples high-capacity neural representation with structured symbolic cognition. Designed to match or exceed performance of top-tier LLMs in symbolic domains, it supports persistent memory, entropic recall, multi-stage symbolic routing, and self-organizing knowledge structures.

This model is ideal for advanced reasoning agents, research assistants, and symbolic math/code generation systems.


Architecture Highlights

  • Backbone: Qwen-14B transformer with rotary embeddings + FlashAttention
  • Symbolic Dim: 8192
  • Symbolic Modules:
    • ThoughtDynamicsLNN (multi-head LSTM attention)
    • LiquidThoughtProcessor
    • CrystallineProcessor (DNAConv GNN)
    • HelicalDNAProcessor (linear helical encoding)
  • Memory: 4096 symbolic states in FP32, retrieved using entropy + contextual similarity
  • Dream Mode: Background symbolic simulation for open-ended cognition
  • Router: Intent classifier + entropy gating for processor path selection

Files Included

File Description
model.bin Transformer weights (LFS)
model.safetensors Memory-safe weights, optimized for loading
memory.pt 4096-symbolic vector bank
config.json Model and architectural metadata
generation_config.json Top-p, temperature, decoding settings
tokenizer.json Full tokenizer with symbolic tag support
added_tokens.json Tags like <D_LIM>, <PROOF>, <BY_MEASURE>, etc.
special_tokens_map.json Special token mapping for tokenizer

Intended Uses

  • Multi-step conversational agents with true memory
  • Long-form symbolic theorem generation and proof planning
  • Scientific dialogue, symbolic simulations, math/code synthesis
  • Reasoning in fuzzy, discontinuous, or non-smooth problem domains

Limitations

  • Memory requires curation and seeding for maximum utility
  • Symbolic cognition is not instruction-tuned for general QA
  • FlashAttention and symbolic modules increase VRAM usage during generation

Citations

Please cite "SymbioticLM" when using symbolic memory components in research or applications.

Downloads last month
33
Safetensors
Model size
14.8B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for reaperdoesntknow/Symiotic-14B

Finetuned
Qwen/Qwen3-14B
Finetuned
(19)
this model
Quantizations
2 models

Dataset used to train reaperdoesntknow/Symiotic-14B

Collection including reaperdoesntknow/Symiotic-14B