File size: 3,255 Bytes
9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c 9555e49 95b472c | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 | ---
language: en
license: cc-by-nc-4.0
tags:
- zetagrid
- cpu-da
- tcn
- fractal
- 25b
- nlp
- text-generation
model_name: RTH-LM (25B)
base_model: RthItalia/seed-7b
parameters: 25B
datasets:
- custom
metrics:
- loss
---
# π Model Card: RTH-LM (25B)
## Model Details
- **Name:** RTH-LM (25B)
- **Architecture:** Fractal Gated Causal TCN (Temporal Convolutional Network)
- **Parameters:** 7B (Physical) / 25B (Effective Fractal Capacity)
- **Author:** Christian Quintino De Luca (RTH Italia)
- **Release Date:** February 2026
- **License:** CC BY-NC 4.0 (Research) / Commercial (Enterprise)
- **Paper (Figshare):** [https://doi.org/10.6084/m9.figshare.31376560](https://doi.org/10.6084/m9.figshare.31376560)
RTH-LM (25B) is a **Fractal TCN (Temporal Convolutional Network)** Language Model, designed for high-efficiency inference on CPU/Consumer Hardware and massive scalability on GPUs.
Unlike Traditional Transformers, ZetaGrid uses a **Gated Causal TCN backbone** with **Fractal Scaling**, allowing it to model long-range dependencies with significantly lower memory overhead during inference.
---
## π Model Specs
| Feature | Specification |
| :--- | :--- |
| **Parameters** | 25 Billion (25B) |
| **Architecture** | Fractal Gated TCN (Non-Transformer) |
| **Layers** | 32 (Phase 2) |
| **Context Window** | 256 - 1024 (Fractal Expansion Capable) |
| **Training Data** | 1.48 GB Cleaned Text (Wiki/Books) |
| **Final Loss** | **1.0675** (Phase 2) |
| **Quantization** | QULP 2-bit (Supported) |
---
## π Usage (Inference)
### Prerequisites
You need the `cpu_da` framework or the Python inference script.
```bash
# Clone the repo
git clone https://github.com/rth-italia/cpu-da
cd cpu-da
```
### Running the Model (Python)
Ensure you have `zeta25b_step15000.pt` (Weights) and `zetagrid_25b_production.npy` (Genome).
```python
import torch
from ZETAGRID_INFERENCE import load_model, generate
# Load 25B Model
model = load_model("zeta25b_step15000.pt", genome="zetagrid_25b_production.npy")
# Generate
text = generate(model, "The future of AI is")
print(text)
```
### QULP 2-bit Inference (Ultra-Low Memory)
To run on consumer CPUs with <2GB RAM:
```bash
python QULP_INFERENCE.py --model zeta25b_2bit.qulp
```
---
## 𧬠Architecture: The "Fractal Soul"
ZetaGrid is **NOT** a Transformer. It is a TCN-based organism.
- **Genome:** A fixed 7GB "DNA" bank of weights (`zetagrid_25b_production.npy`).
- **Phenotype:** The model layers are "grown" from this genome on the fly.
- **Training:** Only the "Soul" (LoRA Adapters + Norms) is trained (~300MB), making the model extremely portable.
- **Fractal Scaling:** The 25B model can be fractally expanded to 50B, 100B+ by duplicating layers and adding self-linear noise.
---
## π Performance
- **Phase 1 (Evolution):** 200 Generations of Genome Optimization.
- **Phase 2 (Gradient):** 15,000 Steps of TCN+LoRA Fine-Tuning.
- **Convergence:** Beat target loss of 1.5, achieving **1.0675**.
- **Capabilities:** Narrative coherence, English syntax mastery, abstract reasoning.
---
## π License
CC BY-NC 4.0 (Creative Commons Non-Commercial) for Research.
**Commercial Use:** Requires a license from **RTH Italia** (Cpu-DA Project).
For inquiries: licensing@rth-italia.com
|