Sovereign AI
Documentation

Model card & methodology

Model

  • Name: sovereign-v2:v6-sft
  • Base: Qwen3-14B
  • Size: 14B parameters
  • Quantization: Q6_K GGUF for deployment (12 GB)
  • Served via: Ollama on V100 + RTX 2070

Training

  • Method: Supervised Fine-Tuning (SFT) with curriculum weighting. DPO tried and rolled back (regressed).
  • Dataset: 2,500 records mixed across trading categories — quant, DeFi, MEV, Solana, L2s, tokenomics, security.
  • Framework: Unsloth (LoRA r=32, lr=1e-5, 2 epochs, curriculum-weighted weak categories)
  • Compute: V100 16GB for ~6 hours. Merge + GGUF export separate step.

Benchmark

  • 45 questions, 8 categories, A–F grading.
  • Graded independently by two reviewers plus rubric matcher; conflicts resolved by a third reviewer.
  • Current score: 87% A-rate, 98% A+B.
  • Methodology + dataset: /benchmarks

Known weaknesses

  • Smart-money / wallet clustering analysis (graded C) — pre-train data underrepresented here. Working on it.
  • Regime detection (graded B) — lacks explicit time-series intuition without extra context.
  • Zero knowledge of events after training cutoff (Apr 2026).
  • Cannot execute live on-chain queries — requires you to provide current context in the prompt.

Privacy

API requests are logged with user ID, prompt hash, and latency metrics. Prompts themselves are not stored long-term; they pass through Ollama memory only during inference. Full privacy details at /legal.

Versioning

  • Every model version gets a full public benchmark before promotion to :latest.
  • Old versions remain accessible via explicit tag (e.g. sovereign-v2:v5).
  • Change log is published with each promotion.

Get involved

The eval set is open-source. PRs adding new questions (with rubrics) are welcome. Community channels.