What We Built
A system that gets smarter every night while we sleep. This week it proved it.
The SBPI Semantic Layer tracks competitive brand power across the micro-drama vertical — a $14B market (Deloitte projects $7.8B in in-app revenue alone for 2026). W12-2026 marks the system's inflection point: DramaBox overtook ReelShort for #1, Google entered the category, 4 new companies crossed the materiality threshold, and 3 weeks of RDF data unlocked all SPARQL insight queries for the first time.
The Pipeline
Two data flows feed the system. Both run without human intervention after setup.
Ecosystem Graph Flow
Headless Claude agents read session records, extract entity relationship statements, and post them to the InfraNodus knowledge graph. 27 sessions processed in 16 minutes. 96% success rate. The latest competitive intel graph (microco-competitive-W12-2026) holds 96 nodes, 268 edges, 8 clusters at 0.724 modularity.
Semantic Layer Flow
Weekly competitive intelligence data converts to RDF triples, validates against SHACL shapes, loads into an Oxigraph SPARQL endpoint, and gets queried by the nightly insights runner. W12-2026 ingestion: 2,588 triples (up from 1,672), SHACL-validated, 10 weekly movers detected, 22 dimension anomalies surfaced, 9 predictive signals generated. 7 query types. 11 SPARQL files. Scheduled daily at 6:13 AM.
OTK Lineage
This architecture has a name. In 2001, a European research consortium built OTK — the Ontology-based Knowledge Management Toolkit. Their pipeline: Extract → Structure → Store → Query → Present. Our pipeline in 2026 uses the same architecture with different tools. The difference that matters: AI agents perform extraction, and the extraction improves with each cycle because the ontology constrains what counts as a valid fact.
Production Stack
| Component | Tool | Purpose |
|---|---|---|
| Ontology | OWL 2 (Turtle) | Domain model: 30+ classes, 50+ properties |
| Validation | SHACL (pySHACL) | Schema enforcement on every data load |
| Triple Store | Oxigraph | SPARQL endpoint (RocksDB backend) |
| ETL | Python (rdflib) | JSON → RDF conversion + store loading |
| Queries | SPARQL 1.1 | 11 query files, 8 analysis patterns |
| Predictions | Python | Momentum + anomaly detection + confidence |
| Scheduling | launchd | Daily nightly insights at 6:13 AM |
| Knowledge Graph | InfraNodus | Entity extraction, clusters, gaps |
| Batch Processing | claude -p | Headless session → entity extraction |
| Deployment | Cloudflare Pages | Live dashboards and editorial sites |
The Ontology as IP
The SBPI ontology (sbpi.ttl) is not a generic schema. It is a domain-specific model of how competitive brand power works in entertainment verticals. The dimension weights encode expert judgment that took months of client engagements to calibrate.
Five Scoring Dimensions
| Dimension | Code | Weight | What It Measures |
|---|---|---|---|
| Content Strength | cs | 20% | Volume, quality, and exclusivity of content produced |
| Narrative Ownership | no | 20% | Control over press coverage, thought leadership, recognition |
| Distribution Power | dp | 25% | App store rankings, global availability, partnerships |
| Community Strength | cm | 20% | Size, engagement intensity, and loyalty of audience |
| Monetization Infrastructure | mi | 15% | Revenue generation: payments, ads, subscriptions, coin systems |
Distribution Power carries the highest weight (25%) because in a mobile-first vertical, app store presence and partnership reach determine whether content reaches audiences. Monetization Infrastructure carries the lowest (15%) because several dominant players subsidize the vertical from existing revenue streams.
Performance Tiers
Dominant (85-100)
Market leaders with category-defining presence. Currently: no companies in micro-drama reach this tier, but DramaBox (82.75) and ReelShort (82.0) are approaching.
Strong (70-84)
Established competitors with sustainable advantages. DramaBox (82.75), ReelShort (82.0), Disney (76.55). W12 milestone: DramaBox overtakes ReelShort for #1.
Emerging (55-69)
Growing players gaining structural power. CandyJar (65.85), Google/100 Zeros (63.65), JioHotstar (62.25), Holywater (61.65), Netflix (60.8), ShortMax (56.65).
Niche (40-54)
Specialized players with narrow but defensible positions.
Limited (<40)
Pre-launch, resource-constrained, or strategically irrelevant.
Validation: SHACL Shapes
Every data load runs through SHACL shape validation. Bad data fails before entering the store. This is schema enforcement, not error handling.
Enforced Constraints
| Shape | Validates | Constraints |
|---|---|---|
| CompanyShape | Company instances | Exactly 1 name, slug matches ^[a-z][a-z0-9-]*$, isPlatformGiant required |
| ScoreRecordShape | Score records | Exactly 1 company + week, composite 0-100, exactly 5 dimension scores |
| DimensionScoreShape | Dimension scores | Exactly 1 dimension, value 0-100 integer |
| WeekShape | Week instances | Label matches ^W[0-9]{1,2}-[0-9]{4}$ |
| SignalShape | Market signals | signalText required |
| AttestationShape | Provenance records | Confidence 0.0-1.0, sourceType required |
Attestation Layer
Every score carries provenance metadata tracking confidence and source quality:
Attestation:
confidence: 0.85 # 0.0-1.0 scale
sourceType: "expert_judgment"
# Options: primary_data, secondary_analysis, expert_judgment, automated_inference
The attestation upgrade engine progressively improves scores as evidence quality increases. Signal URLs upgrade confidence from 0.85 to 0.90. Scoring rationales add a second attestation at 0.95. The system tracks not just what it knows, but how well it knows it.
The Karpathy Auto-Research Pattern
Andrej Karpathy described the future of AI as "auto-research" — headless AI agents that continuously process information, extract structured knowledge, and produce insights without human prompting. We implemented this literally.
Headless Batch Processing
The batch-mapupdate.sh script uses claude -p (Claude's headless CLI mode) to process session records. Each session takes ~20 seconds:
1. claude -p reads session file + extracts entity statements (text only) 2. curl POSTs entities to InfraNodus REST API 3. curl GETs full graph stats (nodes, edges, clusters, gaps) 4. Shell appends MOC entry with parsed results 5. Dashboard regenerated and deployed to Cloudflare Pages
27 sessions processed in a single batch run. 26 succeeded, 1 failed (entity extraction produced raw output instead of parsed entities). 96% success rate. No human interaction required.
Nightly Insights Runner
The nightly-insights.py scheduler runs 7 SPARQL queries daily at 6:13 AM:
| Query | Schedule | What It Detects |
|---|---|---|
| Weekly Movers | Nightly | Biggest week-over-week delta changes |
| Dimension Anomalies | Nightly | Dimension-composite gaps >20 points |
| Predictive Signals | Nightly | Momentum patterns predicting next-week movements |
| Tier Transitions | Weekly | Companies crossing tier boundaries |
| Distribution-Community Gap | Weekly | High distribution but low community |
| Attestation Coverage | Weekly | Source backing quality per score record |
| Platform vs. Pure-Play | Weekly | Structural scoring differences by company type |
W12-2026 Results & W13 Forecast
W12 delivered the system's first complete insight cycle: 10 weekly movers, 22 dimension anomalies, and 9 predictive signals — all three query types returning data for the first time since the semantic layer went live.
W12 Actual Movers (Top 10 by delta magnitude)
| Company | Composite | Delta | Tier | Key Signal |
|---|---|---|---|---|
| DramaBox | 82.75 | +4.0 | Strong | New #1. $500M valuation, Disney Accelerator, only profitable pure-play ($10M net) |
| JioHotstar | 62.25 | +3.95 | Emerging | IPL launch imminent, 100 microdramas planned, 300M subscriber base |
| COL Group / BeLive | 44.55 | +3.15 | Niche | FILMART 2026 execution, 1,700 titles live, SaaS model provable |
| KLIP | 22.35 | -2.65 | Limited | Structural squeeze — JioHotstar's 300M subscribers dwarfs KLIP's reach |
| Amazon | 50.2 | -2.6 | Niche | Downgraded T2→T3. Only major platform giant with zero microdrama strategy |
| Disney | 76.55 | +2.3 | Strong | Locker Diaries #1, vertical feed for Disney+, DramaBox Accelerator investment |
| ReelShort | 82.0 | -2.05 | Strong | Head of Production defection, confirmed loss-making at $400M revenue |
| Netflix | 60.8 | -2.0 | Emerging | Mobile redesign = acknowledgment, not commitment. Zero production activity |
| Viu | 48.15 | -1.85 | Niche | DramaBox SE Asia expansion pressures Viu's core territory |
| Mansa | 19.35 | +1.85 | Limited | First original content, Africa-first positioning |
W13 Predictive Signals
| Company | Direction | Momentum | Confidence | Signal |
|---|---|---|---|---|
| JioHotstar | ▲ BULLISH | +9.45 | 85% | 2-week acceleration. IPL live on March 26 — execution converts planning to production. |
| COL Group / BeLive | ▲ BULLISH | +7.25 | 75% | Post-FILMART momentum. SaaS clients now signing. |
| Amazon | ▼ BEARISH | -5.8 | 75% | 2-week decline accelerating. Now the only major platform with zero strategy. |
| Disney | ▲ BULLISH | +5.55 | 75% | Triple play: own content + invest in pure-plays + platform integration. |
| DramaBox | ▲ BULLISH | +5.25 | 70% | New #1 with $100M raise at $500M. Momentum accelerating from +1.25 to +4.0. |
| Netflix | ▼ BEARISH | -5.0 | 65% | Gap between awareness and action widening. Mobile engagement erosion documented. |
| GoodShort | ▲ BULLISH | +4.5 | 65% | $220M revenue, $160-200K per series. Category's most capital-efficient producer. |
| Lifetime / A+E | ▲ BULLISH | +4.15 | 65% | Taye Diggs partnership. Vertical as IP extension layer — clearest legacy TV signal. |
| ReelShort | ▼ BEARISH | -2.6 | 60% | Talent leaving, profitability question intensifying. Crown slipping to DramaBox. |
W12 unlocked the full SPARQL insight pipeline. Weekly Movers required 2+ weeks of deltas. Dimension Anomalies required composite scores with dimension breakdowns. Predictive Signals required 2+ consecutive weeks with momentum. With W10+W11+W12 data in the store (2,588 triples), all three produce actionable intelligence. The headline: DramaBox (82.75) overtakes ReelShort (82.0) for #1, and the momentum signals predicted it — DramaBox has been BULLISH for 3 consecutive weeks.
W12 New Entrants
Google / 100 Zeros 63.65
Google TV dedicated microdrama offering. Bachelor creator Mike Fleiss, McG, Simon Fuller on slate. First window on Google TV, then multi-platform. The biggest structural shift of the week.
Holywater / My Drama 61.65
$22M Series A. Fox Entertainment equity stake + 200 titles over 2 years. Dhar Mann deal. 55M downloads. Revenue tripled in 2025. Most momentum of any challenger.
ShortMax 56.65
100M downloads, 3,888% YoY growth. Now #3 entertainment on Google Play. Promoted from watchlist. Eroding ReelShort's app store position.
GammaTime 46.15
$14M seed from Alexis Ohanian, Kim Kardashian, Kris Jenner. Sandra Yee Ling (ex-ReelShort Head of Production) hired. CSI creator writing originals.
Confidence Scoring
Every prediction carries a confidence score computed from a documented formula. No black box.
base = 0.60 # minimum: 2 consecutive weeks, same direction Adjustments: +0.10 if avg |delta| > 3.0 # strong movement +0.10 if avg |delta| > 5.0 # very strong (cumulative with above) +0.05 if both |delta| > 2.0 # consistent magnitude Cap: 0.95 # never assert certainty from 2 weeks of data
Dimension Anomalies (Hidden Strengths & Vulnerabilities)
22 anomalies detected in W12 — cases where a single dimension score diverges significantly from the composite. These expose structural strengths masked by overall ranking or vulnerabilities hidden by headline scores.
| Company | Dimension | Dim Score | Composite | Gap | Type |
|---|---|---|---|---|---|
| COL / BeLive | Monetization | 92 | 44.55 | +47.5 | Hidden Strength |
| Amazon | Distribution | 80 | 50.2 | +29.8 | Hidden Strength |
| Amazon | Content | 22 | 50.2 | -28.2 | Vulnerability |
| Disney | Narrative | 93 | 76.55 | +16.5 | Hidden Strength |
| Disney | Content | 55 | 76.55 | -21.6 | Vulnerability |
| DramaBox | Community | 65 | 82.75 | -17.8 | Vulnerability |
Showing 6 of 22 anomalies. Full set available via SPARQL query dimension-anomalies.rq.
The Accuracy Experiment
A controlled comparison framework is built and awaiting the first evaluation cycle:
| Method | Description | Expected Accuracy |
|---|---|---|
| Persistence | Predict no change (delta = 0) | ~33% directional |
| Naive Momentum | Same delta as last week | ~40-50% |
| Mean Reversion | Regression toward tier midpoint | ~40-50% |
| KG-Augmented | SBPI engine (momentum + anomaly + confidence) | Target: >60% |
| LLM Zero-Shot | LLM with no KG context | TBD |
| LLM + KG | LLM with full semantic layer context | TBD |
The hypothesis: KG-augmented predictions outperform statistical baselines because the ontology encodes structural information (dimension weights, tier boundaries, company categories) that pure statistical methods cannot access.
The Scaling Path
Knowledge graphs follow a power-law value curve. Below 1,000 facts, a graph is a reference document. At 10,000-100,000 facts, it becomes a reasoning engine. At 1,000,000+ facts, it becomes a prediction platform.
| Scale | Triples | What Unlocks | Timeline |
|---|---|---|---|
| Current | 2,588 | Single-vertical weekly intelligence, 3 query types live | Now |
| 10K | 10,000 | Cross-vertical comparison (micro-drama vs K-drama vs anime) | Q2 2026 |
| 100K | 100,000 | Temporal pattern library across all verticals + historical data | Q3 2026 |
| 1M | 1,000,000 | Full entertainment landscape with investor signal detection | Q4 2026 |
| 100M | 100,000,000 | Multi-industry competitive intelligence platform | 2027 |
| 1B | 1,000,000,000 | General-purpose business intelligence ontology as a service | 2028+ |
Each row is architecturally supported today. The ETL pipeline ingests any JSON state file. The ontology extends to any vertical by adding instances, not changing the schema. The SPARQL queries generalize — weekly-movers.rq works identically whether tracking 22 companies or 22,000.
Why the IP Is Defensible
1. The Ontology Itself
The dimension weights, tier boundaries, and scoring methodology encode years of domain expertise. Copying the code is trivial. Replicating the judgment embedded in sbpi:DistributionPower weight 0.25 requires understanding why distribution matters more than monetization in a mobile-first vertical. That understanding comes from 7+ client engagements and dozens of competitive analysis cycles.
2. The Attestation Chain
Every fact has a confidence score and source type. As the graph grows, the attestation layer creates a trust gradient. Facts backed by primary data and expert judgment outrank automated inference. This trust gradient is itself queryable: "Show me all ReelShort scores where confidence > 0.90" returns only the most reliable data.
3. The Prediction Track Record
Each prediction cycle writes RDF instances to the store with timestamps and provenance. Over time, this creates an auditable prediction history. A prediction engine with a proven track record is worth orders of magnitude more than one without. The accuracy experiment quantifies this value every cycle.
Capital Deployment Map
| Investment | Amount | Return |
|---|---|---|
| Additional verticals (3-5) | $15K-25K | 5x data volume, cross-vertical correlation signals |
| Historical data backfill (2+ years) | $10K-15K | Temporal pattern library, seasonal signal detection |
| LLM integration for prediction | $5K-10K | Zero-shot vs KG-augmented accuracy comparison |
| Automated source ingestion | $10K-20K | News feeds, app store data, social signals → RDF |
| Platform productization | $30K-50K | Self-service dashboard, API access, white-label reports |
Total: $70K-120K to reach 100K+ triples and prove the prediction accuracy thesis.
Each client engagement (FrameBright, Fiserv, Long Zhu, AHA) produces intelligence briefs worth $2K-10K using this infrastructure. The marginal cost of each additional engagement decreases as the ontology and tooling improve. The infrastructure is the product.
The Self-Improving Cycle
The system is autopoietic — it improves itself through use.
Evidence From Production
- Long Zhu → Produced the Layered Ontology Architecture (K1-K3 + O1-O3). Didn't exist before we needed it.
- FrameBright → Validated the two-site editorial pattern. Now deploys in hours instead of days.
- Fiserv → Proved Brand Power Score scales to corporate brands without schema changes.
- MicroCo (SBPI) → Production semantic layer: 2,588 triples, 24 companies, all 3 SPARQL query types live. W12 predicted DramaBox rise, confirmed by scoring.
- Batch Pipeline → 27 sessions processed autonomously. System writes its own history into structured knowledge. W12 ETL added 916 triples in one cycle.
Parametric vs. Non-Parametric Knowledge
This is where the billion-node thesis comes from. The distinction between what an LLM "knows" and what a knowledge graph knows is the foundation of the entire business model.
Parametric Knowledge (LLM Weights)
When an LLM "knows" that Paris is the capital of France, that knowledge is embedded in its trillions of numerical weights. It is static, hard to update without retraining, and prone to hallucinations because it's a probabilistic guess of what comes next.
- Frozen at training time
- Cannot cite its sources
- Expensive to update (retraining costs millions)
- Broad but shallow — knows a little about everything
Non-Parametric Knowledge (Our KGs)
Knowledge stored in an external structured format — RDF triples, SPARQL-queryable, with provenance and confidence scores. It is explicit, verifiable, and can be updated instantly without retraining any model.
- Updated in real-time (weekly ETL cycles)
- Every fact has an attestation chain
- Update cost: near zero (append new triples)
- Narrow but deep — expert-level in specific domains
The ATLAS paper argues that for a knowledge graph to be a "peer" to an LLM, it needs to reach billion-node critical mass to match the sheer density of facts an LLM has memorized in its parameters. That's the general case. Our hypothesis is that expert-designed, domain-dense knowledge graphs outperform LLM parametric knowledge in concept-specific work at 1/100th the total scale.
Why Domain Density Changes the Math
A general-purpose knowledge graph like Wikidata has ~100 billion triples but knows approximately nothing about micro-drama competitive dynamics. GPT-4 has seen the Wikipedia page for ReelShort but cannot tell you that ReelShort's Distribution Power score dropped 0.55 points last week, or that JioHotstar's momentum is +11.0 at 85% confidence.
Our 2,588 triples contain more actionable intelligence about the micro-drama vertical than the entire parametric knowledge of any LLM. No LLM knows that DramaBox overtook ReelShort in W12-2026, that Amazon's Distribution Power score (80) masks a Content Strength vulnerability (22), or that JioHotstar has 9.45 cumulative momentum at 85% confidence. That's the domain density argument: a small graph with expert-designed ontology beats a large model with generic training in any task that requires structured reasoning over specific facts.
The IP Engine Model
Every intelligence briefing is not just a deliverable. It is a CapEx investment into a proprietary asset. The revenue pays for the construction of a knowledge graph that will eventually be worth millions in recurring licensing.
The Variables
| Variable | Symbol | Definition | Current Value |
|---|---|---|---|
| Node Density | Nd | Unique entities and relationships extracted per briefing | ~50-150 per engagement |
| Ontological Alpha | α | Uniqueness of schema vs. public sets (Wikidata, DBpedia). This is the moat. | High — SBPI dimensions, tier logic, attestation chains have no public equivalent |
| Extraction Efficiency | Ec | Cost (compute + agentic labor) to move a fact from unstructured briefing to structured KG node | ~$0.02/triple (claude -p + InfraNodus API) |
| Decay Rate | λ | How fast domain information becomes obsolete | Low for structural data (ontology), high for scores (weekly refresh) |
"By increasing Ontological Alpha through expert-driven schema design, ShurIQ creates a domain-dense KG that outperforms GPT-4's parametric knowledge in specific expert tasks, even at 1/100th the total scale."
Dual Revenue: Service + Asset Accumulation
Every $2K-10K engagement produces two things simultaneously:
Revenue (Deliverable)
Intelligence brief, editorial site, competitive analysis, Brand Power Score
Value: $2K-10K per engagement
Lifecycle: Consumed by client
Compounding: No — each deliverable is standalone
IP Asset (Knowledge Graph)
50-150 new triples, dimension calibrations, pattern library entries, prediction history
Value: Compounds with every engagement
Lifecycle: Permanent (ontology is additive)
Compounding: Yes — every fact makes the next extraction better
The investor pitch: you're not paying for consultants' time. You're funding the construction of a proprietary database that will eventually be worth millions in recurring licensing. The consulting revenue subsidizes the R&D.
The Karpathy Auto-Research Loop
Andrej Karpathy's "Software 2.0" thesis applied to ontology discovery. Agents don't just extract — they propose, critique, rank, and refine the knowledge graph structure itself.
What We've Built So Far
Steps 1-2 are operational today. The batch-mapupdate pipeline runs headless claude -p agents that read session records and extract entity statements. The InfraNodus API processes these into graph nodes with cluster analysis and gap detection.
Steps 3-4 are partially implemented through SHACL validation (catches schema violations) and the prediction experiment (ranks signal value by accuracy over time).
Step 5 is operational through the wrapup skill and session capture pipeline — human review of extracted entities before they become permanent graph entries.
The Use of Funds Flywheel
Each dollar invested in ShurIQ doesn't pay for a consultant's time. It builds a flywheel where service revenue funds IP accumulation, and the IP makes each subsequent service faster and more valuable.
| Investment Area | Activity | Scalable Asset (IP) |
|---|---|---|
| Agentic Extraction | Automating the Dolma-style pipeline for private briefings. Headless claude -p batch processing. |
A proprietary, locally-hosted "ATLAS" of your client's industry. Updated weekly, queryable via SPARQL. |
| Ontology Design | Mapping dimension weights, tier boundaries, and scoring logic for specific verticals. | A "Schema Library" that can be licensed to other firms. Each vertical adds a new ontology module. |
| Post-Training / Fine-Tuning | Fine-tuning small language models (SLMs) on the specific KG data. | Models that don't hallucinate and "think" like your best analysts. Domain-specific, not general-purpose. |
| Prediction Engine | Running accuracy experiments comparing KG-augmented vs statistical baselines. | A proven prediction track record with documented methodology. The accuracy data IS the moat. |
| Attestation Infrastructure | Building the confidence scoring and source verification pipeline. | Trust gradient across all data. Clients can query "show me only facts backed by primary data." |
The Compounding Math
At current scale (2,588 triples from ~$55K in engagement work), each triple costs ~$21. At the automated batch processing rate (20 seconds per session, ~$0.02 per extracted triple), the marginal cost drops 1,050x. W12 added 916 triples in a single ETL cycle — the system is accelerating. The fixed cost is ontology design. The variable cost approaches zero.
From Service to Platform
This is not service-based reporting. This is IP-based infrastructure. The $100K reports are subsidized R&D that pays for the construction of a database that will eventually be worth millions in recurring licensing. Every engagement makes the next one cheaper, faster, and more accurate — because the knowledge graph grows, and the ontology sharpens.
The Experiment That Proves It
The prediction accuracy experiment is the mechanism that converts this thesis from an argument into evidence.
Experimental Design
For each weekly scoring cycle, we record predictions from 4+ methods against the same companies, then evaluate accuracy when new data loads.
| Method | What It Uses | What It Proves If It Wins |
|---|---|---|
| Persistence | Nothing (predict no change) | Market is random — no method beats chance |
| Naive Momentum | Last week's delta only | Simple statistics are sufficient |
| Mean Reversion | Tier midpoints only | Markets self-correct — no structural model needed |
| KG-Augmented | Full ontology (dimensions, tiers, attestation, signals) | Expert-designed ontology adds predictive value beyond statistics |
| LLM Zero-Shot | Raw question, no KG context | Parametric knowledge alone is sufficient |
| LLM + KG | Full SPARQL context injected into prompt | Non-parametric KG amplifies LLM reasoning |
The hypothesis chain:
- KG-Augmented > Statistical Baselines → Proves ontology design has predictive value
- LLM + KG > LLM Zero-Shot → Proves non-parametric knowledge amplifies LLM reasoning
- Accuracy improves as graph grows → Proves the compounding thesis
If all three hold, the investment thesis is proven: every dollar into the knowledge graph produces exponentially more predictive value over time.
84 predictions generated for W13-2026 (21 companies x 4 methods). W12 evaluation completed with 21 actual scores matched against W12 predictions. The pipeline is live: prediction_experiment.py generates predictions, stores them as JSON, and evaluates accuracy when new data lands. The experiment now has production data flowing through it.