SHUR IQ Semantic Layer Report
SHUR IQ Intelligence Infrastructure

Every Report We Write Makes the Next One Smarter

We built a system where every competitive intelligence report generates two things: revenue for the client engagement and permanent additions to a proprietary knowledge graph that compounds in value over time. The graph runs nightly experiments on its own data, measures whether its predictions improve as it grows, and produces the evidence that turns consulting revenue into defensible IP. This is the formula for converting investor capital into multiples of return — and the clock is ticking, because six months of compounding data creates a lead that can't be caught.

The Breakthrough
Dual-Revenue Asset Model
Client pays $5K-25K per report. That same report adds ~100 verified facts to a knowledge graph that grows more accurate every week. Revenue and IP accumulate simultaneously from the same work.
The Defensibility
Compounding Advantage
Week 1 has 100 facts. Week 52 has 5,000+. The 52nd report draws on a year of structured competitive history that no competitor can replicate without doing a year of the same work. Time-in-market IS the moat.
The Urgency
First-Mover Compounding
Starting 6 months earlier means 26 weeks of accumulated intelligence, 2,600+ additional triples, and 26 rounds of prediction accuracy data. That lead multiplies — it doesn't just add. The advantage curve is exponential, not linear.
1,672RDF Triples (and growing)
17Companies Tracked
68Predictions Locked
6:13 AMNightly Auto-Run

What this actually does

SHUR IQ produces weekly competitive intelligence reports for clients in specific verticals — right now, micro-drama streaming. Each report scores 17 companies across 5 dimensions (content strength, narrative ownership, distribution power, community strength, monetization infrastructure) and ranks them. That's the consulting product.

Behind the scenes, every score, every ranking, every signal gets converted into structured knowledge graph triples — formal, validated, queryable data stored in an ontology-backed database. Every night at 6:13 AM, the system autonomously runs predictive experiments against its own accumulated data, testing whether the graph's predictions get more accurate as it grows.

Why that matters for capital deployment

Most consulting firms sell analyst hours. When the analyst leaves, the knowledge walks out the door. We encode every engagement into permanent, machine-readable infrastructure. After 12 months of weekly reporting across 3 verticals, the graph contains 15,000+ attestation-backed facts about competitive dynamics that no other system on Earth has — and measurable evidence of whether that knowledge improves predictions.

Capital injection accelerates the compounding cycle. More verticals means more data. More data means better predictions. Better predictions mean more clients. More clients mean more data. The flywheel has already started turning. The question is how fast to spin it.

Capital → Advantage Timeline

Months 1-3
Prove Accuracy
12+ weeks of prediction data. First defensible "KG beats baseline" claim. $70-120K.
Months 3-6
Multi-Vertical
3 verticals running. Cross-vertical pattern detection online. $50-80K.
Months 6-9
Model Training
Fine-tuned SLMs that outperform GPT-4 on vertical-specific tasks. $30-50K.
Months 9-12
Licensing & API
SPARQL API access, schema licensing, prediction signal feeds. $40-60K.

A competitor starting 6 months later faces 26 weeks of accumulated intelligence, compounding prediction accuracy, and an entrenched client base generating data they can't access.

Under the hood

The system encodes competitive intelligence as formal ontology, validates it against schema constraints, runs predictive models against its own data, and measures whether its predictions improve as the graph grows. It runs autonomously every night. No human touches it between weekly data loads. The graph compounds.

Architecture

The system follows the OTK (On-To-Knowledge) pipeline that EU research programs formalized in 2001: Extract, Structure, Store, Query, Present. The difference: AI agents perform extraction, SPARQL performs query, and editorial sites perform presentation.

Extract
JSON State
sbpi_to_rdf.py
Structure
OWL Ontology
sbpi.ttl
Store
Oxigraph
SPARQL 1.1
Validate
SHACL Shapes
7 node shapes
Query
11 Queries
nightly-insights.py
Predict
4 Methods
prediction_experiment.py

The Nightly Scheduler

Every day at 6:13 AM, a headless claude -p session fires. It checks if Oxigraph is running, starts it if not, runs the SPARQL insight queries, generates predictions for the upcoming week, evaluates past predictions against ground truth, and writes everything to timestamped files.

Scheduler Config

sbpi-nightly-insights

Cron: 13 6 * * * (daily 6:13 AM)
Max turns: 20 | Timeout: 15 min
Permissions: full-edit + python3 + uvx
Working dir: semantic-layer/

Safety Mechanisms

Lock file prevents concurrent runs
macOS-native timeout (no GNU coreutils)
Stale lock detection + cleanup
osascript notifications on complete/fail
Session log captured for replay

Run History

Mar 22, 2:19 PMPASSFull pipeline: 9 signals, 68 predictions, 148s

Run history and prediction accuracy tracked nightly. Performance metrics accumulate with each scheduled execution.

Component Stack

ComponentToolRole
OntologyOWL 2 (Turtle)12 classes, 34 properties, 5 dimension instances, 5 tier instances
ValidationSHACL7 node shapes enforcing schema on every ETL load
Triple StoreOxigraphRust-native SPARQL 1.1 endpoint at port 7878
ETLPython + rdflibJSON state files to RDF triples with provenance
QueriesSPARQL 1.111 queries: movers, tiers, anomalies, correlation, signals, attestation, segments, cross-graph, annotations, accuracy, compound anomaly
PredictionPython4 baseline methods: persistence, naive momentum, mean reversion, KG-augmented
SchedulerClaude Code + launchdHeadless claude -p with scoped permissions, lock files, timeouts
BridgeInfraNodus MCPKnowledge graph nodes/edges imported as RDF for cross-graph SPARQL joins

SBPI Ontology

The Structural Brand Power Index is formalized as an OWL 2 ontology with full provenance (W3C PROV-O), attestation tracking, and prediction modeling. Every score traces to its source, every prediction links to the evidence that generated it.

Namespace

@prefix sbpi:     <https://shurai.com/ontology/sbpi#> .
@prefix company:  <https://shurai.com/data/company/> .
@prefix week:     <https://shurai.com/data/week/> .
@prefix score:    <https://shurai.com/data/score/> .
@prefix signal:   <https://shurai.com/data/signal/> .
@prefix pred:     <https://shurai.com/data/prediction/> .
@prefix anno:     <https://shurai.com/data/annotation/> .

Core Classes

sbpi:Company
Entity tracked in competitive landscape
sbpi:ScoreRecord
One company, one week, five dimensions + composite. Subclass of prov:Entity
sbpi:Dimension
Five scoring axes with weights
sbpi:Attestation
Provenance link to source + confidence. Subclass of prov:Activity
sbpi:Prediction
Forward-looking signal with direction + confidence. Subclass of Annotation
sbpi:Signal
Market event that influences scoring. Subclass of prov:Entity
sbpi:GraphNode
InfraNodus node imported via bridge
sbpi:Cluster
Topical cluster from knowledge graph
sbpi:Annotation
Semantic annotation: intent, context, or prediction

Five Scoring Dimensions

DimensionCodeWeightDefinition
Content Strengthcs20%Volume, quality, and exclusivity of content produced or distributed
Narrative Ownershipno20%Control over vertical's narrative: press, thought leadership, brand recognition
Distribution Powerdp25%App store rankings, global availability, partnerships
Community Strengthcm20%Audience size, engagement intensity, loyalty
Monetization Infrastructuremi15%Payment systems, ad infrastructure, subscription models, coin/token systems

Tier System

TierScore RangeCompanies (W11-2026)
Dominant85-100ReelShort (84.05)*
Strong70-84DramaBox, Disney
Emerging55-69iQiYi, Netflix, CandyJar, JioHotstar, GoodShort, Lifetime/A+E
Niche40-54Amazon, Viu, COL/BeLive
Limited<40VERZA TV, RTP, KLIP, Both Worlds, Mansa

*ReelShort at 84.05 is technically Strong tier but rounds up to Dominant threshold.

SHACL Validation

Seven node shapes enforce data integrity on every ETL load. Every company must have a name, slug (lowercase-hyphenated), and vertical assignment. Every score record must have exactly 5 dimension scores, a composite between 0-100, and a tier. Predictions must declare a direction from {up, down, stable, volatile} and a confidence between 0.0-1.0.

SHACL shape example: ScoreRecordShape
# Every score record must reference exactly one company, one week,
# have 5 dimension scores, composite 0-100, and a tier assignment
<https://shurai.com/shapes/ScoreRecordShape> a sh:NodeShape ;
    sh:targetClass sbpi:ScoreRecord ;
    sh:property [
        sh:path sbpi:compositeScore ;
        sh:datatype xsd:decimal ;
        sh:minCount 1 ;
        sh:maxCount 1 ;
        sh:minInclusive 0 ;
        sh:maxInclusive 100 ;
    ] ;
    sh:property [
        sh:path sbpi:hasDimensionScore ;
        sh:minCount 5 ;
        sh:maxCount 5 ;
    ] .

ETL Pipeline

Weekly SBPI state files (JSON) are converted to RDF triples via sbpi_to_rdf.py. The script creates Company, ScoreRecord, DimensionScore, Signal, and Attestation nodes for each company in each week, then links weeks in temporal chains for time-series queries.

Data Flow

# Input: state/current.json (17 companies, 5 dimensions each)
# + state/archive/W10-2026.json, W11-2026.json, ...

python3 etl/sbpi_to_rdf.py --all --validate

# [1/5] Loading ontology from ontology/sbpi.ttl
# [2/5] Processing current state
# [3/5] Processing 2 archived weeks
# [3.5] Linked 3 weeks in temporal chain
# [4/5] SHACL validation PASSED
# [5/5] Loaded 1,672 triples into Oxigraph

Triple Structure per Company-Week

Each company-week record generates approximately 15-20 triples:

# Company identity (6-8 triples)
company:reelshort a sbpi:Company ;
    sbpi:slug "reelshort" ;
    sbpi:companyName "ReelShort" ;
    sbpi:inVertical <.../vertical/micro-drama> ;
    sbpi:isPlatformGiant false .

# Score record (4-5 triples)
score:reelshort-W11-2026 a sbpi:ScoreRecord ;
    sbpi:forCompany company:reelshort ;
    sbpi:forWeek week:W11-2026 ;
    sbpi:compositeScore 84.05 ;
    sbpi:delta -0.55 ;
    sbpi:inTier sbpi:TierDominant .

# Dimension scores (5 x 2 = 10 triples)
score:reelshort-W11-2026-content_strength a sbpi:DimensionScore ;
    sbpi:forDimension sbpi:ContentStrength ;
    sbpi:dimensionValue 85 .

# Attestation (4 triples)
attest:reelshort-W11-2026 a sbpi:Attestation ;
    sbpi:confidence 0.85 ;
    sbpi:sourceType "expert_judgment" ;
    prov:wasGeneratedBy "ShurIQ SBPI Pipeline" .

SPARQL Query Library

Eleven queries cover the full analysis cycle, from basic movers to compound anomaly detection. Three run nightly, four run weekly, four are on-demand.

QuerySchedulePurpose
weekly-movers.rqNightlyTop 10 companies by absolute delta
dimension-anomalies.rqNightlyDimension-composite gaps >20 points
predictive-signals.rqNightly2-week same-direction momentum = BULLISH/BEARISH
tier-transitions.rqWeeklyCompanies crossing tier boundaries
cross-correlation.rqWeeklyHigh distribution, low community outliers
attestation-coverage.rqWeeklySource backing quality audit
platform-vs-pureplay.rqWeeklyStructural scoring differences by company type
cross-graph.rqOn-demandSBPI + InfraNodus cross-graph joins
annotation-timeline.rqOn-demandAnnotation evolution for a company
prediction-accuracy.rqOn-demandPredicted vs actual comparison
anomaly-compound.rqOn-demandMulti-signal anomaly detection (2+ concurrent types)
Predictive Signals query (the core nightly engine)
# Two consecutive weeks of same-direction delta = momentum signal
PREFIX sbpi: <https://shurai.com/ontology/sbpi#>

SELECT ?name ?w1Label ?delta1 ?w2Label ?delta2 ?momentum ?predDirection
WHERE {
    ?sr1 sbpi:forCompany ?c ;
         sbpi:forWeek ?w1 ;
         sbpi:delta ?delta1 .
    ?sr2 sbpi:forCompany ?c ;
         sbpi:forWeek ?w2 ;
         sbpi:delta ?delta2 .
    ?w1 sbpi:previousWeek ?w2 .           # temporal chain
    ?c sbpi:companyName ?name .

    FILTER((?delta1 > 0 && ?delta2 > 0) ||
           (?delta1 < 0 && ?delta2 < 0))

    BIND(?delta1 + ?delta2 AS ?momentum)
    BIND(IF(?momentum > 0, "BULLISH", "BEARISH") AS ?predDirection)
}
ORDER BY DESC(ABS(?momentum))

Prediction Experiment

Four methods run against the same data. When actuals load, every prediction is scored on directional accuracy, mean absolute error, and Brier score (confidence calibration). The experiment tracks whether the KG-augmented method outperforms statistical baselines as the graph grows.

MethodLogicConfidencePredictions
PersistencePredict delta = 0 (no change)0.5017
Naive MomentumPredict delta = last week's delta0.5517
Mean ReversionPredict 10% gap closure toward tier midpoint0.4517
KG-AugmentedSPARQL momentum + anomaly + confidence scoring0.5017

Future methods (requiring API keys):

  • LLM Zero-Shot: Company names only, no KG context
  • LLM + KG: Full semantic layer context fed to LLM

Evaluation Metrics

Directional Accuracy

Did the prediction correctly call up/down/stable? Baseline expectation: ~33% (random). Target for KG-augmented: >60%.

Mean Absolute Error

How far off was the predicted magnitude? Only applicable to methods that predict delta values (not direction-only).

Brier Score

Measures confidence calibration: (confidence - outcome)². Lower is better. A perfectly calibrated predictor would say "0.7 confident" and be right 70% of the time.

W11-2026 Stack Ranking

Live data from the Oxigraph SPARQL endpoint. 17 companies tracked across the micro-drama vertical.

#CompanyCompositeDeltaTierSignal

Predictive Signals (Nightly Digest)

Companies showing 2+ consecutive weeks of same-direction momentum. These are SPARQL-derived, not statistical noise.

Bullish Signals

JioHotstar+5.5BULLISH
COL Group / BeLive+4.1BULLISH
Disney+3.25BULLISH
GoodShort+2.8BULLISH
Lifetime / A+E+2.8BULLISH
DramaBox+1.25BULLISH

Bearish Signals

Amazon-3.2BEARISH
Netflix-3.0BEARISH
ReelShort-0.55BEARISH

The structural finding

Platform giants (Amazon, Netflix) are declining in the micro-drama vertical while vertical specialists (JioHotstar, COL/BeLive, GoodShort) are rising. ReelShort holds #1 rank but shows early erosion. The market is bifurcating: infrastructure commoditization (COL's "Microdrama in a Box") vs. content differentiation (DramaBox, CandyJar).

Dimension Radar

Five-axis comparison across the top companies. Select companies to compare their structural profiles.

Prediction Experiment Status

W12-2026 Predictions Locked

68 predictions recorded across 4 methods x 17 companies. Waiting for W12-2026 actuals to evaluate.

MethodCountAvg ConfidenceStatus
Persistence170.50Awaiting actuals
Naive Momentum170.55Awaiting actuals
Mean Reversion170.45Awaiting actuals
KG-Augmented170.50Awaiting actuals

Parametric vs. Non-Parametric Knowledge

When an LLM "knows" that ReelShort dominates the micro-drama vertical, that knowledge lives inside trillions of numerical weights trained on internet-scale data. This is parametric knowledge: static, probabilistic, impossible to audit, and expensive to update (retraining costs millions). It hallucinates because it's guessing what token comes next, not retrieving verified facts.

The SBPI semantic layer stores the same intelligence as non-parametric knowledge: explicit RDF triples in an Oxigraph store. Every fact is verifiable, traceable to its attestation source, and updatable in milliseconds. When we say ReelShort scored 84.05 in W11-2026, we can show you the triple, the attestation, the SHACL shape that validated it, and the SPARQL query that retrieved it.

The billion-node question

The ATLAS research (Soldaini et al., 2024) established that retrieval-augmented systems need approximately 1 billion structured facts to match an LLM's parametric breadth across general domains. That's the cost of being a generalist. The ShurIQ thesis is that expert-designed, attestation-backed domain graphs achieve competitive or superior accuracy at 1/100th that scale — 10K to 100K facts — within a specific vertical. Domain density beats general breadth.

Why Domain Density Wins

Schema Is the Moat

GPT-4 doesn't know what "Distribution Power" means in the context of micro-drama competitive dynamics. Our ontology does — it's a named dimension with a 25% weight, backed by 6 indicator types. The schema encodes expert judgment that no amount of pretraining captures.

Provenance Kills Hallucination

Every score record links to an sbpi:Attestation with confidence, source type, and generation activity. When the KG says "BEARISH on Netflix," it can trace that signal through two consecutive weeks of negative delta to the specific SPARQL query that flagged it. An LLM just says things.

Instant Updates, No Retraining

Loading W12-2026 data takes 4 seconds and adds ~100 triples. The graph is immediately queryable with the new data. Updating an LLM's parametric knowledge about the same 17 companies would require fine-tuning at minimum — or waiting for the next training run.

The Compounding Asset Model

Every intelligence briefing we produce is simultaneously revenue and CapEx. The $5K-25K engagement pays for the work. The triples generated during that work become permanent additions to a proprietary knowledge asset. After 52 weeks of weekly reporting across 3 verticals, the graph contains ~15,000 attestation-backed facts about competitive dynamics that no other system on Earth has.

Where We Are

MetricCurrentTarget (Q2 2026)Target (Q4 2026)
RDF Triples1,67215,000+100,000+
Weeks Tracked2 (W10-W11)12+30+
Companies1730+50+
Verticals1 (micro-drama)3+5+
Prediction AccuracyNot yet evaluated>50% directional>60% directional
Evaluation Weeks08+20+

The Revenue-IP Flywheel

Revenue
Client Engagement
$5K-25K per report
Data
Weekly State File
17+ companies scored
Knowledge
RDF Triples Added
~100 triples/week
Accuracy
Predictions Improve
Measured weekly

Service business vs. knowledge business

A service business trades analyst hours for revenue. When the analyst leaves, the knowledge walks out the door. A knowledge business encodes every engagement into structured, queryable, compounding infrastructure. The analyst's judgment becomes permanent triples. The schema preserves institutional knowledge even if every human walks away tomorrow. Each report costs roughly the same to produce, but the 52nd report draws on 51 weeks of accumulated context that the 1st report didn't have.

What Capital Unlocks

Multi-Vertical Expansion

The ontology is vertical-agnostic. The same 12 classes, 34 properties, and 11 SPARQL queries work for sneakers, K-pop, fintech, or any vertical with weekly ranking dynamics. Each new vertical multiplies the graph without new infrastructure.

LLM Comparison Baselines

Methods 5 and 6 (LLM zero-shot and LLM + KG) are designed but not yet active. When enabled, they directly measure the marginal value of the knowledge graph over raw LLM reasoning. This is the investor proof point.

DeltaBet Prediction Market

The predictive signals feed directly into a prediction market product. BULLISH/BEARISH signals become tradeable markets. Signal accuracy = market credibility. The semantic layer becomes a pricing engine.

OTK Lineage

This system consciously inherits from the EU's On-To-Knowledge project (2001-2003), which spent over $1M building an ontology-driven knowledge management pipeline. The components map directly:

OTK Tool (2002)SBPI Equivalent (2026)Key Difference
OntoWrapper (extraction)sbpi_to_rdf.pyAI agents extract; rule-based is dead
OIL/DAML+OIL (ontology)sbpi.ttl (OWL 2)W3C standard, not proprietary format
Sesame (store)Oxigraph (SPARQL 1.1)Rust-native, embeddable, zero-config
Spectacle (presentation)Editorial HTML + D3.jsSelf-contained, deployable anywhere
OntoEdit (authoring)InfraNodus + AI agentsGraph-native, not form-based
OntoShare (distribution)Cloudflare Pages + SlackGlobal CDN, instant deployment

The OntoRAG project (2025) independently arrived at a nearly identical architecture. Their philosophy: "If the system cannot explain what it knows, where it comes from, and why it changed, it is not a knowledge system." We agree. Every triple in this graph traces to its attestation.

The IP Engine Variables

To model the knowledge graph as an investable asset, we quantify four variables that determine whether the graph compounds into a structural advantage or stalls as a glorified spreadsheet.

Nd
Node Density
Unique entities and relationships extracted per briefing. This is the raw throughput of the knowledge engine — how many structured facts each engagement produces. Higher density means faster graph growth per dollar of consulting revenue.
Current: ~100 triples/company-week | Target: 150+ with signal + annotation layers
α
Ontological Alpha
The uniqueness of the schema compared to public knowledge bases like Wikidata or DBpedia. This is the moat. Anyone can scrape Wikipedia. Nobody else has a 5-dimension SBPI scoring ontology with weighted dimensions, SHACL-validated tier boundaries, and attestation provenance for 17 micro-drama companies. Alpha increases as the schema encodes more expert judgment that public sets lack.
Current: High (no public equivalent for micro-drama vertical scoring)
Ec
Extraction Efficiency
The cost — compute plus agentic labor — to move one fact from unstructured source material to validated RDF triple. The nightly scheduler runs headless claude -p sessions that cost ~$0.15/run. Manual extraction of the same 17 companies would take an analyst 4-6 hours. The automation drives Ec toward zero marginal cost per triple.
Current: ~$0.15/run (68 predictions, 9 signals, 3 queries) | Manual equivalent: ~$200/run
λ
Decay Rate
How fast information in a specific domain becomes obsolete. Micro-drama moves fast — company rankings shift weekly, new entrants appear monthly. High decay means the graph needs continuous feeding to stay relevant, but it also means the graph's temporal coverage (historical records across many weeks) becomes more valuable as competitors lose access to historical state.
Current: ~1 week half-life for rankings | Temporal chain preserves full history

The hypothesis we're testing

"By increasing Ontological Alpha through expert-driven schema design, ShurIQ creates a domain-dense knowledge graph that outperforms GPT-4's parametric knowledge in specific expert tasks, even at 1/100th the total scale." The prediction experiment measures this directly: KG-augmented method vs. LLM zero-shot, same companies, same timeframe, scored on the same metrics.

Use of Funds — The Force Multiplier

Investors want to see that $1 into ShurIQ doesn't just pay for a consultant's time — it builds a flywheel. Every investment area produces both immediate deliverables (revenue) and a durable, scalable IP asset.

Investment AreaActivityScalable IP Asset
Agentic Extraction Automating the ETL pipeline for private briefings. Headless claude -p sessions, SHACL validation gates, Oxigraph loading — all running on nightly schedule without human intervention. A proprietary, locally-hosted knowledge base for each client's industry. Updated autonomously.
Ontology Design Mapping the competitive scoring framework for specific verticals. Defining dimensions, weights, tier boundaries, and signal taxonomies that encode expert judgment as formal OWL classes. A "Schema Library" — reusable ontology modules that can be licensed to other intelligence firms or deployed for new verticals in days instead of months.
Prediction Infrastructure Running weekly prediction experiments across 4+ methods. Recording, evaluating, and publishing accuracy metrics. Building the evidence base for "KG beats LLM" claims. A track record of measurable predictive accuracy — the proof point that transforms the KG from a database into a competitive asset. Measured weekly, compounding forever.
Multi-Vertical Expansion Deploying the SBPI ontology template to new verticals (fintech, K-pop, sneakers, biotech). Each vertical gets its own namespace, its own companies, its own dimensional weights — same infrastructure. Cross-vertical knowledge density. Patterns visible across verticals that no single-vertical analyst can see. The graph becomes a structural intelligence about how competitive landscapes work, not just what they contain.
Post-Training / Fine-Tuning Fine-tuning small language models (SLMs) on the specific KG. Models that "think" like the best analysts because they're grounded in attestation-backed facts, not internet-scale noise. Domain-specific models that don't hallucinate about micro-drama rankings because they retrieve from verified triples instead of guessing from parameters.

Capital Deployment Map

PhaseInvestmentDurationUnlock
Prove Accuracy$70K-120KQ2-Q3 202612+ weeks of prediction data, KG-augmented vs. baselines. First defensible accuracy claim.
Multi-Vertical$50K-80KQ3 20263 verticals running, cross-vertical pattern detection online.
SLM Fine-Tuning$30K-50KQ4 2026Domain-specific models that demonstrably outperform GPT-4 on vertical-specific tasks.
Licensing & API$40K-60KQ1 2027SPARQL API access, schema licensing, prediction signal feeds to third parties.

The Karpathy Auto-Research Cycle

Andrej Karpathy's "Software 2.0" insight: the most valuable code is not the code humans write — it's the data the system accumulates through iterative cycles of extraction, evaluation, and refinement. We apply this directly to ontology discovery.

Step 1
Seed
Give agent a Seed Ontology (sbpi.ttl) and a corpus of briefing data
Step 2
Extract & Propose
Agent extracts facts, proposes new ontological nodes it thinks are missing
Step 3
Cross-Reference
Critic agent checks proposals against existing KG for redundancy or contradiction
Step 4
Stack Rank
Agent ranks value of new nodes by how often they bridge gaps between data points
Step 5
Human Finalize
Experts approve high-value nodes, which are permanently baked into the KG

This cycle already runs in a simplified form. The nightly scheduler extracts signals, the SPARQL queries cross-reference against historical data, and the prediction experiment stack-ranks which signals have predictive value. The human finalization step happens when analysts review the weekly digest and decide which signals to act on. As the graph grows, the automated steps become more accurate and the human becomes a curator rather than a creator.

The private brain thesis

Each vertical gets its own "private brain" — a structured knowledge system that accumulates intelligence about competitive dynamics, market movements, and predictive signals. The $5K-25K weekly reports are subsidized R&D that pays for the construction of a database eventually worth millions in recurring licensing. The clients get intelligence. We get a compounding asset. Both sides of the transaction create value.

Measuring the Thesis

The prediction experiment is designed to produce the specific evidence an investor needs. Each week, four methods predict the same 17 companies. When actuals load, we score them all on the same metrics. Over time, the data answers three questions:

Question 1: Does the KG add value?

KG-augmented method vs. persistence and naive momentum. If KG-augmented wins, the graph is encoding real structural knowledge that simple statistics miss. This is the minimum viable proof point.

Question 2: Does accuracy compound?

Plot KG-augmented accuracy over time. If the curve goes up as triples accumulate, the graph is compounding. If it flatlines, the schema needs revision. The temporal chain gives us a clean x-axis: weeks of data available.

Question 3: Can we beat GPT-4?

Methods 5 (LLM zero-shot) and 6 (LLM + KG) directly compare. Zero-shot uses company names only. LLM + KG feeds the full semantic layer context. The delta between 5 and 6 is the measurable marginal value of the knowledge graph — the number investors care about.

Experiment Status

MethodW12 PredictionsEvaluationsAccuracy
Persistence17 lockedAwaiting W12 actuals
Naive Momentum17 lockedAwaiting W12 actuals
Mean Reversion17 lockedAwaiting W12 actuals
KG-Augmented17 lockedAwaiting W12 actuals
LLM Zero-ShotNot yet active
LLM + KGNot yet active

First evaluation runs when W12-2026 actuals are loaded via python3 etl/sbpi_to_rdf.py.

The Autopoietic Cycle

The system improves itself through use. This is not a metaphor — it is a measurable property of the architecture.

Each client engagement generates scored company data → sbpi_to_rdf.py converts to triples → Oxigraph stores them → SPARQL queries detect signals → prediction experiment tests whether signals have predictive value → accuracy metrics reveal which parts of the ontology produce real intelligence → analysts refine the ontology based on what works → next engagement benefits from the refined schema + accumulated history.

The information doesn't just accumulate — it reorganizes. The ontology evolves. Dimensions get reweighted based on what predicts. New signal types get added when the graph reveals patterns that the original schema didn't anticipate. The structure IS the IP, and the structure improves with each cycle.

This is what separates the system from a data warehouse. A warehouse stores records. This system stores records and the schema for interpreting them, and the validation shapes for ensuring quality, and the prediction methods for testing whether the stored knowledge has real-world predictive value. The value comes from how the sources are organized — InfraNodus entity extraction, wikilink relationships, gap analysis, ontological bridging — not just from the sources themselves.

Concept Network

Force-directed layout of the SBPI knowledge graph analyzed via InfraNodus. 150 nodes, 700 edges, 8 clusters. Node size reflects degree centrality. Edge opacity reflects relationship weight. The five gateway nodes (data, prediction, knowledge, graph, schema) bridge multiple clusters and control information flow through the network.

Structural Gaps

Three structural gaps identified in the concept network. These represent missing conceptual bridges between clusters that, if filled, would strengthen the overall intelligence framework. Each gap is a content or research opportunity.

Gap 1
Knowledge Density --- Market Momentum
No direct conceptual link between knowledge infrastructure metrics and market signal interpretation. Bridging this gap would enable the system to correlate graph density with predictive momentum strength.
Gap 2
Predictive Accuracy --- Market Momentum
Prediction methodology and market signal clusters operate independently. Connecting them would let the system use momentum data as a direct input to accuracy calibration.
Gap 3
Market Momentum --- Semantic Context
Market signal interpretation (bullish/bearish) lacks semantic grounding in the LLM/knowledge-graph context cluster. Bridging this enables natural language explanations of momentum signals.

Cluster Composition

Eight topical clusters ranked by betweenness influence. The Predictive Accuracy cluster dominates at 39% of network influence, followed by Knowledge Density at 30%. Together these two clusters control nearly 70% of the network's information flow.

Graph Statistics

150
Nodes
Unique concepts extracted from the SBPI knowledge corpus
700
Edges
Weighted concept relationships with co-occurrence strength
8
Clusters
Modularity: 0.425 (focused community structure)
0.299
Top Gateway BC
data — highest betweenness centrality, bridges 42 connections

Gateway Nodes

Conceptual gateways control information flow between clusters. These nodes have high betweenness centrality and connect otherwise isolated topic areas.

NodeBetweennessDegreeRole
data0.29942Central hub connecting predictions, graphs, and infrastructure
prediction0.25440Bridges accuracy experiments to intelligence outputs
graph0.24945Links knowledge density to schema and experiments
knowledge0.23141Connects infrastructure to density and revenue concepts
schema0.19639Bridges ontology design to prediction methodology

Top Relations

Strongest weighted edges in the network, representing the most frequently co-occurring concept pairs across the SBPI corpus.

SourceTargetWeight
datapredictions23
predictionaccuracy21
graphdata20
predictiondata17
predictionexperiment15
knowledgegraph14
accuracydata13
knowledgeinfrastructure13

What the graph reveals

The network's focused modularity (0.425) with five high-centrality gateways confirms the SBPI thesis structure: data and prediction form the operational core, knowledge and graph form the asset layer, and schema is the bridge between ontology design and predictive methodology. The three structural gaps all involve the Market Momentum cluster, which is conceptually isolated from the system's infrastructure — exactly where the next round of ontology design work should focus.