We built a system where every competitive intelligence report generates two things: revenue for the client engagement and permanent additions to a proprietary knowledge graph that compounds in value over time. The graph runs nightly experiments on its own data, measures whether its predictions improve as it grows, and produces the evidence that turns consulting revenue into defensible IP. This is the formula for converting investor capital into multiples of return — and the clock is ticking, because six months of compounding data creates a lead that can't be caught.
SHUR IQ produces weekly competitive intelligence reports for clients in specific verticals — right now, micro-drama streaming. Each report scores 17 companies across 5 dimensions (content strength, narrative ownership, distribution power, community strength, monetization infrastructure) and ranks them. That's the consulting product.
Behind the scenes, every score, every ranking, every signal gets converted into structured knowledge graph triples — formal, validated, queryable data stored in an ontology-backed database. Every night at 6:13 AM, the system autonomously runs predictive experiments against its own accumulated data, testing whether the graph's predictions get more accurate as it grows.
Most consulting firms sell analyst hours. When the analyst leaves, the knowledge walks out the door. We encode every engagement into permanent, machine-readable infrastructure. After 12 months of weekly reporting across 3 verticals, the graph contains 15,000+ attestation-backed facts about competitive dynamics that no other system on Earth has — and measurable evidence of whether that knowledge improves predictions.
Capital injection accelerates the compounding cycle. More verticals means more data. More data means better predictions. Better predictions mean more clients. More clients mean more data. The flywheel has already started turning. The question is how fast to spin it.
A competitor starting 6 months later faces 26 weeks of accumulated intelligence, compounding prediction accuracy, and an entrenched client base generating data they can't access.
The system encodes competitive intelligence as formal ontology, validates it against schema constraints, runs predictive models against its own data, and measures whether its predictions improve as the graph grows. It runs autonomously every night. No human touches it between weekly data loads. The graph compounds.
The system follows the OTK (On-To-Knowledge) pipeline that EU research programs formalized in 2001: Extract, Structure, Store, Query, Present. The difference: AI agents perform extraction, SPARQL performs query, and editorial sites perform presentation.
Every day at 6:13 AM, a headless claude -p session fires. It checks if Oxigraph is running, starts it if not, runs the SPARQL insight queries, generates predictions for the upcoming week, evaluates past predictions against ground truth, and writes everything to timestamped files.
sbpi-nightly-insights
Cron: 13 6 * * * (daily 6:13 AM)
Max turns: 20 | Timeout: 15 min
Permissions: full-edit + python3 + uvx
Working dir: semantic-layer/
Lock file prevents concurrent runs
macOS-native timeout (no GNU coreutils)
Stale lock detection + cleanup
osascript notifications on complete/fail
Session log captured for replay
| Mar 22, 2:19 PM | PASS | Full pipeline: 9 signals, 68 predictions, 148s |
Run history and prediction accuracy tracked nightly. Performance metrics accumulate with each scheduled execution.
| Component | Tool | Role |
|---|---|---|
| Ontology | OWL 2 (Turtle) | 12 classes, 34 properties, 5 dimension instances, 5 tier instances |
| Validation | SHACL | 7 node shapes enforcing schema on every ETL load |
| Triple Store | Oxigraph | Rust-native SPARQL 1.1 endpoint at port 7878 |
| ETL | Python + rdflib | JSON state files to RDF triples with provenance |
| Queries | SPARQL 1.1 | 11 queries: movers, tiers, anomalies, correlation, signals, attestation, segments, cross-graph, annotations, accuracy, compound anomaly |
| Prediction | Python | 4 baseline methods: persistence, naive momentum, mean reversion, KG-augmented |
| Scheduler | Claude Code + launchd | Headless claude -p with scoped permissions, lock files, timeouts |
| Bridge | InfraNodus MCP | Knowledge graph nodes/edges imported as RDF for cross-graph SPARQL joins |
The Structural Brand Power Index is formalized as an OWL 2 ontology with full provenance (W3C PROV-O), attestation tracking, and prediction modeling. Every score traces to its source, every prediction links to the evidence that generated it.
@prefix sbpi: <https://shurai.com/ontology/sbpi#> . @prefix company: <https://shurai.com/data/company/> . @prefix week: <https://shurai.com/data/week/> . @prefix score: <https://shurai.com/data/score/> . @prefix signal: <https://shurai.com/data/signal/> . @prefix pred: <https://shurai.com/data/prediction/> . @prefix anno: <https://shurai.com/data/annotation/> .
| Dimension | Code | Weight | Definition |
|---|---|---|---|
| Content Strength | cs | 20% | Volume, quality, and exclusivity of content produced or distributed |
| Narrative Ownership | no | 20% | Control over vertical's narrative: press, thought leadership, brand recognition |
| Distribution Power | dp | 25% | App store rankings, global availability, partnerships |
| Community Strength | cm | 20% | Audience size, engagement intensity, loyalty |
| Monetization Infrastructure | mi | 15% | Payment systems, ad infrastructure, subscription models, coin/token systems |
| Tier | Score Range | Companies (W11-2026) |
|---|---|---|
| Dominant | 85-100 | ReelShort (84.05)* |
| Strong | 70-84 | DramaBox, Disney |
| Emerging | 55-69 | iQiYi, Netflix, CandyJar, JioHotstar, GoodShort, Lifetime/A+E |
| Niche | 40-54 | Amazon, Viu, COL/BeLive |
| Limited | <40 | VERZA TV, RTP, KLIP, Both Worlds, Mansa |
*ReelShort at 84.05 is technically Strong tier but rounds up to Dominant threshold.
Seven node shapes enforce data integrity on every ETL load. Every company must have a name, slug (lowercase-hyphenated), and vertical assignment. Every score record must have exactly 5 dimension scores, a composite between 0-100, and a tier. Predictions must declare a direction from {up, down, stable, volatile} and a confidence between 0.0-1.0.
# Every score record must reference exactly one company, one week, # have 5 dimension scores, composite 0-100, and a tier assignment <https://shurai.com/shapes/ScoreRecordShape> a sh:NodeShape ; sh:targetClass sbpi:ScoreRecord ; sh:property [ sh:path sbpi:compositeScore ; sh:datatype xsd:decimal ; sh:minCount 1 ; sh:maxCount 1 ; sh:minInclusive 0 ; sh:maxInclusive 100 ; ] ; sh:property [ sh:path sbpi:hasDimensionScore ; sh:minCount 5 ; sh:maxCount 5 ; ] .
Weekly SBPI state files (JSON) are converted to RDF triples via sbpi_to_rdf.py. The script creates Company, ScoreRecord, DimensionScore, Signal, and Attestation nodes for each company in each week, then links weeks in temporal chains for time-series queries.
# Input: state/current.json (17 companies, 5 dimensions each) # + state/archive/W10-2026.json, W11-2026.json, ... python3 etl/sbpi_to_rdf.py --all --validate # [1/5] Loading ontology from ontology/sbpi.ttl # [2/5] Processing current state # [3/5] Processing 2 archived weeks # [3.5] Linked 3 weeks in temporal chain # [4/5] SHACL validation PASSED # [5/5] Loaded 1,672 triples into Oxigraph
Each company-week record generates approximately 15-20 triples:
# Company identity (6-8 triples) company:reelshort a sbpi:Company ; sbpi:slug "reelshort" ; sbpi:companyName "ReelShort" ; sbpi:inVertical <.../vertical/micro-drama> ; sbpi:isPlatformGiant false . # Score record (4-5 triples) score:reelshort-W11-2026 a sbpi:ScoreRecord ; sbpi:forCompany company:reelshort ; sbpi:forWeek week:W11-2026 ; sbpi:compositeScore 84.05 ; sbpi:delta -0.55 ; sbpi:inTier sbpi:TierDominant . # Dimension scores (5 x 2 = 10 triples) score:reelshort-W11-2026-content_strength a sbpi:DimensionScore ; sbpi:forDimension sbpi:ContentStrength ; sbpi:dimensionValue 85 . # Attestation (4 triples) attest:reelshort-W11-2026 a sbpi:Attestation ; sbpi:confidence 0.85 ; sbpi:sourceType "expert_judgment" ; prov:wasGeneratedBy "ShurIQ SBPI Pipeline" .
Eleven queries cover the full analysis cycle, from basic movers to compound anomaly detection. Three run nightly, four run weekly, four are on-demand.
| Query | Schedule | Purpose |
|---|---|---|
weekly-movers.rq | Nightly | Top 10 companies by absolute delta |
dimension-anomalies.rq | Nightly | Dimension-composite gaps >20 points |
predictive-signals.rq | Nightly | 2-week same-direction momentum = BULLISH/BEARISH |
tier-transitions.rq | Weekly | Companies crossing tier boundaries |
cross-correlation.rq | Weekly | High distribution, low community outliers |
attestation-coverage.rq | Weekly | Source backing quality audit |
platform-vs-pureplay.rq | Weekly | Structural scoring differences by company type |
cross-graph.rq | On-demand | SBPI + InfraNodus cross-graph joins |
annotation-timeline.rq | On-demand | Annotation evolution for a company |
prediction-accuracy.rq | On-demand | Predicted vs actual comparison |
anomaly-compound.rq | On-demand | Multi-signal anomaly detection (2+ concurrent types) |
# Two consecutive weeks of same-direction delta = momentum signal PREFIX sbpi: <https://shurai.com/ontology/sbpi#> SELECT ?name ?w1Label ?delta1 ?w2Label ?delta2 ?momentum ?predDirection WHERE { ?sr1 sbpi:forCompany ?c ; sbpi:forWeek ?w1 ; sbpi:delta ?delta1 . ?sr2 sbpi:forCompany ?c ; sbpi:forWeek ?w2 ; sbpi:delta ?delta2 . ?w1 sbpi:previousWeek ?w2 . # temporal chain ?c sbpi:companyName ?name . FILTER((?delta1 > 0 && ?delta2 > 0) || (?delta1 < 0 && ?delta2 < 0)) BIND(?delta1 + ?delta2 AS ?momentum) BIND(IF(?momentum > 0, "BULLISH", "BEARISH") AS ?predDirection) } ORDER BY DESC(ABS(?momentum))
Four methods run against the same data. When actuals load, every prediction is scored on directional accuracy, mean absolute error, and Brier score (confidence calibration). The experiment tracks whether the KG-augmented method outperforms statistical baselines as the graph grows.
| Method | Logic | Confidence | Predictions |
|---|---|---|---|
| Persistence | Predict delta = 0 (no change) | 0.50 | 17 |
| Naive Momentum | Predict delta = last week's delta | 0.55 | 17 |
| Mean Reversion | Predict 10% gap closure toward tier midpoint | 0.45 | 17 |
| KG-Augmented | SPARQL momentum + anomaly + confidence scoring | 0.50 | 17 |
Future methods (requiring API keys):
Did the prediction correctly call up/down/stable? Baseline expectation: ~33% (random). Target for KG-augmented: >60%.
How far off was the predicted magnitude? Only applicable to methods that predict delta values (not direction-only).
Measures confidence calibration: (confidence - outcome)². Lower is better. A perfectly calibrated predictor would say "0.7 confident" and be right 70% of the time.
Live data from the Oxigraph SPARQL endpoint. 17 companies tracked across the micro-drama vertical.
| # | Company | Composite | Delta | Tier | Signal |
|---|
Companies showing 2+ consecutive weeks of same-direction momentum. These are SPARQL-derived, not statistical noise.
| JioHotstar | +5.5 | BULLISH |
| COL Group / BeLive | +4.1 | BULLISH |
| Disney | +3.25 | BULLISH |
| GoodShort | +2.8 | BULLISH |
| Lifetime / A+E | +2.8 | BULLISH |
| DramaBox | +1.25 | BULLISH |
| Amazon | -3.2 | BEARISH |
| Netflix | -3.0 | BEARISH |
| ReelShort | -0.55 | BEARISH |
Platform giants (Amazon, Netflix) are declining in the micro-drama vertical while vertical specialists (JioHotstar, COL/BeLive, GoodShort) are rising. ReelShort holds #1 rank but shows early erosion. The market is bifurcating: infrastructure commoditization (COL's "Microdrama in a Box") vs. content differentiation (DramaBox, CandyJar).
Five-axis comparison across the top companies. Select companies to compare their structural profiles.
68 predictions recorded across 4 methods x 17 companies. Waiting for W12-2026 actuals to evaluate.
| Method | Count | Avg Confidence | Status |
|---|---|---|---|
| Persistence | 17 | 0.50 | Awaiting actuals |
| Naive Momentum | 17 | 0.55 | Awaiting actuals |
| Mean Reversion | 17 | 0.45 | Awaiting actuals |
| KG-Augmented | 17 | 0.50 | Awaiting actuals |
When an LLM "knows" that ReelShort dominates the micro-drama vertical, that knowledge lives inside trillions of numerical weights trained on internet-scale data. This is parametric knowledge: static, probabilistic, impossible to audit, and expensive to update (retraining costs millions). It hallucinates because it's guessing what token comes next, not retrieving verified facts.
The SBPI semantic layer stores the same intelligence as non-parametric knowledge: explicit RDF triples in an Oxigraph store. Every fact is verifiable, traceable to its attestation source, and updatable in milliseconds. When we say ReelShort scored 84.05 in W11-2026, we can show you the triple, the attestation, the SHACL shape that validated it, and the SPARQL query that retrieved it.
The ATLAS research (Soldaini et al., 2024) established that retrieval-augmented systems need approximately 1 billion structured facts to match an LLM's parametric breadth across general domains. That's the cost of being a generalist. The ShurIQ thesis is that expert-designed, attestation-backed domain graphs achieve competitive or superior accuracy at 1/100th that scale — 10K to 100K facts — within a specific vertical. Domain density beats general breadth.
GPT-4 doesn't know what "Distribution Power" means in the context of micro-drama competitive dynamics. Our ontology does — it's a named dimension with a 25% weight, backed by 6 indicator types. The schema encodes expert judgment that no amount of pretraining captures.
Every score record links to an sbpi:Attestation with confidence, source type, and generation activity. When the KG says "BEARISH on Netflix," it can trace that signal through two consecutive weeks of negative delta to the specific SPARQL query that flagged it. An LLM just says things.
Loading W12-2026 data takes 4 seconds and adds ~100 triples. The graph is immediately queryable with the new data. Updating an LLM's parametric knowledge about the same 17 companies would require fine-tuning at minimum — or waiting for the next training run.
Every intelligence briefing we produce is simultaneously revenue and CapEx. The $5K-25K engagement pays for the work. The triples generated during that work become permanent additions to a proprietary knowledge asset. After 52 weeks of weekly reporting across 3 verticals, the graph contains ~15,000 attestation-backed facts about competitive dynamics that no other system on Earth has.
| Metric | Current | Target (Q2 2026) | Target (Q4 2026) |
|---|---|---|---|
| RDF Triples | 1,672 | 15,000+ | 100,000+ |
| Weeks Tracked | 2 (W10-W11) | 12+ | 30+ |
| Companies | 17 | 30+ | 50+ |
| Verticals | 1 (micro-drama) | 3+ | 5+ |
| Prediction Accuracy | Not yet evaluated | >50% directional | >60% directional |
| Evaluation Weeks | 0 | 8+ | 20+ |
A service business trades analyst hours for revenue. When the analyst leaves, the knowledge walks out the door. A knowledge business encodes every engagement into structured, queryable, compounding infrastructure. The analyst's judgment becomes permanent triples. The schema preserves institutional knowledge even if every human walks away tomorrow. Each report costs roughly the same to produce, but the 52nd report draws on 51 weeks of accumulated context that the 1st report didn't have.
The ontology is vertical-agnostic. The same 12 classes, 34 properties, and 11 SPARQL queries work for sneakers, K-pop, fintech, or any vertical with weekly ranking dynamics. Each new vertical multiplies the graph without new infrastructure.
Methods 5 and 6 (LLM zero-shot and LLM + KG) are designed but not yet active. When enabled, they directly measure the marginal value of the knowledge graph over raw LLM reasoning. This is the investor proof point.
The predictive signals feed directly into a prediction market product. BULLISH/BEARISH signals become tradeable markets. Signal accuracy = market credibility. The semantic layer becomes a pricing engine.
This system consciously inherits from the EU's On-To-Knowledge project (2001-2003), which spent over $1M building an ontology-driven knowledge management pipeline. The components map directly:
| OTK Tool (2002) | SBPI Equivalent (2026) | Key Difference |
|---|---|---|
| OntoWrapper (extraction) | sbpi_to_rdf.py | AI agents extract; rule-based is dead |
| OIL/DAML+OIL (ontology) | sbpi.ttl (OWL 2) | W3C standard, not proprietary format |
| Sesame (store) | Oxigraph (SPARQL 1.1) | Rust-native, embeddable, zero-config |
| Spectacle (presentation) | Editorial HTML + D3.js | Self-contained, deployable anywhere |
| OntoEdit (authoring) | InfraNodus + AI agents | Graph-native, not form-based |
| OntoShare (distribution) | Cloudflare Pages + Slack | Global CDN, instant deployment |
The OntoRAG project (2025) independently arrived at a nearly identical architecture. Their philosophy: "If the system cannot explain what it knows, where it comes from, and why it changed, it is not a knowledge system." We agree. Every triple in this graph traces to its attestation.
To model the knowledge graph as an investable asset, we quantify four variables that determine whether the graph compounds into a structural advantage or stalls as a glorified spreadsheet.
claude -p sessions that cost ~$0.15/run. Manual extraction of the same 17 companies would take an analyst 4-6 hours. The automation drives Ec toward zero marginal cost per triple."By increasing Ontological Alpha through expert-driven schema design, ShurIQ creates a domain-dense knowledge graph that outperforms GPT-4's parametric knowledge in specific expert tasks, even at 1/100th the total scale." The prediction experiment measures this directly: KG-augmented method vs. LLM zero-shot, same companies, same timeframe, scored on the same metrics.
Investors want to see that $1 into ShurIQ doesn't just pay for a consultant's time — it builds a flywheel. Every investment area produces both immediate deliverables (revenue) and a durable, scalable IP asset.
| Investment Area | Activity | Scalable IP Asset |
|---|---|---|
| Agentic Extraction | Automating the ETL pipeline for private briefings. Headless claude -p sessions, SHACL validation gates, Oxigraph loading — all running on nightly schedule without human intervention. |
A proprietary, locally-hosted knowledge base for each client's industry. Updated autonomously. |
| Ontology Design | Mapping the competitive scoring framework for specific verticals. Defining dimensions, weights, tier boundaries, and signal taxonomies that encode expert judgment as formal OWL classes. | A "Schema Library" — reusable ontology modules that can be licensed to other intelligence firms or deployed for new verticals in days instead of months. |
| Prediction Infrastructure | Running weekly prediction experiments across 4+ methods. Recording, evaluating, and publishing accuracy metrics. Building the evidence base for "KG beats LLM" claims. | A track record of measurable predictive accuracy — the proof point that transforms the KG from a database into a competitive asset. Measured weekly, compounding forever. |
| Multi-Vertical Expansion | Deploying the SBPI ontology template to new verticals (fintech, K-pop, sneakers, biotech). Each vertical gets its own namespace, its own companies, its own dimensional weights — same infrastructure. | Cross-vertical knowledge density. Patterns visible across verticals that no single-vertical analyst can see. The graph becomes a structural intelligence about how competitive landscapes work, not just what they contain. |
| Post-Training / Fine-Tuning | Fine-tuning small language models (SLMs) on the specific KG. Models that "think" like the best analysts because they're grounded in attestation-backed facts, not internet-scale noise. | Domain-specific models that don't hallucinate about micro-drama rankings because they retrieve from verified triples instead of guessing from parameters. |
| Phase | Investment | Duration | Unlock |
|---|---|---|---|
| Prove Accuracy | $70K-120K | Q2-Q3 2026 | 12+ weeks of prediction data, KG-augmented vs. baselines. First defensible accuracy claim. |
| Multi-Vertical | $50K-80K | Q3 2026 | 3 verticals running, cross-vertical pattern detection online. |
| SLM Fine-Tuning | $30K-50K | Q4 2026 | Domain-specific models that demonstrably outperform GPT-4 on vertical-specific tasks. |
| Licensing & API | $40K-60K | Q1 2027 | SPARQL API access, schema licensing, prediction signal feeds to third parties. |
Andrej Karpathy's "Software 2.0" insight: the most valuable code is not the code humans write — it's the data the system accumulates through iterative cycles of extraction, evaluation, and refinement. We apply this directly to ontology discovery.
sbpi.ttl) and a corpus of briefing dataThis cycle already runs in a simplified form. The nightly scheduler extracts signals, the SPARQL queries cross-reference against historical data, and the prediction experiment stack-ranks which signals have predictive value. The human finalization step happens when analysts review the weekly digest and decide which signals to act on. As the graph grows, the automated steps become more accurate and the human becomes a curator rather than a creator.
Each vertical gets its own "private brain" — a structured knowledge system that accumulates intelligence about competitive dynamics, market movements, and predictive signals. The $5K-25K weekly reports are subsidized R&D that pays for the construction of a database eventually worth millions in recurring licensing. The clients get intelligence. We get a compounding asset. Both sides of the transaction create value.
The prediction experiment is designed to produce the specific evidence an investor needs. Each week, four methods predict the same 17 companies. When actuals load, we score them all on the same metrics. Over time, the data answers three questions:
KG-augmented method vs. persistence and naive momentum. If KG-augmented wins, the graph is encoding real structural knowledge that simple statistics miss. This is the minimum viable proof point.
Plot KG-augmented accuracy over time. If the curve goes up as triples accumulate, the graph is compounding. If it flatlines, the schema needs revision. The temporal chain gives us a clean x-axis: weeks of data available.
Methods 5 (LLM zero-shot) and 6 (LLM + KG) directly compare. Zero-shot uses company names only. LLM + KG feeds the full semantic layer context. The delta between 5 and 6 is the measurable marginal value of the knowledge graph — the number investors care about.
| Method | W12 Predictions | Evaluations | Accuracy |
|---|---|---|---|
| Persistence | 17 locked | Awaiting W12 actuals | — |
| Naive Momentum | 17 locked | Awaiting W12 actuals | — |
| Mean Reversion | 17 locked | Awaiting W12 actuals | — |
| KG-Augmented | 17 locked | Awaiting W12 actuals | — |
| LLM Zero-Shot | Not yet active | — | — |
| LLM + KG | Not yet active | — | — |
First evaluation runs when W12-2026 actuals are loaded via python3 etl/sbpi_to_rdf.py.
The system improves itself through use. This is not a metaphor — it is a measurable property of the architecture.
Each client engagement generates scored company data → sbpi_to_rdf.py converts to triples → Oxigraph stores them → SPARQL queries detect signals → prediction experiment tests whether signals have predictive value → accuracy metrics reveal which parts of the ontology produce real intelligence → analysts refine the ontology based on what works → next engagement benefits from the refined schema + accumulated history.
The information doesn't just accumulate — it reorganizes. The ontology evolves. Dimensions get reweighted based on what predicts. New signal types get added when the graph reveals patterns that the original schema didn't anticipate. The structure IS the IP, and the structure improves with each cycle.
This is what separates the system from a data warehouse. A warehouse stores records. This system stores records and the schema for interpreting them, and the validation shapes for ensuring quality, and the prediction methods for testing whether the stored knowledge has real-world predictive value. The value comes from how the sources are organized — InfraNodus entity extraction, wikilink relationships, gap analysis, ontological bridging — not just from the sources themselves.
Force-directed layout of the SBPI knowledge graph analyzed via InfraNodus. 150 nodes, 700 edges, 8 clusters. Node size reflects degree centrality. Edge opacity reflects relationship weight. The five gateway nodes (data, prediction, knowledge, graph, schema) bridge multiple clusters and control information flow through the network.
Three structural gaps identified in the concept network. These represent missing conceptual bridges between clusters that, if filled, would strengthen the overall intelligence framework. Each gap is a content or research opportunity.
Eight topical clusters ranked by betweenness influence. The Predictive Accuracy cluster dominates at 39% of network influence, followed by Knowledge Density at 30%. Together these two clusters control nearly 70% of the network's information flow.
data — highest betweenness centrality, bridges 42 connectionsConceptual gateways control information flow between clusters. These nodes have high betweenness centrality and connect otherwise isolated topic areas.
| Node | Betweenness | Degree | Role |
|---|---|---|---|
data | 0.299 | 42 | Central hub connecting predictions, graphs, and infrastructure |
prediction | 0.254 | 40 | Bridges accuracy experiments to intelligence outputs |
graph | 0.249 | 45 | Links knowledge density to schema and experiments |
knowledge | 0.231 | 41 | Connects infrastructure to density and revenue concepts |
schema | 0.196 | 39 | Bridges ontology design to prediction methodology |
Strongest weighted edges in the network, representing the most frequently co-occurring concept pairs across the SBPI corpus.
| Source | Target | Weight |
|---|---|---|
data | predictions | 23 |
prediction | accuracy | 21 |
graph | data | 20 |
prediction | data | 17 |
prediction | experiment | 15 |
knowledge | graph | 14 |
accuracy | data | 13 |
knowledge | infrastructure | 13 |
The network's focused modularity (0.425) with five high-centrality gateways confirms the SBPI thesis structure: data and prediction form the operational core, knowledge and graph form the asset layer, and schema is the bridge between ontology design and predictive methodology. The three structural gaps all involve the Market Momentum cluster, which is conceptually isolated from the system's infrastructure — exactly where the next round of ontology design work should focus.