Every AI agent reads the same pages and rebuilds the same semantic understanding from scratch. There is no shared semantic layer. STP is that layer.
The modern web was designed for human eyes. When an AI agent needs to read a webpage, it reverses engineers meaning from a presentation layer never intended for machines. The waste is structural.
"The web was built for humans. Search engines retrofitted machine readability on top. STP asks a different question: what if we designed the data layer for agents first, and let humans have a translation on request?"
STP embeds a structured semantic block inside any webpage via a script tag. Browsers skip it. Agents parse it and skip the DOM entirely. Where HTML communicates presentation and JSON communicates data, STP communicates meaning.
8 relation types. Canonical concept IDs. Confidence scores with provenance chains. Designed to degrade gracefully — pages without STP fall back to HTML parsing. Pages with STP are just faster, cheaper, and structurally richer to consume.
Each layer is safe to ship without the next. The reading layer is inherently safe — nothing executes. The action layer requires a complete security specification first. The A2A protocol can only follow.
The RAG comparison is the most practically important — it's what agents actually use today for web reading. The 39× number is robust even with minimal STP blocks because typed structure is categorically richer than prose fragments.
| COMPARISON | SAVINGS | NOTE |
|---|---|---|
| STP vs raw HTML | 161× | Real but not the fair comparison |
| STP vs stripped text | 48× | Assumes well-authored STP block |
| STP vs RAG (5×512 chunks) | 39× — and more structured | Use this number. |
| MODEL | CONVENTIONAL (PAGES / CTX) | STP (PAGES / CTX) |
|---|---|---|
| Claude Sonnet 200K | 53 pages | 2,545 pages |
| Gemini 1.5 Pro 1M | 264 pages | 12,723 pages |
| BENCHMARK | CONVENTIONAL | STP | IMPROVEMENT |
|---|---|---|---|
| Task completion time | 9.27s | 0.42s | 21.8× faster |
| Bytes processed | 102KB HTML | 892 bytes | 116× less |
| LLM calls required | 1 | 0 | Eliminated |
| Selector breaks | 1 (CSS → XPath retry) | 0 | None possible |
| Crawler compression | 288KB HTML | 2,620 bytes STP | 112.5× |
Honest caveat: 48× assumes a well-authored STP block. Real-world numbers in early adoption are probably 20–35× vs extracted text, climbing toward 48× as tooling matures. $9/day saved at 1,000 pages/day at current Claude Sonnet pricing. $928/day at crawler scale (100K pages).
An agent reading an ML paper from 2023 needs to know that the field's confidence in that claim has since dropped. A static knowledge graph gives it the claim. STP's temporal layer gives it the claim, its current standing, and the event that caused the revision.
LLM causes EmergentBehavior relation weakens and changes type to relates_to.Every layer of STP is interactive and runnable. Not slides. Not mockups. Working code that demonstrates what the protocol actually does.
Every milestone documented as it ships. The honest numbers, the caveats, the security threats before the features.
12 interactive prototypes. Full security threat model. Benchmark vs conventional agents. 11 development posts. All open source from day one.