Builder Prize · Submission

Operating Model Deployment Kit

Switching LLMs used to mean starting over. Memories sticking halfway, decisions buried in walls of generated text, copy-pasting yesterday's best work between chats. Now the operating model travels with the work. Same hierarchy in every system.

BuilderSolo · MarTech
SurfaceClaude · Cursor · Hyperagent · any LLM
LiveMarTech, ongoing
VersionV1 · 2026-05-04

What changed.

Prompts losing context between sessions. Memories sticking halfway. Copy-pasting "best work" examples between chats so I don't lose them. Decisions and silent debt buried under walls of generated text. Endless revision cycles to drag the work up to the level it needs to be. And worst, having to switch LLMs and start over. Every builder I knew was running their own version of these workarounds. So I built a way through.

The kit is a set of files any adopter can drop in. Answer a few setup questions to tailor it to your context. After that, you work as usual. When something material changes (a new manager, a new AI model, a new regulation), the kit revisits itself and updates only what shifted.

The shift

From scattered notes to named hierarchy.

Loose context per tool became content with a place: top-level principles, mid-level decisions, per-instance tweaks. Each piece names its own level so any LLM reads it at the right weight.

Whose work changes

Operator first, then the team, then the org.

Solo builders get session-to-session compounding. Teams share the substrate so context survives the person who created it. Org-direction conversations get a common starting vocabulary across functions.

How it runs

Modular files in whichever tool you use.

The kit drops into Claude memory, Cursor rules, Hyperagent skills, plain markdown. The same hierarchy is present in every location. Hand any single file to any LLM and it carries its own meaning.

43
Skills
11 kit-native + 32 bundled
14
Templates
One per layer + governance starters
27
Diagrams
Architecture, phases, mechanisms, roll-ups
10
Setup questions
One kit, many adopters

The structure travels. Same hierarchy in Claude, Cursor, Hyperagent, plain markdown. Every system reads the content at the right level.

How it's built.

v1 made you answer 10 setup questions before the kit would do anything. That kept adopters from skipping the basics, but it turned the kit into a one-shot setup tool. Then my uncle asked for help with his app. He'd already hired someone, watched scope creep, and was thousands of dollars in with nothing to show. I'm resource-constrained too. I couldn't sit with him to rebuild. But I could give him my process. He kept hitting situations the setup hadn't covered. The kit needed to keep asking questions as he went, not just at the start. v1.1 redesigned it: setup questions land first, follow-ups arrive as new context surfaces, and targeted refreshes fire when something significant shifts (a new manager, a new AI model, a new regulation, a new task). The kit gets stronger the more you use it, not just on day one.

What's in the box

43 skills, 14 templates, 8 decision records, 27 diagrams.

Each skill covers a specific task: deploying the kit, validating an output, recording a decision. Templates give you the right shape for each layer. Decision records hold the kit's own design rationale on file. Diagrams show the parts and how they connect.

Agree on mature, drive toward it

Less prescribing, more auto-approvals.

The kit names what "mature" looks like up front and iterates toward it, asking questions along the way. When evidence is there, it moves on. When it isn't, refusal kicks in: status checks need evidence, release-readiness needs an evaluation baseline, reproducibility needs the model version, prompt, tools called, and data snapshot all logged. Best practices reinforce as you go, instead of getting relitigated each session.

The kit follows its own rules

The kit is its own worked example.

The same patterns the kit teaches (setup, decisions on file, status checks, refreshes when context shifts) are the ones the kit applies to itself. You don't take the structure on faith. The kit is the demo.

Take what fits

You don't have to adopt the whole thing.

Take a layer template. Borrow the decision-record format. Copy the maturity-scorecard. Or hand any single file to any LLM and ask it to extract what fits. Every file carries its own structure, so partial use is valid.

The kit isn't an alternative to Claude memory or Cursor rules or Hyperagent skills. It's the SHAPE of the content you put into them. Same hierarchy. Every system reads it at the right level.

What's new about it.

The reimagining wasn't "add AI to my notes folder." It was that loose context per tool became content with a place: a top-level principle, a recorded decision, a per-instance tweak. Each piece names its own level. The same hierarchy lands inside every system you load it into. Wherever the content travels, every system reads it at the right weight.

The technique to steal

Naming + hierarchy + supersession is a translation contract.

Any rule, decision, or pattern documented this way travels across LLMs. Drop the same files into Claude memory, Cursor rules, Hyperagent skills, or plain markdown. Every system reads them the same way. The structure lives in the content, not the tool. If your function runs across multiple AI surfaces, the technique transfers.

What this makes possible

Leverage compounds across people, tools, and time.

Solo operators compound their own work session by session. Teams share the substrate so context survives the session and the person. Org-direction conversations start from a common vocabulary across functions. AI gives individuals leverage. The kit makes that leverage compound. First for the operator. Then the team. Then the organization.

"The best way out is always through."

For anyone getting started building, and especially those who've already got their feet wet and felt the loose ends: try it, tell me what breaks, and let's mature it into something stronger.