Layer 01: Model Reality
Layer 02: System Design
Layer 03: Reliability
Layer 04: Delivery
Blueprint
DepthPilot is not organized as a pile of topics. It is organized around what the learner can actually deliver. The knowledge network decides what to learn, course modes decide how to learn, and acceptance criteria decide whether the learner truly mastered it.
Knowledge Backbone
The knowledge network defines nodes, prerequisites, and mastery proof. The teaching blueprint translates those nodes into concept lessons, guided builds, project work, and assessments. You need both layers or the curriculum drifts into topic sprawl.
Open the knowledge networkLayer 01: Model Reality
Layer 02: System Design
Layer 03: Reliability
Layer 04: Delivery
Path 01: Understand the Model
Path 02: Build Reliable Workflows
Path 03: Ship the Product
Teaching principle
Every lesson begins with a concrete AI problem the user actually faces, instead of dropping abstract terminology first.
Teaching principle
A lesson is only complete when it produces an artifact such as a screenshot, a running page, a finished config, or an acceptance report.
Teaching principle
Core ideas are anchored in official docs or primary materials before we turn them into structured teaching content.
Teaching principle
Passing a quiz is not enough. The learner must be able to recreate the method in their own workflow.
Skill Paths
People who use AI frequently but still cannot explain why results vary so much
Upgrade from using AI to diagnosing and designing AI systems
Final Capability
You can tell whether the problem comes from prompts, context, data, tools, or evaluation.
Understand token, context, and eval constraints
Break giant prompts into structured context architecture
Build a minimum eval loop and improve the system with failures
People who want to follow a guide, configure tools fast, and actually get them running
Move from step-by-step setup to independent configuration, debugging, and delivery
Final Capability
You can configure OpenClaw, Supabase, and Creem on your own and verify that they really work.
Finish one full tool setup from zero
Understand key configuration points and common failure modes
Turn the setup result into part of a real project
People who want to turn AI capability into a real product or internal system
Connect concept lessons, guided builds, and project work into one product-building path
Final Capability
You can ship an AI product with auth, billing, content, learning data, and a trust layer.
Build a minimum working prototype
Add identity, billing, and data loops
Create content sourcing, review, and update mechanisms
Course Network
Foundations · Concept lesson · 24 min
Teach learners when the model should answer, when it should clarify, when it needs retrieval or tools, and when it should stop.
How We Teach
Start from a real confident-but-wrong failure
Explain the boundary between model ability, evidence, authority, and tools
Force the learner to rewrite one task as a decision ladder
User Outcomes
Separate fluent output from real reliability
Route tasks into direct answer, clarification, retrieval, tool use, and refusal paths
Turn one confident failure into a safer workflow boundary
Validation
Multi-question quiz
Task-routing exercise
Reflective rewrite
Deliverables
1 task-routing sheet
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners finally see the gap between sounding right and being reliable
The lesson creates an immediate sense of control over AI behavior
Foundations · Concept lesson · 26 min
Move learners from 'writing more detailed text' to 'designing an interface the system can actually verify'.
How We Teach
Start from a real failure where downstream systems had to guess meaning
Explain the distinct roles of framing, schema, and action execution
Force the learner to convert a free-form output into a contract
User Outcomes
Know when to use natural-language instruction, structured outputs, or function calling
Rewrite one real task as an output contract with clear fields and types
Understand why failures must be visible instead of silently guessed through
Validation
Multi-question quiz
Output-contract rewrite exercise
Reflective redesign
Deliverables
1 output contract draft
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners realize prompt design can be treated like interface design
The lesson produces a strong sense of system control
Foundations · Concept lesson · 18 min
Show why AI systems are constrained by token budget from the start.
How We Teach
Start with the real problem of prompts getting longer and worse
Explain the relationship between tokens and context
Give one workflow decomposition exercise
User Outcomes
Understand why token budget shapes product boundaries
Separate persistent information from on-demand injection
Start reading your workflow through a budget lens
Validation
Instant quiz
Reflection
Knowledge card capture
Deliverables
1 knowledge card
1 reflection
1 quiz result
Motivation Hooks
Learners realize they were stuffing prompts in the wrong place
They can map the lesson back to their real workflow immediately
Systems · Concept lesson · 22 min
Move learners from writing giant prompts to designing context architecture.
How We Teach
Show a failure caused by an oversized prompt
Break down a three-layer context structure
Ask the learner to design their own context split
User Outcomes
Separate system rules, task state, and live evidence
Diagnose whether a failure is a prompt issue or an architecture issue
Start rewriting giant prompts into structure
Validation
Instant quiz
Structured reflection
Workflow rewrite draft
Deliverables
1 context architecture draft
1 knowledge card
1 quiz result
Motivation Hooks
The cognitive reversal is strong: the problem is not the prompt, it is the architecture
The lesson can be applied to a real system immediately
Systems · Concept lesson · 29 min
Teach learners to treat retrieval sources as governed assets with owners, freshness classes, and expiry behavior instead of a one-time document dump.
How We Teach
Start from a sourced answer that was still wrong because the source was outdated
Break freshness into classes, metadata, ownership, and fallback behavior
Force the learner to govern one real knowledge source instead of discussing RAG in the abstract
User Outcomes
Separate retrieval quality from freshness and governance quality
Define owners, review cadence, and expiry thresholds for real knowledge sources
Diagnose whether a grounded answer failed because the evidence was stale, mixed, or unmanaged
Validation
Multi-question quiz
Freshness register draft
Reflective redesign
Deliverables
1 freshness register
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners realize that retrieval quality can collapse even when search technically works
The lesson immediately sharpens internal-doc assistants, policy bots, and support copilots
Systems · Concept lesson · 28 min
Teach the learner when evidence is required, how to retrieve it, and how to tie the final answer back to freshness and provenance.
How We Teach
Start from a confident answer that should never have been given without evidence
Explain evidence routing, freshness, and provenance as system responsibilities
Force the learner to turn one real question into a retrieval chain
User Outcomes
Separate retrieval design from simply making the context longer
Design a real workflow with query, filtering, injection, and citation steps
Diagnose whether a failure came from missing evidence, stale evidence, or noisy retrieval
Validation
Multi-question quiz
Retrieval-chain design exercise
Reflective redesign
Deliverables
1 retrieval-chain draft
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners feel the shift from trusting the model to controlling evidence
The lesson improves internal docs search, knowledge bases, and policy-answering workflows immediately
Systems · Concept lesson · 30 min
Teach learners to design tool workflows around routing, action boundaries, recovery order, and reusable skills instead of hoping the agent improvises safely.
How We Teach
Start from a workflow that feels magical until it touches a real external action
Break the chain into routing, authority, execution, and recovery responsibilities
Require the learner to rewrite one workflow as an action-ready operating design
User Outcomes
Separate model reasoning from tool execution and confirmation logic
Design one real workflow across clarify, retrieve, decide, act, verify, and handoff stages
Turn repeated operator behavior into a reusable SOP or skill
Validation
Multi-question quiz
Workflow-routing exercise
Reflective redesign
Deliverables
1 workflow-routing sheet
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners stop confusing tool access with system readiness
The lesson creates immediate leverage for agents, automations, and operator-facing AI systems
Evaluation · Concept lesson · 20 min
Upgrade from vague confidence to verifiable system improvement.
How We Teach
Start from real failures, not abstract benchmarks
Give a minimum eval loop template
Ask the learner to convert their failures into eval samples
User Outcomes
Collect real failure samples
Define a minimum eval set
Tie eval results to launch or rollback decisions
Validation
Instant quiz
Failure-sample collection task
Eval draft
Deliverables
1 minimum eval set draft
1 knowledge card
1 quiz result
Motivation Hooks
Learners finally see why launches regress so often
Eval loops create a clear sense of control
Evaluation · Concept lesson · 30 min
Build real performance and cost judgment by eliminating system waste before chasing model pricing alone.
How We Teach
Start from the real problem of demos that become slow and expensive in production
Explain critical path, caching, batching, async work, and graceful downgrade
Force the learner to audit one workflow for latency and cost
User Outcomes
Separate user-perceived latency from total system latency
Spot waste in duplicated requests, bloated context, oversized outputs, and missing caching
Draft a minimum latency and cost audit for one workflow
Validation
Multi-question quiz
Latency/cost audit exercise
Reflective redesign
Deliverables
1 latency/cost audit draft
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners finally see that much of the waste is not model price at all
The lesson can directly reduce spend and improve responsiveness in a real product
Evaluation · Concept lesson · 32 min
Teach learners to turn safety into boundaries, confirmation steps, and graceful failure instead of more warning text inside prompts.
How We Teach
Start from a real injection or unauthorized-action scenario
Explain trust boundaries, action boundaries, and graceful downgrade
Force the learner to run a guardrail audit on their own workflow
User Outcomes
Separate untrusted text, system rules, and real-world actions by trust level
Recognize prompt injection, prompt leak, and unauthorized-action paths
Design input isolation, confirmation, and downgrade behavior for one real workflow
Validation
Multi-question quiz
Guardrail-audit exercise
Reflective redesign
Deliverables
1 trust-boundary draft
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners stop trying to merely 'make the model obey'
The lesson has immediate value for agents, knowledge workflows, and tool-using systems
Evaluation · Concept lesson · 32 min
Teach learners to route requests by value, risk, evidence need, and budget while preserving the right to clarify, retrieve, abstain, or escalate.
How We Teach
Start from the cost and reliability damage caused by sending everything through one answer path
Break routing into task classes, thresholds, and fallback logic
Require the learner to write one explicit unsupported-answer policy
User Outcomes
Design a routing matrix for different task classes instead of defaulting to one model path
Define when the system should answer directly and when it should not answer yet
Judge routing policy with evals, abstention quality, latency, and cost
Validation
Multi-question quiz
Routing-policy draft
Reflective redesign
Deliverables
1 routing matrix draft
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners stop confusing strong products with always-answer products
The lesson creates immediate leverage for support bots, copilots, and internal tools
Evaluation · Concept lesson · 30 min
Teach the learner how to record, replay, and label bad runs so they can localize the broken layer instead of guessing.
How We Teach
Start from a production failure the team cannot reproduce
Explain the roles of traces, replay, and failure labels
Force the learner to redesign one bad case as a replayable debugging record
User Outcomes
Know what a minimum useful trace must contain
Replay failures before deciding whether to change prompting, retrieval, tool use, or orchestration
Turn bad runs into assets for later evaluation and prioritization with failure labels
Validation
Multi-question quiz
Trace-template design exercise
Reflective redesign
Deliverables
1 minimum trace template
1 knowledge card
1 multi-question quiz result
Motivation Hooks
Learners feel the shift from guessing at failures to locating them
The lesson immediately improves real agent and workflow debugging practice
Evaluation · Concept lesson · 31 min
Teach learners to replace vague quality judgments with dimensions, anchors, thresholds, and grader rules another operator can reuse.
How We Teach
Start from a workflow that everyone says is improving but nobody can score clearly
Break evaluation into named dimensions, anchors, and override rules
Require the learner to build one rubric that another reviewer could actually apply
User Outcomes
Turn one abstract quality goal into a rubric with dimensions and score anchors
Keep dimension scores and hard-stop rules instead of relying on one total score
Use rubric evidence to decide what to fix first
Validation
Multi-question quiz
Rubric draft
Reflective redesign
Deliverables
1 scoring rubric
1 grader note
1 multi-question quiz result
Motivation Hooks
Learners stop treating quality as taste
The lesson creates an immediate bridge from eval ideas to launch and rollback decisions
Evaluation · Concept lesson · 30 min
Teach learners to define hard stops, review-queue ownership, and handoff packets so the system stops before unsupported answers or unsafe actions compound.
How We Teach
Start from a workflow that currently over-answers because nobody wants to escalate
Break the problem into hard-stop triggers, ownership, SLA, and handoff evidence
Force the learner to design one live review queue instead of talking about human-in-the-loop in slogans
User Outcomes
Define escalation triggers for risk, authority, evidence, and policy-sensitive requests
Design a handoff packet another human can act on immediately
Treat escalation as a quality path instead of product embarrassment
Validation
Multi-question quiz
Escalation policy draft
Reflective redesign
Deliverables
1 escalation policy
1 handoff packet
1 multi-question quiz result
Motivation Hooks
Learners feel the shift from answer pressure to operational judgment
The lesson immediately improves trust in support, approval, and action-taking workflows
Delivery · Guided build · 55 min
Run checkout, webhook, customer portal, and in-app entitlement as one chain.
How We Teach
Set up product, portal, and webhook in test mode first
Map them to the app routes and sync logic
Validate with in-app entitlements and DB state
User Outcomes
Create a test product and wire env vars correctly
Run Creem Checkout, Portal, and local webhook forwarding
Understand why you cannot trust success_url alone
Validation
Payment flow screenshots
Webhook self-check
Subscription row check
Entitlement verification
Deliverables
1 complete billing chain
1 billing verification checklist
1 webhook troubleshooting recap
Motivation Hooks
Learners quickly see payment become real product access
Billing, entitlement, and product behavior finally connect
Delivery · Guided build · 45-60 min
Get OpenClaw running step by step with real validation instead of guessing.
How We Teach
State the final artifact, environment requirements, and common failures up front
Drive every step with a checklist
Define success criteria and troubleshooting hints for each step
User Outcomes
Finish environment prep, config fill-in, and startup verification
Understand what 3 to 5 critical config items actually control
Know what to check first when things fail
Validation
Checklist complete
Runtime screenshots
Minimum troubleshooting recap
Deliverables
1 working OpenClaw environment
1 set of screenshots
1 troubleshooting note
Motivation Hooks
The learner sees the tool actually running inside one lesson
The feedback loop is much stronger than passive reading
Delivery · Guided build · 50 min
Build the full chain from database and auth to live page state.
How We Teach
Show the end state first
Then wire tables, env vars, and helpers step by step
Finish with post-login page behavior
User Outcomes
Create user tables and RLS rules
Get sign-in, sign-out, and session refresh working
Understand why auth cannot live only in the frontend
Validation
Runtime screenshots
Auth flow self-check
Config explanation
Deliverables
1 working auth system
1 self-check checklist
Motivation Hooks
The page-state change is immediately visible
The result can be reused in a real product right away
Delivery · Assessment · 35 min
Audit one retrieval workflow for freshness classes, ownership, metadata, and stale-content handling before it quietly ships old truth as current truth.
How We Teach
Start from one workflow that looks grounded but is vulnerable to stale evidence
Drive the audit through source class, metadata, ownership, and fallback behavior
Require artifacts another operator could use in launch review
User Outcomes
Produce a freshness register that classifies source types and expiry thresholds
Define document-governance rules for version, owner, and approval status
Create a stale-content triage path for clarify, refresh, or escalate decisions
Validation
Freshness register
Governance review
Stale-content triage review
Audit recap
Deliverables
1 freshness register
1 refresh policy
1 stale-content triage sheet
Motivation Hooks
Learners stop confusing retrieval with trustworthy knowledge
The audit immediately improves policy bots, internal search, and support assistants
Delivery · Assessment · 40 min
Audit one real workflow and turn vague safety concerns into a trust-boundary map, confirmation matrix, and containment plan.
How We Teach
Start with the threat-model order before touching prompts
Drive the review with an audit ladder and live red-team attempts
Require a signed-off artifact instead of a discussion-only outcome
User Outcomes
Map trusted and untrusted content instead of treating all text as equal
Inspect where tool actions can be triggered by risky inputs
Produce a real pre-launch guardrail review report
Validation
Audit checklist
Artifact review
Red-team evidence
Recap review
Deliverables
1 trust-boundary map
1 action-confirmation matrix
1 guardrail review report
Motivation Hooks
The learner stops hoping the model behaves and starts designing containment
The review feels like product control, not abstract safety talk
Delivery · Assessment · 35 min
Audit one workflow into a real escalation path with hard stops, queue ownership, SLA, and a handoff packet so risky or unsupported cases stop cleanly.
How We Teach
Start from one workflow that still answers when it should stop
Drive the review through escalation triggers, owner, SLA, and packet design
Require deliverables that make escalation auditable instead of hand-wavy
User Outcomes
Define hard-stop triggers that force escalation instead of unsupported answers
Assign queue ownership and review expectations
Produce a handoff packet template that preserves the evidence a human needs
Validation
Escalation-policy review
Queue review
Handoff-packet review
Audit recap
Deliverables
1 escalation policy
1 review queue scorecard
1 handoff packet template
Motivation Hooks
Learners stop treating escalation as embarrassment
The lab immediately improves trust in support, approval, and action workflows
Delivery · Assessment · 40 min
Audit a real workflow for request waste, context bloat, caching, async opportunities, and budget tradeoffs before changing models.
How We Teach
Establish a baseline before discussing optimizations
Use an audit ladder that starts with waste, not model swaps
Require a reusable report the learner can take back to their product
User Outcomes
Separate user-perceived latency from total background work
Expose waste in request count, prompt size, output size, and retrieval volume
Produce a ranked optimization plan with a performance budget
Validation
Audit checklist
Artifact review
Baseline evidence
Optimization recap
Deliverables
1 request inventory
1 performance budget report
1 ranked optimization backlog
Motivation Hooks
Learners usually discover the biggest waste is not where they assumed
The audit translates directly into cost and speed wins in a real product
Delivery · Assessment · 35 min
Turn one fuzzy AI step into a contract with explicit schema, failure states, and downstream acceptance checks.
How We Teach
Start from a broken downstream integration instead of prompt-writing tips
Use a contract ladder from framing to schema to acceptance checks
Require a spec and checklist that another operator could review
User Outcomes
Separate free-form reasoning from data the rest of the system must trust
Define fields, enums, null behavior, and explicit failure paths for one real task
Produce a contract spec that engineering or ops can actually implement
Validation
Schema checklist
Artifact review
Failure-path review
Replay recap
Deliverables
1 output contract spec
1 schema review checklist
1 accepted failure-state policy
Motivation Hooks
Learners feel the shift from prompting vibes to interface control
The artifact is immediately reusable in product and automation work
Delivery · Assessment · 40 min
Audit one evidence-dependent workflow for retrieval scope, freshness, provenance, and unsupported-answer handling.
How We Teach
Start from an answer that should have been grounded but was not
Run the audit through evidence routing, freshness, and citation checks
Require a report that could be used in a launch review
User Outcomes
Map where the system should retrieve, cite, abstain, or escalate
Spot stale evidence, noisy retrieval, and missing provenance before launch
Produce a retrieval review report and evidence-chain checklist for a real workflow
Validation
Evidence-chain checklist
Artifact review
Citation proof
Audit recap
Deliverables
1 retrieval review report
1 evidence-chain checklist
1 unsupported-answer policy
Motivation Hooks
Learners stop treating retrieval as a vague RAG buzzword
The lesson creates immediate leverage for docs QA and internal knowledge systems
Delivery · Assessment · 35 min
Audit one workflow for task classes, model-path choices, fallback thresholds, and explicit unsupported-answer behavior before it reaches users.
How We Teach
Start from one workflow that currently over-answers or over-spends
Drive the review through task classes, routing thresholds, and unsupported-answer handling
Require policy artifacts that another operator could enforce and review
User Outcomes
Define a routing matrix that matches task value, risk, evidence need, and budget
Write an unsupported-answer policy that covers clarify, retrieve, abstain, and escalate paths
Produce a fallback ladder instead of letting the system improvise under uncertainty
Validation
Routing matrix
Policy review
Fallback ladder review
Audit recap
Deliverables
1 model-routing matrix
1 unsupported-answer policy
1 fallback ladder
Motivation Hooks
Learners feel the shift from prompt optimism to operational policy
The audit immediately improves cost, trust, and failure handling in real products
Delivery · Assessment · 35 min
Turn one workflow into a scoreable review system with dimensions, anchors, hard-stop rules, and grader instructions another reviewer can reuse.
How We Teach
Start from a workflow currently judged by gut feel
Drive the audit through dimensions, anchors, hard stops, and calibration
Require artifacts that make quality reviewable by a second operator
User Outcomes
Write one rubric with dimensions and scoring anchors for a live workflow
Define grader instructions and hard-fail rules instead of relying on overall impressions
Leave with a calibration sheet that makes scoring disagreements visible
Validation
Rubric review
Grader-spec review
Calibration review
Audit recap
Deliverables
1 scoring rubric
1 grader spec
1 calibration sheet
Motivation Hooks
Learners stop arguing in vibes
The lab creates immediate leverage for launches, regressions, and trace review
Delivery · Assessment · 35 min
Audit one tool-using workflow for routing order, confirmation gates, recovery steps, and the operator logic that should become a reusable skill.
How We Teach
Start with one workflow that currently feels agentic but is still fragile
Drive the audit through routing order, tool boundaries, and recovery order
Require a reusable artifact that survives beyond one careful operator
User Outcomes
Map the workflow into explicit clarify, retrieve, act, verify, and handoff stages
Define which actions need evidence, which need confirmation, and which should never auto-run
Produce a routing sheet, a tool-boundary checklist, and one operator skill brief
Validation
Routing sheet
Checklist review
Skill brief review
Audit recap
Deliverables
1 workflow-routing sheet
1 tool-boundary checklist
1 operator skill brief
Motivation Hooks
Learners feel the shift from agent theater to operational control
The artifacts immediately improve internal automations and tool-using teams
Delivery · Project · 2-4 h
Turn the concept lessons and build lessons into one finished deliverable.
How We Teach
Define project scope and acceptance criteria
Advance in stages across content, auth, billing, data, and trust
Require a final demo and recap
User Outcomes
Build the complete product loop independently
Explain your architecture choices
Show a real working result
Validation
Project acceptance
Artifact review
Recap review
Deliverables
1 online or local demo
1 architecture note
1 project recap
Motivation Hooks
The learner leaves with a product, not a notebook
This creates the strongest sense of progress and shareable output
Search Cluster
Search pages should not be traffic bait. They should route users from search vocabulary into a real learning and delivery path.
Prompt Engineering Course
This page targets users who really search for a prompt engineering course, but DepthPilot does not reduce the topic to prompt hacks. It puts prompting back into context architecture, workflow design, and eval loops.
Open pathAI Workflow Course
If the user searches for an AI workflow course, they usually need more than model theory. They need to connect AI into real workflows, tools, access control, and delivery standards.
Open pathOpenClaw Tutorial
This entry page aligns directly with the OpenClaw tutorial search intent. It shows the learner what they will actually gain before sending them into the full guided build, skills page, and project path.
Open pathAI Eval Checklist
Users searching for an AI eval checklist usually do not lack opinions. They lack an executable review frame. This page condenses the minimum eval logic into a checklist-style entry point.
Open pathHuman in the Loop AI
Many people searching for human-in-the-loop AI only want to know whether humans should review output. DepthPilot pushes further: when must the system stop, who owns the queue, and what evidence must travel with the case?
Open pathRAG Freshness Governance
Many teams treat RAG as 'it can search documents now', then assume the system has reliable knowledge. DepthPilot asks the harder questions: who owns the documents, when do they expire, how are versions governed, and what happens when freshness cannot be trusted?
Open pathLLM Evaluation Rubric
Many people searching for an LLM evaluation rubric only want a template. DepthPilot goes further: we turn rubric design into dimensions, anchors, hard-stop rules, and grader instructions that help you decide what broke and what to fix first.
Open pathPrioritize concept and guided-build lessons for the seeded nodes so every path can move from understanding to delivery. Do not chase volume before path completion.