Content Factory
PremiumBuild your first automated pipeline: from prompts to JSON
Content Factory: Build Your Automated Production Line
"Factory efficiency does not depend on how fast workers run, but how smooth the line is designed."
What you will get in this chapter
- A full breakdown of a minimum viable content factory (MVS)
- Structured standards for inputs and outputs
- Practical metrics for QA thresholds and cost control
One-sentence definition
Content factory = structured input + standardized output + reusable production chain.
It is not "AI writes articles". It is AI continuously producing usable content.
Minimum viable content factory (MVS)
| Step | You need | Acceptance result |
|---|---|---|
| Input | 50 keywords + 1 template + 1 data sheet | Stable structured input |
| Production | 3-step prompt chain | Every output can be rendered |
| QA | Rule-based validation | Retry on fail, publish on pass |
| Storage | JSON/MDX files | Reuse across channels |
Qualified signal: out of 100 items, failure rate < 5%, and all can be rendered on the web.
Four stations of the factory
- Input standardization: turn keywords and data into structured tables.
- Strategy/outline: generate reusable outlines and modules.
- Content generation: generate in parts to avoid "long-form drift".
- Quality and publishing: validate + fix + archive.
Input standardization: turn raw material into tables
You must standardize "keywords" and "business data", or the factory becomes manual labor.
Minimum input fields:
keyword: core termintent: search intent (tutorial/comparison/tool)audience: target audiencepriority: importance
{
"keyword": "best ai tools for designers",
"intent": "list",
"audience": "designer",
"priority": "high"
}Output standardization: JSON as the single deliverable
If AI outputs JSON, you can render automatically and reuse across channels.
{
"meta": {
"title": "Best AI Tools for Designers",
"slug": "ai-tools-designers",
"summary": "A curated list of tools for designers"
},
"sections": [
{ "type": "intro", "text": "..." },
{ "type": "list", "items": ["..."] },
{ "type": "faq", "items": [{ "q": "...", "a": "..." }] }
]
}Prompt chain design: split big problems into small steps
Do not use one prompt for everything. Split into 3 steps:
- Generate structure: title, modules, bullet points
- Generate in parts: write section by section to avoid drift
- Format and QA: output valid JSON and check missing fields
This makes issues easier to debug and lowers the failure rate.
Core metrics (must track)
Definition (default):
- Time window: unless stated otherwise, use the last 7 days rolling.
- Data source: use one trusted source (GA4/GSC/platform console/logs) and keep it consistent.
- Scope: only the current product/channel, exclude self-tests and bots.
| Metric | Meaning | Pass line |
|---|---|---|
| Structural completeness | Required fields present | >= 98% |
| Duplication rate | Similarity to related content | <= 60% |
| Readability | Does the opening hit the problem | Value in first 3 paragraphs |
| Fact risk | Are claims verifiable | Key data traceable |
No threshold, no factory.
Principle: if it fails, do not publish. Slow is fine; garbage is not.
Cost control: spend tokens where they matter
Cost per item estimate:
cost = (avg_tokens * generations * unit_price) / 1000Three ways to cut cost:
- Reuse modules: semi-templatize Intro/FAQ blocks
- Caching: reuse outlines instead of regenerating every time
- Retry only failures: re-run only failed sections, not the whole item
Acceptance checklist
Common mistakes
- No structured output -> cannot reuse or auto-render at scale
- Not differentiated -> pages flagged as duplicate content
- No quality threshold -> garbage piles up, indexing drops
Summary
Next chapter: Automated Distribution - deploy content like code and keep it in sync.
AI Practice Knowledge Base