I added native image generation to SystemSculpt because visual production kept slowing down otherwise strong publishing systems. Writing moved fast, but image creation still required too many tool switches and handoffs.
This workflow closes that gap directly inside Obsidian so a single operator can move from concept to production asset with less friction.

Why I shipped this inside the vault
I wanted one environment where I could:
- Draft content strategy.
- Generate image variants.
- Store approved assets with stable paths.
- Ship Markdown with embedded visuals without context switching.
That is the exact operating problem this feature solves. It keeps ideation, production, and publication in one execution loop.
What the feature does in practical terms
- Generate images from prompt templates tied to note context.
- Produce multiple variants in one run for faster selection.
- Save outputs directly into vault folders used by publish workflows.
- Preserve naming and folder discipline so assets remain searchable.
- Reduce the handoff overhead between writing and visual production.

Setup checklist before first run
I recommend this exact setup before generating production images:
- Confirm your plan includes image generation in SystemSculpt Pro.
- Review baseline feature behavior in the image generation docs.
- Configure provider and defaults in settings docs.
- Confirm model availability and limits in model provider docs.
- Create a dedicated vault folder for generated assets.
- Decide naming conventions before your first batch.
That setup prevents most early quality and organization issues.
The production workflow I use
1. Define visual intent inside the content note
Before I prompt anything, I define the asset role:
- Hero image
- Section explainer visual
- Social preview
- Conversion support graphic
I treat "role" as non-negotiable context. Without role clarity, prompt quality drops fast.
2. Generate one concept set with controlled variation
I generate one concept and request 3 to 5 variants. This gives enough optionality without creating review overload.
I keep variation focused on:
- Composition
- Lighting
- Subject emphasis
- Color temperature
I avoid changing every variable at once. Controlled variation makes selection more objective.
3. Run a hard quality gate
I reject anything that fails legibility, relevance, or consistency with page intent.
My quick gate:
- Clear focal point on first glance.
- Composition survives mobile crop.
- No accidental text artifacts.
- Style aligns with the surrounding page.
- File fits the intended slot ratio.
4. Save with stable slugs and deterministic folders
I store approved assets with a predictable path pattern so embeds do not break later:
/assets/images/
/blog/
/<post-slug>/
hero-v1.png
section-workflow-v1.png
social-preview-v1.png
I avoid generic names like image1.png. Clear naming keeps review and updates fast.
5. Embed immediately in Markdown and validate
I place the chosen image in the note right after selection. Immediate embedding catches layout or context issues while the decision is still fresh.

Prompt framework I use for repeatable quality
I use a simple, structured prompt skeleton:
Create a <aspect-ratio> visual for <asset-role>.
Context: <topic + audience + conversion intent>.
Visual direction: <style anchors, composition, lighting>.
Constraints: no text overlay, strong focal point, web-safe contrast.
Return <N> variants with controlled composition differences.
Example fill
Create a 16:9 visual for a blog hero image.
Context: AI workflow implementation for Obsidian operators evaluating automation.
Visual direction: clean desktop workspace, data flow motif, high clarity, neutral palette.
Constraints: no text overlay, one dominant focal element, high contrast for readability.
Return 4 variants with composition changes only.
The key is consistency. I treat prompts like reusable production assets, not one-off experiments.
Implementation details that materially improve output
Keep ratio-specific templates
I maintain separate templates for 16:9, square, and vertical output. Ratio-specific templates reduce crop pain and preserve composition integrity.
Separate exploration from production
I run exploratory prompts in a scratch note, then move approved prompt patterns into a production template note. That keeps published workflows clean.
Version prompts and images together
When I increment an image version, I also version the prompt variant. This gives me clear traceability when performance or quality changes.
Align model choice with task type
I use model settings for intent, not novelty:
- Editorial hero: clarity and composition stability.
- Diagram-like visual: shape consistency and clean edges.
- Campaign visual: emotional contrast and focal urgency.
Pitfalls I see and how I avoid them
Pitfall 1: Prompt drift across a single project
Inconsistent prompt structure creates inconsistent visual language.
I prevent drift by locking a template per project and only changing one parameter group at a time.
Pitfall 2: Over-generation without review discipline
Generating too many options wastes time and blurs decision quality.
I cap most runs at 4 variants, then iterate from the best candidate.
Pitfall 3: File chaos after publishing pressure
Assets become hard to reuse when folder and naming rules are undefined.
I enforce stable folder rules before first publish and reject unstructured saves.
Pitfall 4: Ignoring downstream crop behavior
Desktop-perfect images can fail on mobile cards and social previews.
I check crops for each target slot before final approval.
Pitfall 5: Style mismatch with conversion intent
A visually impressive image can still hurt conversion if it conflicts with page purpose.
I tie every visual decision back to the specific page CTA and audience stage.
QA checklist before publish
Run this checklist for every production asset:
- Asset role is documented in the note.
- Final image passes legibility on desktop and mobile.
- File name includes role and version.
- Embed path resolves correctly in Markdown.
- Visual style matches the page goal.
- Backup variant is stored for fast replacement.

Window-captured execution proof inside Obsidian
This is a direct window capture of the Obsidian app workflow, not a fullscreen desktop recording. I use this kind of recording to verify the exact plugin interaction path before publishing docs or campaign content.

Conversion paths based on your operating mode
I recommend one of two paths depending on execution needs.
Path A: Self-serve implementation inside the vault
If you want to build and iterate internally:
- See plans and feature access
- Read full docs
- Start with image generation docs
- Tune model/provider setup
Path B: Guided rollout for business operations
If image operations are part of a larger AI implementation initiative:
7-day activation plan I recommend
- Day 1: Configure settings, provider, and folder structure.
- Day 2: Build three ratio-specific prompt templates.
- Day 3: Generate and approve one hero asset for a live page.
- Day 4: Add section visuals to one long-form post.
- Day 5: Publish and track qualitative engagement feedback.
- Day 6: Refine prompt template based on observed quality gaps.
- Day 7: Package the workflow as a repeatable internal SOP.
Strong next step
If you want immediate access and self-serve velocity:
If you need a governed implementation path tied to business outcomes: