DEV Community

Ken Deng
Ken Deng

Posted on

Training AI to See Your Client’s Scribbles: Beyond "Make It Pop"

Freelance designers know the drill: a client emails back a marked-up PDF with “move this” and a vague arrow. You spend 20 minutes deciphering the scribble, then manually update your files. This revision chaos eats your profit and time. The promise of AI automation often falls flat here, because generic systems can’t interpret visual feedback.

The V-F-C Framework: Context is Your AI’s Compass

The core failure is treating feedback as text-only. Comments like “unbalanced” or “make it pop” are aesthetic judgments lacking technical instruction. To train an effective system, you must provide structured context. Think of it as a three-part anchor: Visual (V), Feedback Type (F), and Context (C).

  • Visual Anchor (V): V:logo_top_right. This tells the AI what element in the image is being referenced.
  • Feedback Type (F): F:position_shift. This classifies the action needed, often deduced from the client’s markup (an arrow = move, a red X = remove).
  • Context (C): C:vs_v2, C:brand_guideline_pg3. This specifies which version is being critiqued and any relevant reference material.

A tool like Claude.ai can be purposefully used here. Its multimodal input allows you to upload the client’s marked-up screenshot alongside your structured text prompt, guiding it to interpret visual cues against your defined anchors.

Mini-Scenario: A client highlights a color swatch on a PDF and writes “too bright?” Your AI, trained with V-F-C, identifies the visual anchor (V:primary_button), classifies the feedback (F:color_change), and references the context (C:brand_guideline_pg3) to suggest a compliant alternative.

Implementing Your Feedback Interpreter

  1. Define Your Anchors First. Before training, catalog common elements (V:) and feedback actions (F:) for your projects. Create a simple reference sheet.
  2. Engineer Instructional Prompts. Your system prompt must instruct the AI to analyze the visual markup first, then cross-reference it with your provided V-F-C data. It is not a question; it is a command.
  3. Resolve Ambiguity Explicitly. For every comparative comment (“use the spacing from the desktop mock”), explicitly link the versions (C:from_desktop_v1) in your prompt to eliminate pronoun confusion.

The key takeaway is that AI automation for revisions requires you to build a contextual framework. By moving beyond raw text parsing and teaching your system to see feedback through structured Visual, Feedback, and Context anchors, you turn vague client markups into actionable, trackable change requests. This shifts automation from a broken promise to a practical, time-saving asset.

Word Count: 498

Top comments (0)