Skip to main content
Katie Academy

Multimodal Workflow Lab

Intermediate18 minutesLesson 5 of 5

Progress saved locally. Sign in to sync across devices.

Learning objectives

  • Combine several ChatGPT input modes in one coherent task.
  • Choose the role of each input before you start.
  • Keep the final output grounded even when the workflow is rich.

The real power of multimodal work is not novelty. It is composition. When documents, spreadsheets, screenshots, and generated visuals each have a defined role, ChatGPT can help you move through a richer workflow without constantly re-explaining the context.

Show an end-to-end flow from uploaded brief, spreadsheet, screenshot, and final memo.

What you'll learn
  • How to assign a role to each input type.
  • How to sequence a multimodal workflow.
  • How to turn multimodal context into one useful deliverable.
Why this matters

Many practical tasks are naturally multimodal: a policy document plus a spreadsheet, a screenshot plus a memo, a data table plus a visual mockup. Handling them in one workflow can save time and reduce context switching.

But multimodal work also raises the risk of confusion. Clear sequencing and output design matter even more when several inputs are involved.

This lesson matters because multimodal workflows represent the highest-leverage use of ChatGPT in professional work. Most real tasks do not live in a single input format. A product decision involves data, documents, and visual artifacts. A client deliverable draws on meeting notes, spreadsheets, and screenshots. When you can orchestrate these inputs effectively, ChatGPT becomes a workflow tool rather than a question-answering tool. But without clear structure, the same richness of inputs produces muddled, shallow output.

The core idea

A strong multimodal workflow names the role of each input. The document may define the problem. The spreadsheet may provide supporting data. The screenshot may reveal the current interface or workflow pain point. The final output may be a memo, table, or recommendation note.

The reason role assignment matters is that multiple inputs create exponentially more ambiguity than a single input. When you upload one file, the model has to guess one job. When you upload three files, the model has to guess three jobs and how they relate to each other. Without explicit roles, the model often treats each input independently, producing separate summaries instead of an integrated analysis. Or it over-emphasizes the most text-heavy input and underweights the image or the data, simply because language is its most natural modality.

The sequence matters too. Usually you want to understand the source materials first, then analyze, then synthesize. Trying to jump straight to a polished conclusion can create shallow output even when the inputs are rich. A staged approach also lets you verify the model's understanding of each input before asking it to combine them. If the model misreads the spreadsheet in step one, you can correct it before that error propagates into the final synthesis.

How it works

  1. Choose the final artifact first so the workflow has a target.
  2. Assign a role to each input: context, evidence, comparison, or visual reference.
  3. Run the workflow in stages and ask ChatGPT to preserve the chain from input to conclusion.
  4. At each stage, verify the model's understanding before moving to synthesis.

What skilled users do differently

A novice dumps multiple files into a single message and asks for help. The model does its best to make sense of the pile, but the result is usually a surface-level response that touches each input without connecting them meaningfully. The user gets a long answer that feels thorough but lacks the analytical depth that comes from directed composition.

A skilled user designs the workflow before starting the conversation. They know what the final artifact should be: a memo, a recommendation, a comparison table, a brief. They work backward from that artifact to determine which inputs serve which role. Then they sequence the conversation so the model processes each input in the right order. They might start by having the model read the policy document and extract the relevant sections, then upload the spreadsheet and ask for analysis against those extracted requirements, then upload the screenshot and ask whether the current interface reflects the policy correctly.

Skilled users also use the conversation thread strategically. Because ChatGPT retains context within a conversation, each stage builds on the previous one. But this means errors in early stages propagate. So skilled users verify at each step: "Before we continue, confirm your understanding of the three key obligations from the policy document." That checkpoint prevents a misreading from corrupting the final output.

Two worked examples

Example 1: a vague request

Use this document, spreadsheet, and screenshot to help me.

This prompt is weak because it gives no input roles, no analytical direction, and no target artifact. The model will attempt something helpful, but without knowing which input matters for what purpose, the output will be a loose summary of each file rather than an integrated analysis. It is the multimodal equivalent of saying "do something with these."

Example 2: a structured multimodal analysis

I want to run a multimodal workflow.

Inputs:
- Document: defines the onboarding process and current policy
- Spreadsheet: shows drop-off by stage and customer segment
- Screenshot: shows the current onboarding dashboard UI

Task:
1. identify the biggest onboarding problem
2. connect the evidence from the document and spreadsheet
3. note whether the screenshot suggests a usability contributor
4. produce a one-page recommendation memo with next steps and caveats

This version assigns each input a role, sequences the analysis logically (identify the problem, connect evidence, check the UI, synthesize), and defines the final artifact (a one-page memo with caveats). The model now knows what each file is for and how they relate to each other.

Example 3: a content production workflow

I want to produce a client-ready project summary using multiple inputs.

Inputs:
- Spreadsheet: project timeline with milestones, owners, and status (source of truth for progress)
- Document: meeting notes from the last three stakeholder calls (source of qualitative context)
- Screenshot: current project dashboard showing burn-down chart (visual evidence of trajectory)

Workflow:
1. First, extract key milestones from the spreadsheet and flag any that are overdue or at risk.
2. Then, identify the top 3 concerns or decisions from the meeting notes that relate to those milestones.
3. Note whether the burn-down chart aligns with the spreadsheet status or suggests a discrepancy.
4. Produce a two-page project summary with: executive overview, milestone status table, key risks, and recommended next steps.

Include a "confidence notes" section at the end listing any claims that depend on inference rather than explicit data.

This example shows a production-oriented multimodal workflow. Notice the "confidence notes" section at the end. This is the multimodal equivalent of asking for caveats: it forces the model to be transparent about which parts of the synthesis are well-supported and which are extrapolated.

Prompt block

Use this document, spreadsheet, and screenshot to help me.

Better prompt

I want to run a multimodal workflow.

Inputs:
- Document: defines the onboarding process and current policy
- Spreadsheet: shows drop-off by stage and customer segment
- Screenshot: shows the current onboarding dashboard UI

Task:
1. identify the biggest onboarding problem
2. connect the evidence from the document and spreadsheet
3. note whether the screenshot suggests a usability contributor
4. produce a one-page recommendation memo with next steps and caveats

Why this works

The better prompt gives each input a role and defines a single final artifact, which keeps the workflow coherent. This works because multimodal synthesis is fundamentally a composition problem. Each input contributes a different kind of evidence, and the final artifact needs to weave them together. Without explicit roles and sequencing, the model treats each input as an independent task and produces parallel summaries rather than an integrated analysis.

The request for caveats is especially important in multimodal work because the opportunities for error multiply with each input. A misreading of the spreadsheet, a misinterpretation of the screenshot, or a missed detail in the document can all compound in the final synthesis. Asking for caveats forces the model to flag where it is less certain, giving you a verification checklist rather than a polished answer that hides its gaps.

Common mistakes
  • Adding multiple input types without assigning them roles. Each file needs a stated job, or the model treats them all as generic background.
  • Trying to synthesize before understanding each source on its own terms. Jumping to conclusions without verifying the model's understanding of each input produces shallow synthesis that sounds integrated but is not.
  • Producing a polished final answer with no visible chain back to the inputs. If you cannot trace a claim in the output to a specific input, the claim may be fabricated or inferred without evidence.
  • Uploading too many inputs at once without sequencing. The model handles staged workflows better than a single massive prompt with five attachments and a complex task.
  • Skipping verification checkpoints between stages. If the model misreads the first input, every subsequent stage inherits that error. A quick "confirm your understanding" step between stages catches problems early.
Mini lab
  1. Choose a small real task that involves at least two input types (document plus spreadsheet, screenshot plus text, or any other combination).
  2. Define the role of each input in one line before starting.
  3. Run the workflow in stages: have ChatGPT process each input separately first, then ask for synthesis.
  4. Ask ChatGPT for one final artifact that preserves evidence and caveats.
  5. In one sentence, name what the staged approach revealed that a single-prompt approach would have missed.

Do not skip step five. The reflection forces you to articulate why staged workflows produce better results than single-prompt approaches. That understanding transfers to every multimodal task you run in the future.

Key takeaway

Multimodal workflows work best when each input has a job and the final deliverable is clear from the start.