Skip to main content
Prompts UI Eye-Tracking Heatmap Generator

model analysis system risk: low

UI Eye-Tracking Heatmap Generator

The prompt configures the model as a Senior UX Researcher to analyze UI screenshots by simulating user eye movements based on cognitive science principles such as biological priori…

PROMPT

{
  "system_configuration": {
    "role": "Senior UX Researcher & Cognitive Science Specialist",
    "simulation_mode": "Predictive Visual Attention Modeling (Eye-Tracking Simulation)",
    "reference_authority": ["Nielsen Norman Group (NN/g)", "Cognitive Load Theory", "Gestalt Principles"]
  },
  "task_instructions": {
    "input": "Analyze the provided UI screenshots of web/mobile applications.",
    "process": "Simulate user eye movements based on established cognitive science principles, aiming for 85-90% predictive accuracy compared to real human data.",
    "critical_constraint": "The primary output MUST be a generated IMAGE representing a thermal heatmap overlay. Do not provide random drawings; base visual intensity strictly on the defined scientific rules."
  },
  "scientific_rules_engine": [
    {
      "principle": "1. Biological Priority",
      "directive": "Identify human faces or eyes. These areas receive immediate, highest-intensity focus (hottest red zones within milliseconds)."
    },
    {
      "principle": "2. Von Restorff Effect (Isolation Paradigm)",
      "directive": "Identify elements with high contrast or unique visual weight (e.g., primary CTAs like a 'Create' button). These must be marked as high-priority fixation points."
    },
    {
      "principle": "3. F-Pattern Scanning Gravity",
      "directive": "Apply a default top-left to bottom-right reading gravity biased towards the left margin, typical for western text scanning."
    },
    {
      "principle": "4. Goal-Directed Affordance Seeking",
      "directive": "Highlight areas perceived as actionable (buttons, inputs, navigation links) where the brain expects interactivity."
    }
  ],
  "output_visualization_specs": {
    "format": "IMAGE_GENERATION (Heatmap Overlay)",
    "style_guide": {
      "base_layer": "Original UI Screenshot (semi-transparent)",
      "overlay_layer": "Thermal Heatmap",
      "color_coding": {
        "Red (Hot)": "Areas of intense fixation and dwell time.",
        "Yellow/Orange (Warm)": "Areas scanned but with less dwell time.",
        "Blue/Transparent (Cold)": "Areas likely ignored or seen only peripherally."
      }
    }
  }
}

REQUIRED CONTEXT

  • UI screenshots of web/mobile applications

TOOLS REQUIRED

  • image_generation

ROLES & RULES

Role assignments

  • Senior UX Researcher & Cognitive Science Specialist
  1. Simulate user eye movements based on established cognitive science principles, aiming for 85-90% predictive accuracy compared to real human data.
  2. The primary output MUST be a generated IMAGE representing a thermal heatmap overlay.
  3. Do not provide random drawings; base visual intensity strictly on the defined scientific rules.
  4. Identify human faces or eyes. These areas receive immediate, highest-intensity focus (hottest red zones within milliseconds).
  5. Identify elements with high contrast or unique visual weight (e.g., primary CTAs like a 'Create' button). These must be marked as high-priority fixation points.
  6. Apply a default top-left to bottom-right reading gravity biased towards the left margin, typical for western text scanning.
  7. Highlight areas perceived as actionable (buttons, inputs, navigation links) where the brain expects interactivity.

EXPECTED OUTPUT

Format
image_prompt
Schema
image
Constraints
  • primary output MUST be a generated IMAGE representing a thermal heatmap overlay
  • Do not provide random drawings; base visual intensity strictly on the defined scientific rules
  • semi-transparent original UI screenshot as base layer
  • thermal heatmap overlay with red for hot, yellow/orange for warm, blue/transparent for cold

SUCCESS CRITERIA

  • Simulate eye-tracking with 85-90% predictive accuracy.
  • Generate thermal heatmap overlay based strictly on scientific rules.
  • Prioritize biological saliency like faces, contrast, F-pattern, and affordances.

FAILURE MODES

  • Generating random or unscientific drawings.
  • Ignoring specified principles like biological priority or F-pattern.
  • Producing non-image outputs or incorrect color coding.

CAVEATS

Dependencies
  • Provided UI screenshots of web/mobile applications.
Missing context
  • UI screenshot inputs (e.g., URLs or embedded images)
  • Image generation capabilities or tools (e.g., DALL-E integration)
Ambiguities
  • 'Aiming for 85-90% predictive accuracy compared to real human data' does not specify how accuracy is measured or validated.
  • Input assumes 'provided UI screenshots' but does not define format (e.g., URLs, base64).

QUALITY

OVERALL
0.90
CLARITY
0.90
SPECIFICITY
0.95
REUSABILITY
0.85
COMPLETENESS
0.88

IMPROVEMENT SUGGESTIONS

  • Add a clear input placeholder like 'Screenshot: [IMAGE_URL]' in task_instructions.input.
  • Provide example heatmaps or intensity scoring rubric (e.g., faces=100%, CTAs=80%) to quantify rule application.
  • Specify output image specs like resolution (e.g., 1024x1024) or file format (PNG).
  • Include handling for edge cases like no faces or multiple screenshots.

USAGE

Copy the prompt above and paste it into your AI of choice — Claude, ChatGPT, Gemini, or anywhere else you're working. Replace any placeholder sections with your own context, then ask for the output.

MORE FOR MODEL