Feedback from 3 Different Chatbots

I am currently working on a product sense grading prompt. It is more work than you might imagine to get it right. For the results to be genuinely helpful for me and others.

In my newsletter this week I shared the audio of my answering the prompt: Product to enjoy art. Here are how the three different chatbots responded.

One of them got things wrong, they missed things that were in the transcript. Not unlike a real panel, you will see a high, medium and low response. You see overlapping feedback but also where things diverge and people have different opinions on what the output indicated.

I am debating about continuing to use all three to get different POVs but it is likely too much work to keep up after testing is complete.

If you listened and want to compare your results to the prompt that is slowly being programmed to be my doppleganger.


ChatGPT

Executive Summary

Strong, structured product sense with clear prioritization. You framed the mission (“help people enjoy art”), narrowed ambitiously yet pragmatically to novices in the online context, prioritized pain points via frequency × severity, and proposed three distinct solutions, then prioritized with Addresses Need × Engineering Effort and gave a workable verbal flow for the top pick. Biggest gaps: make the “why now” tie back to company strategy, state explicit risks & tradeoffs, and offer goal-aligned metrics. Overall, this reads like a solid Hire with a few polish items away from “Strong Hire.”

Scores

Overall Recommendation: Hire
Gate-Acknowledged Score: 4.3 / 5
Adjusted Potential Score (if gates passed): N/A

What You Did Well

  • Clear mission + scope: “Help people find enjoyment in art,” then focus on novices and online context.

  • Prioritization discipline: Used TAM/Opportunity to choose online; frequency × severity to select “articulation” as the top pain; Need × Effort to pick the phone app.

  • Solution breadth → depth: Three distinct concepts (flip book, translator, phone app) with a concise verbal flow for the phone app.

  • Audience-aware check-ins: Frequent “pause/any questions before I prioritize?” reduced cognitive load.

Where To Improve Most

  • Make “why now” strategic: You cited pandemic/AI shifts; explicitly connect to company right-to-win (distribution, data, partnerships) and user adoption levers.

  • State explicit risks & tradeoffs: (e.g., privacy from photo ingestion, artist IP/licensing, on-device vs cloud inference costs) and how you’d mitigate.

  • Define success with metrics: Tie to mission (enjoyment → discovery/usage outcomes), not generic engagement.

  • Sharpen two-sided view: You noted artists/museums; briefly justify sequencing (start with consumers; later unlock creator marketplace).

Clarifying Questions & Assumptions

  • Scope of “art”: Confirmed 2D/3D visual (paintings, sculpture).

  • Company context: Assumed startup; later sell to museums or stand-alone

  • Constraint posture: Assumed no hard budget to explore solution space.

  • User focus: Chose novice over expert to maximize TAM and utility.

Strategic Setup

  • Mission & goal: Help people enjoy art; refined to “help people find the art they want to enjoy by improving articulation.”

  • Why now (good but thin): Pandemic-driven digital consumption + AI/ownership debates; strengthen by tying to distribution advantages and creator economics.

User Segmentation

  • Primary split: Novice vs. expert; chose novice intentionally.

  • Context segmentation: Online vs. museum vs. home; prioritized online using TAM and frequency of exposure/opportunity.

  • Pick rationale: Online offers highest reach and under-served “enjoyment” mechanics beyond mere exposure.

Pain Points

  • List (focused): Articulation/definition, searching, overexposure/habituation for screensavers/backgrounds.

  • Prioritization: Selected articulation using frequency × severity (search depends on articulation; overexposure affects fewer users).

Solutions

  • Set: (1) Flip book style explorer → vocabulary; (2) Translator chat tool from vague descriptors to style; (3) Phone-camera “I like” collector → weekly recs.

  • Prioritization: Addresses Need × Engineering Effort → pick phone app (habit-piggyback, lighter MVP).

  • Verbal flow present: Download app → capture liked visuals/screens → model infers style → periodic suggestions (screensavers/backgrounds) → optional artist marketplace later.

  • Add next time: 1–2 explicit risks (privacy/IP, cold start, model bias) + mitigations; outline MVP vs V2 (on-device embeddings now, creator marketplace later).

Metrics

If unasked, metrics are optional; here are aligned examples you could state next time:

  • % users who save ≥1 recommended artwork/week (activation).

  • Time-to-first “I like” recommendation ≤7 days (onboarding success).

  • Weekly “kept as background ≥7 days” rate (enjoyment persistence).

  • Creator payout rate / $ per active creator / week (when marketplace launches).

Recommended Practice

  • Keep the check-ins; add a 10-second gate after each major section (“pause/adjust?”).

  • Always name 1–2 concrete risks per top solution with mitigations.

  • Anchor “why now” to right-to-win (data, channels, ecosystem).

Suggested Reading

Mock Count Recommendation

10 mocks (aim for all ≥4). Focus reps on: explicit risk/tradeoff articulation, crisp “why now” with strategic tie-in, and rapid metric framing.


Claude

Product Sense Interview Evaluation

Executive Summary

The candidate demonstrated solid product thinking fundamentals with clear structure and logical flow. They effectively clarified the problem space, established a user-focused goal, and provided creative solutions with good prioritization frameworks. However, critical gaps in explicit prioritization criteria and missing solution details prevented higher scores. The candidate showed strong communication skills and adapted well to interviewer questions, but needs to be more precise with frameworks and include technical implementation details.

Scores

Overall Recommendation: Leaning Hire
Gate-Acknowledged Score: 3/5
Adjusted Potential Score (if gates passed): 4/5

What You Did Well

  • Strong clarifying questions - Asked fundamental questions about art scope, company context, and budget constraints before diving in

  • Clear goal articulation - Established "help people find the art they want to enjoy" as a user-focused mission

  • Creative solution generation - Proposed three distinct solutions (flip book, translator tool, camera app) with different approaches

  • Good communication structure - Used clear transitions and logical sequencing throughout the interview

  • Thoughtful user journey mapping - Walked through specific pain points like screensaver selection and overexposure

Where To Improve Most

  • Missing explicit prioritization criteria - Used vague terms like "opportunity" and "need" without defining concrete frameworks (this capped your score)

  • Insufficient solution depth - Lacked verbal wireframe/flow description for chosen solution with specific screens and user actions

  • No risk/tradeoff discussion - Failed to identify implementation risks or potential downsides of proposed solutions

  • Imprecise segmentation framework - "TAM and opportunity" criteria were not clearly defined or measured

Clarifying Questions & Assumptions

  • Effective fundamentals coverage - Started with core problem definition (art scope) rather than tactical details

  • Smart assumption validation - Confirmed startup context and budget flexibility appropriately

  • Could have probed deeper - Might have asked about target demographics or existing competitor landscape

Strategic Setup

  • Clear mission statement - "Help people find enjoyment in art" was well-articulated and user-focused

  • Solid "why now" reasoning - Connected pandemic digital shift and AI/creativity debates to current market timing

  • Good market insight - Recognized intimidation factor and accessibility barriers in art consumption

User Segmentation

  • Reasonable segment identification - Expert vs novice distinction made sense, then refined to online/museum/home contexts

  • Weak prioritization framework - "TAM and opportunity" criteria were mentioned but not clearly defined or measured

  • Logical segment choice - Online focus was sensible given accessibility and frequency arguments

Pain Points

  • Good quantity and specificity - Identified relevant pain points like articulation difficulty, search challenges, overexposure

  • Attempted prioritization - Used frequency × severity framework appropriately

  • Clear selection rationale - Chose articulation problem with sound reasoning about foundational nature

Solutions

  • Met requirement - Provided three distinct solutions with different approaches

  • Missing verbal wireframe - Lacked detailed flow description with specific screens, states, and user interactions

  • No explicit risks identified - Failed to discuss implementation challenges or potential failure modes

  • Good prioritization attempt - Used "addresses need × engineering effort" framework appropriately

Metrics

Not attempted - candidate focused appropriately on core problem-solving given time constraints.

Examples if needed:

  • Weekly active users creating style profiles

  • Percentage of users who save recommended art within 7 days

  • Average time to find desired art style (reduction metric)

  • User satisfaction score on art recommendations

Recommended Practice

  • Practice explicit framework articulation - Define all prioritization criteria with specific measurements

  • Develop solution wireframing skills - Practice describing user flows with screen-by-screen detail

  • Include risk assessment - Always identify 1-2 implementation risks and mitigation strategies

Suggested Reading

Mock Count Recommendation

10 additional mocks recommended to practice framework precision and solution depth.


Google Gemini

Executive Summary

The candidate delivered an exceptionally strong and well-structured performance, demonstrating a masterful command of product sense frameworks. They excelled at turning an ambiguous prompt into a clear plan, methodically moving from clarifying questions to a user-focused goal, needs-based segmentation, and prioritized pain points. Their use of explicit prioritization frameworks (TAM/Opportunity, Frequency/Severity, Need/Effort) in every major section was best-in-class. The primary area for improvement is pacing; the setup portion of the interview was lengthy, leaving less time for solution depth. While the proposed solutions were creative and well-reasoned, the final discussion lacked an explicit analysis of risks and tradeoffs for the chosen concept. Overall, this was a "Strong Hire" performance.

Scores

Overall Recommendation: Strong Hire Gate-Acknowledged Score: 5 / 5 Adjusted Potential Score: N/A


What You Did Well

  • Excellent Framework-Driven Approach: You began by outlining a clear structure and followed it precisely, making your thought process logical and easy to follow.

  • Consistent, Explicit Prioritization: You were one of the few candidates who correctly applied distinct, well-defined prioritization frameworks at every critical stage: User Segmentation (TAM × Opportunity), Pain Points (Frequency × Severity), and Solutions (Solves Need × Engineering Effort). This is a significant strength.

  • Strong User-Centricity: You established a clear, user-focused goal ("help people find enjoyment in art") and maintained that focus throughout, from identifying the "novice" user to defining their specific pain points.

  • Creative and Differentiated Solutions: You generated three distinct solution concepts (Flip book, Translator tool, Camera feature) that addressed the core user problem from different angles, demonstrating both creativity and strategic thinking.

  • Clear Communication and Check-Ins: Your communication was concise and structured. You effectively used pauses to gather your thoughts and checked in with the interviewer at key transition points.

Where To Improve Most

  • Pacing and Time Management: Your setup (from clarifying questions to pain points) was very thorough but took too long. You must move to the solutions phase more quickly to allow for greater depth, probing, and discussion of risks and GTM. Move to solutions by minute 20-22; use 10-second check-ins after each section to maintain momentum.

  • Explicit Risk & Tradeoff Analysis: For your chosen solution, you did not explicitly discuss the potential risks (e.g., cold start problem, privacy concerns with photo access, accuracy of the AI) or tradeoffs (e.g., prioritizing passive collection over active learning). This is a required component for a top-tier solutions section.

  • Conciseness in Setup: While your reasoning was clear, you can tighten your explanations during the strategy and segmentation phases to save time for the solutioning deep dive, which is the most critical part of the interview.

Clarifying Questions & Assumptions

  • Strong Start: You began perfectly by clarifying the ambiguous term "art" and establishing the scope (startup, no budget constraints).

  • Effective Framing: Your questions successfully reduced the problem space and demonstrated a focus on understanding the fundamentals before proceeding.

Strategic Setup

  • Clear Mission and Goal: You established a strong, user-centric goal upfront: "to help people find enjoyment in art," which you later refined effectively.

  • Compelling "Why Now": Your justification was excellent, linking the problem to current, relevant trends like the pandemic's impact on digital consumption and the rise of AI in creative fields.

  • Acknowledged Market Dynamics: You correctly identified the key players (creators, consumers, institutions) before strategically choosing to focus on the consumer side.

User Segmentation

  • Needs-Based, Not Demographic: You correctly avoided demographic traps and segmented your target "novice" user based on their context and job-to-be-done: enjoying art online, in a museum, or at home.

  • Excellent Prioritization: You used a clear framework (Total Addressable Market × Opportunity/Frequency) to justify your choice to focus on the "online" segment, which was well-reasoned and convincing.

Pain Points

  • Specific and User-Centric: You identified a strong list of 5+ pain points specific to the online novice, such as "defining what I like," "copyrights," and "overexposure."

  • Rigorous Prioritization: Your use of a Frequency × Severity matrix to select the top pain point ("defining and articulating what you like") was exemplary and demonstrated strong analytical thinking.

Solutions

  • Good Breadth and Creativity: You proposed three distinct and viable concepts that directly addressed the prioritized pain point.

  • Clear Verbal Wireframe: You provided a solid verbal walkthrough of your chosen "Camera phone" feature, describing the key user actions and outcomes (take pictures -> get style profile -> receive recommendations).

  • Missing Risk Analysis: As noted, the key missing piece was an explicit discussion of the risks and tradeoffs associated with your prioritized solution. For example, a risk for the camera app is the "cold start" problem (needing many photos to be useful) and a tradeoff is building a standalone app versus integrating into an existing photo platform.

Metrics

  • Not Discussed: This section was not covered, which is acceptable as you were not prompted for it.

  • Example Metrics: For your chosen solution, success could be measured by:

    • Activation: % of new users who add ≥10 photos to their style profile in the first week.

    • Engagement: # of recommendations saved or applied (e.g., as a screensaver) per user per month.

    • Goal Fulfillment: User satisfaction score (NPS or survey) on the question, "How well does this app help you discover art you enjoy?"

Recommended Practice

  • Time-Box Your Setup: Practice mock interviews with a timer. Force yourself to complete the clarifying, strategy, segmentation, and pain points sections within 20 minutes to ensure at least 15 minutes remain for solutions and probing.

  • Drill Risks & Tradeoffs: For every solution you propose in practice, create a habit of immediately listing 1-2 key risks and 1-2 key tradeoffs. This will make it a natural part of your flow.

Suggested Reading


Mock Count Recommendation

Based on this performance, you need 4 more mock interviews. (All 5 required mocks must score a 4 or 5).

Previous
Previous

Reminder: Diagram Your Prompt

Next
Next

Grading Prompts