Built for PMs. By PMs.

Built for PMs. By PMs.

For Founders & PMs

Templates to help you get a jumpstart.

Built for PMs. By PMs.

For Founders & PMs

Templates to help you get a jumpstart.

All Posts

Templates

Research

widget
user pic

Templates

Aug 1, 2025

Hardware Product One-Pager (Spec → Shelf)

1) Hardware Product One-Pager (Spec → Shelf)

H1 (website): Hardware PRD that survives the real world
Subhead: From sensing and power math to drop tests—ship without “surprise failures.”

What you fill in (teaser):

Problem (1–2 lines), Success numbers, Non-goals

Sensing & actuation table, Power budget, Durability targets

Packaging, Serviceability, Launch gates


Full in-app template

Problem
One sentence. Who gets hurt today? Cost of not fixing?

Success Criteria (hard numbers)

First-pass yield at EVT/DVT: __% / __%

RMA rate: ≤ __% at 90 days

Setup time: ≤ __ minutes (p95)

Workflow completion rate (prototype→EVT→DVT): ≥ __%


Non-Goals
(Example: No multi-SKU colorways in v1. No Wi-Fi if BLE meets latency.)

System Requirements

Sensing: variable, unit, accuracy (±), sample rate ( Hz), calibration plan

Actuation: force/torque/speed; fail-safe defaults

Compute & Memory: min MCU/CPU class, RAM/Flash floors

Comms: protocol, throughput, latency budget

Power: chemistry, duty cycle, target life __h; charge time __h; thermal throttling behavior

Durability: UV/chemicals; IP rating target; drop spec (height × faces × repetitions)

Packaging: “in the box,” unboxing steps, retail constraints

Serviceability: repair vs replace, return flow, defect tolerances, SLAs


Key User Flows

Out-of-box → Activated
First success (task done)

Update (firmware)

Troubleshoot → Return


Acceptance Checks (Given-When-Then)

Given 1 m drop on each face ×10, when rebooted, then self-test passes and boots < 5s.

Given enclosure at IP54 test, when exposed to dust/splash, then no ingress.

Given duty cycle X for __ hours, then battery SOC ≥ __%.

Given BLE RSSI −85 dBm, when command sent, then control latency ≤ 200 ms.


Instrumentation & Events

  • device_boot, self_test_pass/fail, drop_event_detected, battery_soc, fw_update_success/fail.


Privacy & Data

  • Default: telemetry stored locally; cloud only with explicit opt-in. Retention: __ days if opted-in.

Tokens + Cost (if any AI on-device)

  • Tokens-per-workflow: ≤ __k

  • Cost-per-workflow: ≤ $.

  • Routing: small on-device model for anomaly; escalate to cloud model only if confidence < __.

Launch Gates
EVT exit, DVT exit, PVT readiness; pilot sample size; GA criteria.


All Posts

background pattern

Research

Aug 22, 2025

Most AI projects fail. Here’s why—and how to ship agents that actually work.

A new report from MIT Media Lab’s Project NANDA—The GenAI Divide: State of AI in Business 2025—looked at hundreds of real deployments and found a harsh truth: about 95% of generative-AI pilots don’t produce business impact. Only ~5% translate into revenue or real savings. The problem isn’t “AI is bad.” The problem is how we build and run it.

The short version: we’re trying to run AI agents like old, deterministic software. That doesn’t work. AI is probabilistic; outputs vary; providers change models; quality moves if you don’t measure it. You need a different playbook—one built around evaluations, simulation, guardrails, and cost/quality routing. Even OpenAI’s leadership keeps saying the quiet part out loud: evals are the work.

Where AI projects go wrong—and how to fix them

Symptom you see

Why it’s happening

What to do differently

Endless pilots that never reach production.

Teams run agents with old software rituals (feature list, demo, launch) instead of an agent playbook (goals, evaluations, simulator, rollout rules).

Write an agent-centric product requirements document (spell out task goals, tools allowed, escalation rules, acceptance thresholds). Pair it with a living evaluation set and a simulator plan from Day 1. The MIT study shows most pilots stall without this rigor.

Demos look great; real users get flaky results.

No graded evaluations—only spot checks.

Build a living evaluation suite that mirrors real workflows and scores usefulness, safety, completeness, latency, and cost—not just “pass/fail.”

“Same prompt, different answer.”

AI is probabilistic; behavior drifts with context and provider updates.

Pin model versions where possible, capture seeds/parameters, and re-run evaluations whenever anything changes. Use a router to balance quality, cost, and latency per task.

Agents pass unit tests but fail in the wild.

You tested components, not long-horizon tasks with tools and UI.

Use a sandbox/simulator to rehearse full tasks before go-live (web environments, multi-site flows, deterministic replicas).

Surprising errors and “confident nonsense.”

Hallucinations and reasoning gaps aren’t monitored.

Add guardrails (grounding, retrieval, policy checks), confidence/deferral to humans, and quality dashboards. Nature and other studies show hallucinations are inherent—so you must detect and contain them.

Costs spike as usage grows.

Usage is token-metered; small inefficiencies compound.

Track cost per task, cap tokens by step, cache and batch where safe, and route simpler tasks to cheaper models. (Yes—pricing is usage-based.)

“It worked last week.”

Hidden ML technical debt and drift: data shifts, prompts change, providers update.

Treat data and prompts as first-class code; add change audits; invest in ML observability. The classic “Hidden Technical Debt in ML Systems” explains why old habits break.

Privacy/security objections stall deployment.

Data handling is unclear; cloud-only deployments.

Offer on-prem/private-cloud options, document data isolation and retention, and build redaction/access controls into the agent itself. (Standard enterprise best-practice patterns.)

The playbook we use at Tastefully.ai

We build agents like systems, not stunts. Four pillars make the difference.

1) In-house Router (quality-, cost-, and latency-aware)

  • Task-aware selection. The router chooses the smallest model that can pass the task’s acceptance criteria; escalates only when needed.

  • Version control. We can pin model/provider versions and re-run evals on every change.

  • Cost guardrails. Token caps per step, caching, and batch policies keep spend predictable.

  • Quality arbitration. For high-stakes steps, the router can cross-check with a second model or a verifier before acting.
    This turns “black-box AI” into managed, measurable behavior instead of hope. (Usage-based pricing is real; routing is how you keep it sane.)

2) In-house Simulator (our “agent gym”)

  • We rehearse real tasks end-to-end: multi-page flows, tool calls, dead ends, and recovery paths.

  • The simulator runs deterministic replicas and scripted obstacles so we can measure repeatability and catch regressions.

  • We keep an evaluation logbook (by scenario and version) so production changes never outpace quality checks.
    Benchmarks like WebArena and REAL show why deterministic web/task replicas matter; we’ve built our own for your workflows.

3) Spec-and-Tickets Generator that isn’t “just a template”

  • It listens to real work (docs, meetings, tickets) and proposes diffs to your spec with clear rationales.

  • Tickets come with acceptance criteria, dependency notes, and confidence tags; low-confidence items auto-escalate for review.

  • Every suggestion is trace-linked to the evidence that justified it (meeting excerpt, doc section, prior decision).
    Result: less “AI slurry,” more auditable decisions your team can trust.

4) Dynamic UI/UX Generator

  • Agents don’t leave you with raw JSON. They generate small, adaptive interfaces—fields, tables, and checklists matched to the task.

  • UI elements are typed and validated (think: schema + state machine), so clicks and updates are idempotent and traceable.

  • This is how we turn probabilistic text into predictable actions your team can review and approve.

“Deterministic” behavior—without pretending AI is deterministic

Truth: you can’t make a large language model fully deterministic in the wild. Nature-level research shows that hallucination risk never vanishes; you manage it.

What you can do is box in the variability so outcomes are repeatable enough for production:

  • Constrain the task (tools, scopes, allowed actions).

  • Define acceptance thresholds per step; fail closed.

  • Use retrieval/grounding for facts; verify or defer on low confidence.

  • Route smartly, pin versions, and re-eval on change.

  • Wrap actions in typed UI/state machines so the end effect is consistent.

That’s how we get “deterministic-feeling” agents—reliable, auditable, and ready for real work.

Why this matters now

Two things are true at once:

  1. Most pilots fail—not for lack of ambition, but for lack of the right operational model.

  2. Evals + simulation + routing are the difference between a cool demo and a durable system. This is fast becoming industry consensus

If you’re done with pilot theater and want production-grade agents—with privacy, quality, and cost under control—we should talk.

black flat screen tv turned on displaying game
user pic

Research

Jan 14, 2025

What Is “Taste”? The Career Edge for Product Managers

Be the editor. Cut nine, ship one.

Rick Rubin didn’t become a legendary producer by mastering every instrument. He mastered taste — knowing what to keep, what to cut, and when to say “that’s done.” He built that taste the unglamorous way: hanging around underground clubs in New York, listening hard, learning what people loved — and what they didn’t.

Film director Alan Parker said a movie is made three times: when you write it, when you shoot it, and when you edit it. No one watches forty hours of raw footage. Editors make the story.
That’s your job as a product manager: you are the editor.

Most product managers aren’t practicing taste. They’re buried under meeting notes, requests, documents, and tickets. The backlog owns them.

This piece shows how to reclaim taste — not as vibes, but as a simple loop you can run every day — and how Tastefully makes that loop concrete without adding another tool to babysit.

What taste is (and isn’t)

  • Taste isn’t costume. Not the turtleneck, not the quotes. It’s the ability to make fewer, better decisions under noise.

  • Taste is trained. Rubin calls himself a reducer: strip to what matters. That’s practice, not luck.

  • Taste is values alignment. Match what your product stands for to what your users actually value.

  • Taste is subtraction. Industry reports suggest most features are rarely used. That is wasted effort. The courage to stop at “enough” is taste.

  • Taste protects attention. Context switching taxes your brain. Guarding focus is not a luxury; it’s throughput.

A simple loop for taste

Think of taste as a repeatable loop — not magic.

  1. Listen
    Pull raw inputs from interviews, support threads, analytics, notes.

  2. Filter
    Score ideas on two axes only: impact on one core behavior and time to first proof.

  3. Decide
    Commit to the top three moves for the next day or week. If everything is a priority, nothing is.

  4. Ship
    Define “done” as a change in behavior with clear acceptance criteria. Ship to the smallest audience that can prove or falsify.

  5. Learn
    Compare actual behavior to the one measure you targeted. Keep, kill, or iterate.

How Tastefully makes this real (today)

No mystique. No future roadmap. Two things you can use right now.

1) Morning “Top Three” — your producer’s note

Every morning, Tastefully ingests yesterday’s calls, chat, calendar, documents, and product data. It clusters the chaos into themes, scores them by one behavior to move × time to proof, and returns three actions for today — with links, owners, and blockers.

Constraint, not clutter. Three moves, not thirty. Space to practice taste.

2) Auto-draft and sync the product requirement document, specs, and tickets

While you work, Tastefully drafts or updates the product requirement document, acceptance criteria, and tickets from real conversations and decisions — so your source of truth stays current without copy-paste.

  • No more stale documents that drift from reality.

  • No more ticket farms with thin context.

  • No more status theater in standup; the work explains itself.

Why engineers will love this

  • Faster unblocks. Tickets arrive ready to start: one behavior to move, crisp acceptance criteria, links to the exact source decisions. Less back-and-forth.

  • Less noise. Duplicates get merged. “Nice to have” fluff gets cut. Smaller backlog, clearer queue.

  • Fewer last-minute changes. Specs stay in sync with real discussions. Less rework.

  • Fewer status pings. Context lives in the ticket. Engineers can build instead of explain.

Use this line: Clear outcomes. Fewer tickets. Faster unblocks.

If you don’t have a product manager yet (engineering managers, this is for you)

  • Get a daily Top Three that actually moves this week’s outcome.

  • Turn standups and call notes into a one-page spec and real tickets.

  • See a pre-prioritized queue: today’s must-do, not a wish list.

  • Ship a small bet to real users within days, not weeks.

How it works (mechanism, not hype)

  1. Ingest yesterday’s calls, chat, calendar, docs, product data.

  2. Cluster related requests and changes into themes.

  3. Score each theme by one behavior it moves × time to first proof.

  4. Return your Top Three with links, owners, blockers.

  5. Auto-draft the product requirement document, acceptance criteria, and tickets from the real conversations and decisions.

  6. You edit once; the system keeps everything in sync.

See it before you try it (sample output)

Sample Top Three (from demo data)

  1. Reduce drop-off on onboarding step 2

    • Action: add inline help; run 48-hour A/B with 10% traffic

    • Evidence: 38% of support threads mention confusion at this step

    • Links: call notes · spec draft · ticket

  2. Close the loop on churn survey

    • Action: ship a one-question in-app prompt; trigger downgrade rescue email

    • Evidence: spike in “price vs. value” mentions in past 7 days

  3. Fix duplicate user profiles

    • Action: merge job; acceptance criteria added to spec

    • Evidence: 14% of failed checkouts tied to duplicate accounts

Spec excerpt (auto-drafted)
Goal: Increase completed sign-ups per visitor
Behavior to change: finish step 2 without bouncing
Acceptance criteria: median time on step 2 down 20%; completion rate +8 points in 7 days
Risks: added latency; misleading copy
Out of scope: redesign entire flow

How to know it’s working (simple, measurable)

  • Days to first proof: idea to first user signal in 72 hours or less.

  • Top Three hit rate: of the last 30 actions, how many moved the one behavior you chose. Aim for 50–70%.

  • Subtractions: at least one meaningful removal every two weeks (feature, meeting, doc, or request).

  • Outcome delta: pick one release metric; if it didn’t move, the work was overhead.

Short FAQ

Do I have to switch tools?
No. We work with your current docs, calendar, chat, and tracker. You edit; we do the typing.

How do I keep junk tickets out?
Set the one behavior you want to move. We only draft tickets tied to it. You approve before anything goes live.

Will this slow my team down?
No. It removes steps. Capture → auto-draft → quick edit → assign.

Will engineers actually use the output?
Yes. Because it is concrete: one behavior, clear acceptance criteria, links to decisions — and fewer tickets.

Try it

Get your Top Three tomorrow morning.
Start free. Limited alpha slots.

a close up of a sign with an apple logo
user pic

Research

Apr 2, 2025

Why great product managers run toward “self-cannibalization”

Creative destruction isn’t a threat to your career. It’s the path to your edge.

Steve Jobs didn’t defend the iPod. He replaced it.
He knew the iPhone would eat iPod sales. He shipped it anyway because that’s the natural order in technology: either you cannibalize yourself, or someone else will.

That same pressure is now sitting on the day-to-day work of product managers.
The repeatable parts—condensing meeting notes, drafting product requirement documents, writing acceptance criteria, creating tickets, syncing status—are ready for automation.

So the real question isn’t “Will automation take my job?”
It’s: Will you design the system that removes your busywork so you can do the part only you can do—taste, judgment, and bold bets?

First principles: what a product manager is paid to do

Strip the role down to load-bearing pieces:

Coordinate
Move information between people and tools. Scheduling, summarizing, ticket typing, status chasing.

Synthesize
Turn raw signals into a single narrative: what changed, why it matters, what to do next.

Decide
Say no to most things, yes to a few, and define the finish line in plain terms.

Practice taste
Hold the line on values, reduce the product to its essence, and protect attention so the important work actually finishes.

Now be honest: which of these should a machine do, and which must a human own?

The automation frontier (today)

Work type

Can a machine do it now?

Why

What you should do

Coordinate (summaries, document updates, ticket typing, status sync)

Yes, reliably

Repeatable rules; context already lives in tools

Hand it off. Keep final review.

Synthesize (draft product requirement sections, extract decisions, link evidence)

Yes, with your edit

Inputs are structured enough; drafts benefit from human judgment

Automate the first draft. You do the edit.

Decide (what gets built now vs. later)

Not yet

Values, leverage, risk tradeoffs

Own it. Use data; keep the call.

Practice taste (subtraction under noise)

No

Judgment shaped by experience

Own it. That’s your moat.

If you cling to coordination, you’ll be replaced.
If you elevate into synthesis, decision, and taste, you become hard to replace.

What “creative destruction” looks like in your week

Before

Mornings rewriting scattered notes into a product requirement document.

Afternoons copying the same context into a tracker.

Evenings lost to “one more pass” on acceptance criteria.

After you self-cannibalize

You start with a top three list distilled from yesterday’s calls, chat, calendar, documents, and product data.

A system drafts the new product requirement section and tickets; you only edit.

You ship smaller bets faster because the finish line is written in plain language: the single user behavior you intend to change and how you’ll know.

The work gets quieter: fewer handoffs, fewer re-types, fewer status theatrics.


“I bounce between analytics, heatmaps, tickets, and waiting days for implementation. Give me one place where insights connect to the change and deploy the best variant for that user.” — A growth-PM's insight
Exactly. Your stack should stop being a museum of dashboards and start being a pipeline from signal → decision → shipped change.

Money talk: why this raises your compensation and builds a moat

Companies don’t pay for keystrokes. They pay for judgment that moves a metric.

A simple, believable model:

Cycle time: Cutting idea-to-proof from 10 days to 3 lets you run ~3 times more real experiments per month.

Hit rate: If your “top three each day” improves the share of actions that move the target behavior from 20% to 35%, your wins per quarter nearly double.

Talent signal: More shipped learning with less drama reads as seniority. Senior people get the raise, the equity, and the hard problems.

Compounding effect (example):

You ship 36 proofs in a quarter instead of 12.

At a 35% hit rate, that’s ~13 wins vs. ~2–3 before.

More wins → larger scope → larger comp. The moat isn’t secrecy; it’s throughput of good judgment.

Or as one commenter put it: don’t get paid for time or effort—get rewarded for judgment. Automation exposes that judgment. That’s good news for people who have it.

The editor’s loop (how you actually work)

Think like a film editor. The camera captured 40 hours; your job is to cut.

  1. Listen
    Pull raw inputs: calls, support threads, analytics, decision notes.

  2. Filter
    Judge ideas on two axes only: impact on one behavior × time to first proof.

  3. Decide
    Pick three moves for the next 24–72 hours. If everything is a priority, nothing is.

  4. Ship
    Define “done” as a behavior change with clear acceptance criteria (Given / When / Then written in plain English). Ship to the smallest audience that can prove or falsify.

  5. Learn
    Compare outcomes to the single measure you targeted. Keep, kill, or iterate. Then repeat.

Automation doesn’t do your craft. It removes the sand in the gears.

Addressing the real objections

“Tools won’t fix a weak growth culture.”
True. The tool should force better thinking by making the path from signal to shipped change embarrassingly short. The habit change is choosing, daily, to ship tiny proofs instead of polishing decks.

“Who owns growth—product or marketing?”
In modern teams, growth is a discipline across the funnel, not a silo. Automation helps by unifying the loop so every builder can practice growth with the same rigor.

“Will automation erase the craft?”
No. It erases retyping. What remains—taste under noise, subtraction, the call to stop—is the craft.

“Specs as the new source code?”
Yes. When your product requirement document is living and executable—linked to evidence, tickets, and acceptance criteria—your spec becomes the control system for change.

Write your replacement plan (this week)

List every task a system should do in your week. If it doesn’t require judgment, it’s a candidate.

Codify inputs and outputs. Example: “After a customer call, produce a product requirement section, acceptance criteria in Given-When-Then, and two tickets—with links to evidence.”

Keep a kill list. Every Friday, remove one feature, meeting, document, or “nice-to-have” request. Taste is subtraction.

Measure what matters:

Days to first proof (idea to real user signal ≤ 72 hours)

Top-three hit rate (share of moves that moved the behavior)

Meaningful subtractions (one every two weeks)

Busy isn’t leverage. Finished, useful work is leverage.

Start here (no ceremony, no process theater)

Wake up to a top-three list in your inbox.
Pulled from yesterday’s calls, chat, calendar, documents, and product data; scored by impact on one behavior × time to first proof.

Edit drafts, don’t start from blank pages.
Real conversations become a living product requirement document, acceptance criteria, and ready-to-start tickets. You edit once; the system keeps everything in sync.

Run the editor’s loop for 30 days. Track only three measures.
If the three don’t move, automate more and cut harder. If they do, good—you just got your real job back.

What Tastefully automates today (so you can do the human part)

Coordination: meeting digests, decision logs, document updates, ticket creation, status sync across your tools.

Synthesis first draft: product requirement sections with linked evidence; acceptance criteria in plain English.

Integrated signals: product analytics, session recordings, support threads, and notes pulled into one view that proposes the next three moves.

Loop discipline: small, targeted releases to the smallest audience that can prove or falsify; automatic follow-ups to learn and log outcomes.

Behind the scenes we obsess over workflow completion rate, tokens per workflow, cost per workflow, model routing, and user privacy—so you can obsess over judgment.

A final word on taste

Rick Rubin calls himself a reducer. Film editors win Oscars because they cut the footage into a story people actually want to watch. That’s the work.

Automation is not your rival. It is the delete key that gives you room to practice taste.
Use it. Cut nine. Ship one.

All Posts

Templates

Research

widget
user pic

Templates

Aug 1, 2025

Hardware Product One-Pager (Spec → Shelf)

1) Hardware Product One-Pager (Spec → Shelf)

H1 (website): Hardware PRD that survives the real world
Subhead: From sensing and power math to drop tests—ship without “surprise failures.”

What you fill in (teaser):

Problem (1–2 lines), Success numbers, Non-goals

Sensing & actuation table, Power budget, Durability targets

Packaging, Serviceability, Launch gates


Full in-app template

Problem
One sentence. Who gets hurt today? Cost of not fixing?

Success Criteria (hard numbers)

First-pass yield at EVT/DVT: __% / __%

RMA rate: ≤ __% at 90 days

Setup time: ≤ __ minutes (p95)

Workflow completion rate (prototype→EVT→DVT): ≥ __%


Non-Goals
(Example: No multi-SKU colorways in v1. No Wi-Fi if BLE meets latency.)

System Requirements

Sensing: variable, unit, accuracy (±), sample rate ( Hz), calibration plan

Actuation: force/torque/speed; fail-safe defaults

Compute & Memory: min MCU/CPU class, RAM/Flash floors

Comms: protocol, throughput, latency budget

Power: chemistry, duty cycle, target life __h; charge time __h; thermal throttling behavior

Durability: UV/chemicals; IP rating target; drop spec (height × faces × repetitions)

Packaging: “in the box,” unboxing steps, retail constraints

Serviceability: repair vs replace, return flow, defect tolerances, SLAs


Key User Flows

Out-of-box → Activated
First success (task done)

Update (firmware)

Troubleshoot → Return


Acceptance Checks (Given-When-Then)

Given 1 m drop on each face ×10, when rebooted, then self-test passes and boots < 5s.

Given enclosure at IP54 test, when exposed to dust/splash, then no ingress.

Given duty cycle X for __ hours, then battery SOC ≥ __%.

Given BLE RSSI −85 dBm, when command sent, then control latency ≤ 200 ms.


Instrumentation & Events

  • device_boot, self_test_pass/fail, drop_event_detected, battery_soc, fw_update_success/fail.


Privacy & Data

  • Default: telemetry stored locally; cloud only with explicit opt-in. Retention: __ days if opted-in.

Tokens + Cost (if any AI on-device)

  • Tokens-per-workflow: ≤ __k

  • Cost-per-workflow: ≤ $.

  • Routing: small on-device model for anomaly; escalate to cloud model only if confidence < __.

Launch Gates
EVT exit, DVT exit, PVT readiness; pilot sample size; GA criteria.