Hamel Husain’s insights for developing and releasing a large language model based product1:

  1. Build a tool that allows you to view and annotate conversations with the LLM for real users. Getting close to the data in this way will allow you to see real trends and blockers. Focus on improve the failures and measure the improvement. Prefer these metrics over squishier metrics.
  2. Get domain experts involved in prompt writing, but tools in their hands to do it.
  3. When data is not available, use LLMs to generate synthetic.
  4. Focus on binary outcomes for LLM success (Pass/Fail) rather than a gradients or scales. This gives you clear outcomes
  5. You can scale by keeping humans closer in the loop and gradually leaning on AI to decide pass/fail. Finally, you can move to a statistical sample approach to maintain on going calibration
  6. AI Projects and Project Management do not look like typical product management where features are discrete and knowable. AI projects are much more experimental and iterative and so framing this up for stakeholders is important.
    1. Update stakeholders on a regular cadence with how experiments are progressing.

1. Husain, H. A Field Guide to Rapidly Improving AI Products – Hamel’s Blog. https://hamel.dev/blog/posts/field-guide/ (2025).