Get In Touch
108 Wooster Street, New York, NY 10012,
hello@div.digital
Ph: +1.949.254.0339
Work Inquiries
hello@div.digital
Ph: +1.949.254.0339
Back

State-of-the-Art Prompting for AI Agents

YC’s Summary of Prompt techniques used by some of the best AI startups

Here’s a summary of key prompt engineering techniques used by some of the best AI startups:

  1. Be Hyper-Specific & Detailed (The “Manager” Approach):
  • Summary: Treat your LLM like a new employee. Provide very long, detailed prompts that clearly define their role, the task, the desired output, and any constraints.
  • Example: Parahelp’s customer support agent prompt is 6+ pages, meticulously outlining instructions for managing tool calls.
  1. Assign a Clear Role (Persona Prompting):
  • Summary: Start by telling the LLM who it is (e.g., “You are a manager of a customer service agent,” “You are an expert prompt engineer”). This sets the context, tone, and expected expertise.
  • Benefit: Helps the LLM adopt the desired style and reasoning for the task.
  1. Outline the Task & Provide a Plan:
  • Summary: Clearly state the LLM’s primary task (e.g., “Your task is to approve or reject a tool call…”). Break down complex tasks into a step-by-step plan for the LLM to follow.
  • Benefit: Improves reliability and makes complex operations more manageable for the LLM.
  1. Structure Your Prompt (and Expected Output):
  • Summary: Use formatting like Markdown (headers, bullet points) or even XML-like tags to structure your instructions and define the expected output format.
  • Example: Parahelp uses XML-like tags like <manager_verify>accept</manager_verify> for structured responses.
  • Benefit: Makes it easier for the LLM to parse instructions and generate consistent, machine-readable output.
  1. Meta-Prompting (LLM, Improve Thyself!):
  • Summary: Use an LLM to help you write or refine your prompts. Give it your current prompt, examples of good/bad outputs, and ask it to “make this prompt better” or critique it.
  • Benefit: LLMs know “themselves” well and can often suggest effective improvements you might not think of.
  1. Provide Examples (Few-Shot & In-Context Learning):
  • Summary: For complex tasks or when the LLM needs to follow a specific style or format, include a few high-quality examples of input-output pairs directly in the prompt.
  • Example: Jazzberry (AI bug finder) feeds hard examples to guide the LLM.
  • Benefit: Significantly improves the LLM’s ability to understand and replicate desired behavior.
  1. Prompt Folding & Dynamic Generation:
  • Summary: Design prompts that can dynamically generate more specialized sub-prompts based on the context or previous outputs in a multi-stage workflow.
  • Example: A classifier prompt that, based on a query, generates a more specialized prompt for the next stage.
  • Benefit: Creates more adaptive and efficient agentic systems.
  1. Implement an “Escape Hatch”:
  • Summary: Instruct the LLM to explicitly state when it doesn’t know the answer or lacks sufficient information, rather than hallucinating or making things up.
  • Example: “If you do not have enough information to make a determination, say ‘I don’t know’ and ask for clarification.”
  • Benefit: Reduces incorrect outputs and improves trustworthiness.
  1. Use Debug Info & Thinking Traces:
  • Summary: Ask the LLM to include a section in its output explaining its reasoning or why it made certain choices (“debug info”). Some models (like Gemini 1.5 Pro) also provide “thinking traces.”
  • Benefit: Provides invaluable insight for debugging and improving prompts.
  1. Evals are Your Crown Jewels:
  • Summary: The prompts are important, but the evaluation suite (the set of test cases to measure prompt quality and performance) is your most valuable IP.
  • Benefit: Evals are essential for knowing why a prompt works and for iterating effectively.
  1. Consider Model “Personalities” & Distillation:

Benefit: Optimizes for both quality (from larger models) and cost/latency (with smaller models).

Summary: Different LLMs have different “personalities” (e.g., Claude is often more “human-like,” Llama 4 might need more explicit steering). You can use a larger, more capable model for complex meta-prompting/refinement and then “distill” the resulting optimized prompts for use with smaller, faster, or cheaper models in production.

Full talk: State-Of-The-Art Prompting For AI Agents

At first, prompting seemed to be a temporary workaround for getting the most out of large language models. But over time, it’s become critical to the way we interact with AI. On the Lightcone Podcast, Garry, Harj, Diana, and Jared break down what they’ve learned from working with hundreds of founders building with LLMs: why prompting still matters, where it breaks down, and how teams are making it more reliable in production. They share real examples of prompts that failed, how companies are testing for quality, and what the best teams are doing to make LLM outputs useful and predictable.