AI Strategy 8 min read September 5, 2025

Prompt Engineering That Actually Works: A Practical Guide for Business

Stop getting generic AI outputs. Learn the iterative, test-driven approach to prompt engineering that turns AI from frustrating to reliable for B2B service businesses.

Prompt Engineering That Actually Works: A Practical Guide for Business

The Shortcut Problem

Most business owners expect AI to read their minds. Type a vague request, hope for brilliance, blame the technology when it doesn’t deliver.

I’ve seen this pattern repeat across dozens of client conversations. Someone tries ChatGPT or Claude for a business task, gets mediocre output, and concludes that “AI isn’t ready yet” or “it doesn’t understand my business.”

The problem is the prompt.

After building production AI systems that generate proposals, analyze sales calls, and automate client communications, I’ve learned that prompt engineering is about iteration, testing, and treating prompts like code.

Why Most AI Prompts Fail

Before we fix the problem, let’s understand what causes it. These five mistakes explain 90% of disappointing AI outputs.

1. Vague Instructions

Bad: “Write me something about marketing.”

Better: “Write a 300-word LinkedIn post about how B2B service businesses can use AI to automate proposal generation. Target audience: agency owners aged 35-55. Tone: professional but approachable. Include one specific example.”

Vague prompts force the AI to guess what you want. Specific prompts eliminate guesswork.

2. Missing Business Context

AI doesn’t know your company’s voice, your industry jargon, or your client relationships. You need to provide that context explicitly.

We learned this the hard way when implementing an automated email system for a client. The AI was generating technically correct emails, but they didn’t sound like the client’s brand at all. The fix was adding brand voice guidelines and example emails to every prompt.

3. Overloading Single Prompts

Asking AI to do five things at once leads to mediocre results on all five. Break complex tasks into sequential steps, where each prompt focuses on one thing.

Instead of: “Research this prospect, write an email, suggest a meeting time, and format it for our CRM.”

Try: Step 1: Research the prospect. Step 2: Draft the email using the research. Step 3: Format for CRM import.

4. No Format Specification

Do you want a paragraph, a bulleted list, a table, or JSON? If you don’t specify the format, you’ll get whatever the AI feels like generating.

Always include: “Return your response as [format] with [specific structure].“

5. One-and-Done Mentality

The biggest mistake is treating prompts as one-shot attempts. Effective prompt engineering requires iteration. Your first prompt will almost never be your final prompt.

The Iterative Approach: Test-Driven Prompting

Here’s what actually works: treat prompts like code. Build incrementally. Test against success criteria. Refine based on failures.

The TDD Loop for Prompts

Step 1: Define success criteria first.

What does “good” look like? Be specific. If you’re generating proposals, success might mean: “Includes company background, addresses their stated pain points, proposes a pilot project, and matches our brand voice.”

Step 2: Write the simplest prompt that might work.

Start basic. Don’t over-engineer from day one. You need to see what the AI produces before you know what to fix.

Step 3: Test against real examples.

Run your prompt with actual inputs from your business. Not hypotheticals. Real client names, real situations, real data.

Step 4: Identify failure modes.

Where does it break? Does it hallucinate facts? Miss the tone? Include irrelevant information? Fail on edge cases? Document every failure pattern you find.

Step 5: Refine and add specificity.

Each failure mode needs a fix. If the AI adds too much context about well-known companies, add: “Do not include publicly known information about the company.” If it sounds too formal, add example outputs that match your desired tone.

Step 6: Repeat until consistent.

Keep testing until you get reliable, predictable results across different inputs. This might take 5 iterations or 50. Don’t stop until you trust the output.

A Real Example: AI Hallucination Discovery

When building a client information system, we discovered that the AI was fabricating client relationships. It would confidently state that “Company X has been a client since 2019” when no such relationship existed.

The fix wasn’t obvious. We had to add explicit instructions: “Only include information that is explicitly provided in the input data. Do not infer or assume any client relationships. If information is missing, state that it is unknown.”

One prompt modification eliminated the problem entirely. But we only found it by testing with real data and catching the failure.

Six Techniques from Anthropic That Work

Anthropic’s prompt engineering documentation outlines techniques in order of impact. Here’s how to apply them in B2B contexts.

1. Be Clear and Direct

Explicit beats implicit. Don’t assume the AI understands what you want.

Before: “Help me with this proposal.”

After: “You are a proposal writer for a B2B AI automation consultancy. Write a proposal for a marketing agency that wants to automate their client reporting workflow. Include: problem statement (their words), proposed solution, timeline (4 weeks), investment (€5,000-€8,000 range), and expected ROI (hours saved per month).“

2. Use Examples (Few-Shot Learning)

Providing 2-3 examples of desired output dramatically improves consistency. This is particularly powerful for maintaining brand voice and handling edge cases.

Include both good examples and edge cases. Show the AI what you want and what to avoid.

3. Let It Think (Chain of Thought)

For complex decisions, tell the AI to reason step by step. This reduces errors in multi-step analysis.

Example: “Before generating the proposal, first list the client’s stated pain points from the meeting notes. Then identify which of our services addresses each pain point. Finally, write the proposal using only relevant services.”

4. Use Structured Tags

Separate context from instructions using clear delimiters. This makes prompts scannable and maintainable.

<context>
You are writing for a B2B AI consultancy.
Target audience: service business owners.
Tone: professional, direct, no jargon.
</context>

<input>
{client_meeting_notes}
</input>

<task>
Generate a follow-up email summarizing the meeting and proposing next steps.
</task>

5. Assign a Role

“You are a senior business analyst with 10 years of experience in B2B services” sets expectations for expertise level and communication style.

Match the role to the task. An email needs a different role than a technical analysis.

6. Chain Complex Tasks

Break big workflows into sequential prompts. Each step produces output that feeds the next step.

For lead research: Prompt 1 gathers company information. Prompt 2 identifies decision-makers. Prompt 3 drafts outreach messaging. Each prompt is simpler and more reliable than trying to do everything at once.

Building Your Prompt Library

Once you have prompts that work, document them. Version control them. Treat them as business assets.

For recurring tasks, create templates:

  • Client onboarding questionnaire analysis
  • Proposal generation from meeting notes
  • Meeting summary extraction
  • Lead qualification scoring
  • Email response drafting

Test new prompts against the old before switching. A “better” prompt that breaks edge cases isn’t actually better.

Share working prompts across your team. Prompts that work for one person should work for everyone using the same inputs.

Measuring Prompt Performance

Define success metrics upfront:

  • Accuracy: Does the output contain correct information?
  • Consistency: Does it work 9 out of 10 times with different inputs?
  • Tone: Does it match your brand voice?
  • Completeness: Does it include all required elements?

Track time saved. If a prompt generates proposals that need 30 minutes of editing, compare that to manual creation time. The goal is net time saved, not perfection.

Know when to stop iterating. At some point, additional refinement has diminishing returns. If your prompt works 95% of the time, that’s probably good enough.

The Compound Effect

Good prompts compound over time. A proposal prompt that saves 2 hours per deal, used 50 times per year, saves 100 hours annually. Multiply that across email drafting, research, reporting, and other tasks.

The businesses that see 4-5x ROI from AI use well-engineered prompts.

Getting Started

Pick one workflow. One task you do repeatedly. Build one prompt using the iterative approach described above. Test it with real data. Refine until it works consistently.

Then move to the next workflow.

Ready to see how prompt engineering and AI automation can transform your operations? Schedule a free AI Readiness Assessment and we’ll show you exactly where to start.

#prompt engineering #AI best practices #ChatGPT #Claude #B2B automation
Thom Hordijk
Written by

Thom Hordijk

Founder

Get posts like this in your inbox every week

Weekly insights on AI and automation for B2B service businesses. No hype, just what works.