✍️ Prompts

What Is Prompt Engineering?

The discipline of reliably extracting high-quality outputs from language models.

Beginner9 min readFebruary 15, 2026ProBotica Editorial

Why the Same Model Gives Wildly Different Results

A language model is a function: given input text, it produces output text. The same model will give completely different responses to semantically similar inputs phrased differently. Ask "What are some ideas for marketing?" and you get a generic list. Ask "You are a senior B2B SaaS marketing strategist. List 5 demand-generation tactics for a €1M ARR company targeting enterprise CTOs in the DACH region, ranked by expected ROI within 90 days" and you get an entirely different, far more useful response.

The model did not change. The prompt did. This is the core insight of prompt engineering: **the quality of an AI output is determined at least as much by the quality of the prompt as by the capability of the model.**

Prompt engineering is not about manipulating or tricking AI systems. It is about communicating clearly, specifying context and constraints, providing examples of desired output, and structuring requests to leverage the model's capabilities most effectively. It is, in essence, the art of communicating with a very capable but literal system.

Anatomy of a Production Prompt

Production prompts for business applications typically have four components:

**System Prompt** (also called the instruction prompt or prefix): Defines the AI's role, personality, rules, knowledge boundaries, and output format. This is set by the developer and hidden from the end user. A good system prompt specifies: who the AI is ("You are ProBotica's lead qualifier agent"), what it knows ("You have access to our pricing tiers and qualification criteria"), what format to use ("Return a JSON object with fields: score, tier, reasoning, nextAction"), and what it must not do ("Never guarantee pricing; always suggest a demo for scores above 70").

**Context**: Relevant background information the model needs to complete the task. For a support agent, this might include the customer's account history, the product documentation, and recent tickets. For a lead qualifier, it might include the company's ICP (ideal customer profile) definition.

**Few-Shot Examples**: Demonstrations of the desired input-output behaviour. The single most effective technique for controlling output format and style. If you show the model three examples of how you want it to respond, it will generalise that pattern far more reliably than any amount of verbal instruction.

**Task**: The specific instruction for this invocation. Clear, specific, imperative phrasing. "Analyse the following support ticket and classify it as billing, technical, feature-request, or account-management. Return only the JSON object."

text
SYSTEM:
You are ProBotica's B2B lead qualifier. Your task is to assess
inbound leads and return a structured JSON qualification score.

Scoring criteria:
- Company size 50-500 employees: +20 points
- Has dedicated IT/ops team: +15 points
- Currently uses manual workflows: +25 points
- Budget mentioned > €3k/month: +20 points
- Decision maker is contact: +20 points

Output format: { "score": 0-100, "tier": "hot|warm|cold",
                 "reasoning": "...", "nextAction": "..." }

USER:
Lead: "Hi, we're a 200-person logistics company. Our ops team of
8 manually processes 400 invoices per week in Excel. I'm the COO
and we have budget for the right solution."

A production-quality lead qualification prompt with scoring criteria and typed JSON output.

Core Techniques: From Few-Shot to Chain-of-Thought

**Zero-shot prompting**: No examples provided. Works for simple, well-defined tasks the model has seen in training. Unreliable for complex or format-specific outputs.

**Few-shot prompting**: 2-5 examples of input → output pairs. Dramatically increases reliability and controls format. This is the most practical technique for business applications. The examples do not need to be from your actual use case — even synthetic examples of the desired format improve performance significantly.

**Chain-of-Thought (CoT)**: Adding "Let's think step by step" or structuring the prompt to elicit explicit reasoning before the final answer. Dramatically improves performance on multi-step reasoning, mathematics, and logical analysis. The model "shows its work," and the intermediate steps improve the final answer.

**Structured output / JSON mode**: Instructing the model to return structured data (JSON, XML) rather than prose. Combined with output validation (Zod, Pydantic), this makes AI outputs type-safe and programmatically processable. This pattern is the foundation of every AI-powered business workflow.

**Role prompting**: Assigning a specific persona ("You are a senior tax attorney," "You are a hostile code reviewer"). Activates relevant knowledge patterns and tonal conventions in the model.

**Self-consistency**: Running the same prompt multiple times and taking the majority answer. Increases reliability on reasoning tasks at the cost of latency and compute.

Tip

The single fastest improvement for any production prompt: add 3 worked examples of ideal input/output pairs. Few-shot prompting is consistently the highest-ROI prompt engineering technique across all task types.

Key Takeaways

  • Prompt engineering is the craft of writing instructions that reliably elicit correct, useful, formatted AI outputs.
  • System prompts define persona, rules, and output format; user prompts carry the specific task.
  • Few-shot examples are the single most effective technique for controlling output format and style.
  • Chain-of-Thought prompting dramatically improves performance on multi-step reasoning tasks.
  • Prompts are code: they should be versioned, tested, and reviewed like any other engineering artefact.