NEU

Zylon in a Box: Plug & Play Private KI. Holen Sie sich einen vorkonfigurierten On-Premise-Server, der lokal einsatzbereit ist, ohne Cloud-Abhängigkeit.

Zylon in a Box: Plug & Play Private KI. Holen Sie sich einen vorkonfigurierten On-Premise-Server, der lokal einsatzbereit ist, ohne Cloud-Abhängigkeit.

Zylon in a Box: Plug & Play Private KI. Holen Sie sich einen vorkonfigurierten On-Premise-Server, der lokal einsatzbereit ist, ohne Cloud-Abhängigkeit.

Veröffentlicht am

·

8 minutes

When AI lives in your building, prompt engineering comes back (and that's a good thing)

Ana Carolina Sanchotene Silva

Kurze Zusammenfassung

Frontier cloud models hide ambiguity behind computing power, Small Language Models running on your own servers don't and that's exactly why they're the right fit for most enterprise work. This post explains why on-premise AI is a hardware and economics story before it's a model story, why SLMs are well-matched to the bounded, high-volume tasks that actually move the P&L, and how prompt engineering is the cheap, learnable discipline that unlocks them. The one-page SLM prompting guide is included at the end.

Prompt engineering has been declared dead for a while now. The story goes: as models get smarter, you just describe what you want in plain language and they figure out the rest. No structure, no constraints, no formatting discipline required.

That story is true for the big cloud models running on someone else's infrastructure, with your data leaving your perimeter every time you press enter. For companies that can't or won't send their data to a third-party cloud — banks, insurers, hospitals, manufacturers, engineering firms, public sector, defense contractors — the math is entirely different. Private AI means running models on your own servers. And running models on your own servers, today, means working with smaller, leaner models. That's not a limitation to apologize for. It's the right architecture for most of the work you need to automate.

The hardware story behind the model story

It helps to understand why SLMs matter at all. ChatGPT-class and Claude-class models are the output of an enormous compute apparatus. Training and serving them economically depends on tens of thousands of high-end accelerators, exotic networking, and power footprints that look more like data center campuses than racks. That kind of infrastructure is owned and operated by a small number of hyperscalers and AI labs. Replicating it inside a bank, a hospital, a manufacturing plant, or a ministry isn't a budget question, it's a different business entirely. Which is exactly why the conversation about enterprise AI has to start somewhere else.

The frontier model running in someone else's cloud is, by definition, dependent on that cloud — its uptime, its pricing, its export controls, its terms of service, its location, its retention policies, and its willingness to keep selling to your sector. That dependency is both technical and strategic.

Small Language Models in the 7B to 70B parameter range — the Qwen, Mistral, Gemma, and Llama families and their derivatives — are designed to run on hardware you can actually buy, install, and own. A handful of GPUs in a server you control. On-prem, in a sovereign cloud, behind your firewall. The data never leaves. The model doesn't change unless you want it to. The reliance on a third party effectively goes away.

That shift is the whole point of private, on-premise AI for regulated industries. You're not chasing the absolute frontier of intelligence; you're trading a few percentage points of generalist capability for sovereignty, predictable cost, auditability, and the ability to keep operating regardless of what happens upstream. For most enterprise work, that's a great trade.

Most business work doesn't need maximum intelligence

Here is the part that gets under-appreciated in the cloud-vs-local debate: the vast majority of work that costs companies real time and money isn't open-ended reasoning. It's bounded and repetitive: invoices processed a thousand times a month, contract clauses extracted across hundreds of documents, maintenance reports drafted from the same structured telemetry every week, the list goes on and on.

None of this needs PhD-level open-ended brilliance, it needs reliability, auditability, and privacy. A well-configured 7B–70B model running on your own infrastructure delivers exactly that and arguably delivers it better than a hyperscaler model would, because you can pin the version, integrate with your internal tools without fears of your API keys getting leaked, log every call, and be sure no token ever crosses your perimeter.

The big cloud models are extraordinary generalists, and where you genuinely need open-ended reasoning, creative writing, or hard multi-step thinking, use them. But pointing the world's most capable generalist at a structured invoice extraction job is like hiring a heart surgeon to take blood pressure. It works but is it optimized in cost and resources?

The reframe: SLMs are specialists, not consolation prizes

The most useful mental shift for an enterprise AI program is to stop ranking models on a single "intelligence" axis and start matching tools to tasks.

  • Frontier cloud model: a brilliant generalist consultant you fly in for the hardest, most ambiguous problems, when the data and the use case allow it.

  • SLM on your servers: a focused, deeply embedded operator that handles the repetitive, high-volume work where you can provide rich context such as historical data, clear formatting guidelines, defined outputs, well-understood edge cases.

Both can coexist in an enterprise AI strategy. A few things SLMs are genuinely strong at, today, and worth leaning into:

  • Structured extraction and classification — pulling fields from documents, tagging tickets, normalizing data. With a good prompt and examples, accuracy is often comparable to the largest models for the same task.

  • Drafting from templates — maintenance reports, internal memos, customer responses, regulatory filings, anywhere the shape of the answer is known.

  • Retrieval-augmented Q&A over your own corpus — the model doesn't need to know the world; it needs to read your documents well, which SLMs do.

  • Agentic workflows on bounded tools — with clear tool definitions and tight prompts, SLMs drive multi-step automations reliably.

  • Local, fast, cheap inference at scale — once it's on your hardware, the marginal cost of an additional call collapses and that changes what's economically worth automating.

What you trade

SLMs have shorter effective context windows than the largest cloud models. Many open-weight models in this range advertise long context (some Qwen3 variants reach 256k tokens, it's enough to fit Crime and Punishment by Dostoevsky but it would be a bit too small for Ulysses by James Joyce), but practical retrieval quality may degrade well before the advertised limit. Architecture, not advertised numbers, is what holds up in production.

Precise instructions get precise outputs while vague ones get inconsistency. Multi-step reasoning holds up when you scaffold the steps, and falls apart when chains run long without checkpoints. They don't know what "professional" or "concise" means at your firm unless you tell them which, to be fair, is also true of the big cloud models. And they pay more attention to instructions at the top of a prompt than the bottom, so put the critical rules first.

None of this makes them bad. It makes them predictable. They reward careful engineering and they punish hand-waving. That's a feature for regulated industries, where vague systems are a compliance problem, not a productivity win.

Prompt engineering is the cheap part of the stack

Here's the encouraging part. Hardware is expensive. Building good retrieval is real engineering work. Prompt engineering, by comparison, is almost free, and it's the single highest-leverage skill your team can develop right now.

The teams getting real value from private AI aren't the ones who installed the cleverest model. They're the ones who built the discipline to communicate with it properly: writing system prompts that define role and context precisely, specifying constraints with numbers instead of adjectives, showing the model what good looks like, anticipating what can go wrong, and building validation into every critical step.

Do that, and a 32B model on a server in your own data center will quietly carry an enormous amount of operational load, reliably, cheaply, and without any external reliance you didn't sign up for. That's not a worse version of frontier AI, it's a better fit for the actual job since SLMs force users to define structures, guardrails, and provide examples, the results are more likely to be consistent over time and different models, if you ever decide to update.

We put the principles that moved the needle in our own deployments into a one-page reference. Download the SLM prompting guide and keep it close.

Guide

SLM prompting: the essentials

Principle 01 - Always define the role

Small models need to be told who they are in this context. Without a role, they default to "generic assistant" mode, which is rarely what you want.

✗ weak

Write a response to this customer complaint
Write a response to this customer complaint
Write a response to this customer complaint

✓ strong

You are a customer service representative for a B2B software company. Write a professional email response (2–3 paragraphs, max 200 words) that acknowledges the issue, empathizes with the client, and offers a concrete resolution path
You are a customer service representative for a B2B software company. Write a professional email response (2–3 paragraphs, max 200 words) that acknowledges the issue, empathizes with the client, and offers a concrete resolution path
You are a customer service representative for a B2B software company. Write a professional email response (2–3 paragraphs, max 200 words) that acknowledges the issue, empathizes with the client, and offers a concrete resolution path

Principle 02 - Use exact commands, not vague guidance

Words like "concise", "brief", or "a few" are meaningless to small models. Replace them with hard limits. When you tell a 7B model to "keep it brief", it will write 500 words. When you say "maximum 80 words", it will be concise, although SLMs are unable to count well and the exact number of characters might vary around 10%.

✗ weak

 Keep it concise. Mention key features. Make it compelling
 Keep it concise. Mention key features. Make it compelling
 Keep it concise. Mention key features. Make it compelling

✓ strong

Exactly 3 sections:
Section 1 Hook (30 words)
Section 2 Features (60 words, list 4 items)
Section 3 CTA (20 words)
Total: 110 words maximum.
If you exceed 110 words: cut

Exactly 3 sections:
Section 1 Hook (30 words)
Section 2 Features (60 words, list 4 items)
Section 3 CTA (20 words)
Total: 110 words maximum.
If you exceed 110 words: cut

Exactly 3 sections:
Section 1 Hook (30 words)
Section 2 Features (60 words, list 4 items)
Section 3 CTA (20 words)
Total: 110 words maximum.
If you exceed 110 words: cut

For instance, if you need a paragraph with exactly three phrases, it works to tell the model to count and enumerate:

✗ weak

Section 1 Hook (30 words - three phrases)
Section 1 Hook (30 words - three phrases)
Section 1 Hook (30 words - three phrases)

✓ strong

Section 1 Hook (30 words - three phrases. Count the phrases: 1, 2, 3)
Section 1 Hook (30 words - three phrases. Count the phrases: 1, 2, 3)
Section 1 Hook (30 words - three phrases. Count the phrases: 1, 2, 3)

Principle 03 - Show, don't only tell — use examples

For any output format you care about, include at least one complete example of what good looks like. For critical tasks, also show what bad looks like and explain why. Small models calibrate on examples far better than on abstract descriptions.

✗ weak

Format each action item professionally with an owner and deadline
Format each action item professionally with an owner and deadline
Format each action item professionally with an owner and deadline

✓ strong

Format each action item like this:
 [Owner]: [Task] by [Date]

Example:
Maria: Send revised contract to client by Friday
Tom: Schedule Q3 review by EOD Tuesday
Format each action item like this:
 [Owner]: [Task] by [Date]

Example:
Maria: Send revised contract to client by Friday
Tom: Schedule Q3 review by EOD Tuesday
Format each action item like this:
 [Owner]: [Task] by [Date]

Example:
Maria: Send revised contract to client by Friday
Tom: Schedule Q3 review by EOD Tuesday

Principle 04 - Front-load critical instructions

Models at 7B–14B parameters have a tendency to drift from instructions placed late in a long prompt. Put your most important constraints in the first third of the prompt. Then repeat them in a validation checklist at the end.

✓ structure template

## Your role
[Who you are + what task you're doing]

## Critical rules  put constraints HERE first
- Output only valid JSON
- Never add explanatory text outside the JSON
- Never invent values not present in the input

## Input
[your data]

## Before you respond verify:
1. Is the output valid JSON? (check again)
2. Are all values sourced from the input

## Your role
[Who you are + what task you're doing]

## Critical rules  put constraints HERE first
- Output only valid JSON
- Never add explanatory text outside the JSON
- Never invent values not present in the input

## Input
[your data]

## Before you respond verify:
1. Is the output valid JSON? (check again)
2. Are all values sourced from the input

## Your role
[Who you are + what task you're doing]

## Critical rules  put constraints HERE first
- Output only valid JSON
- Never add explanatory text outside the JSON
- Never invent values not present in the input

## Input
[your data]

## Before you respond verify:
1. Is the output valid JSON? (check again)
2. Are all values sourced from the input

Principle 05 - Handle edge cases explicitly

Never leave a small model to decide what to do when the input is missing, ambiguous, or unexpected. It will guess but there’s a chance it’s wrong. Every "if X is missing" case you leave unspecified is a failure mode waiting in production.

✓ edge case handling

## Edge case rules
- If date is missing: use "date_unknown"
- If amount has 1 decimal (e.g. 12.5): add zero 12.50
- If post is fewer than 10 characters: return category "Other"
- If field is ambiguous between 2 categories: use the first one mentioned
## Edge case rules
- If date is missing: use "date_unknown"
- If amount has 1 decimal (e.g. 12.5): add zero 12.50
- If post is fewer than 10 characters: return category "Other"
- If field is ambiguous between 2 categories: use the first one mentioned
## Edge case rules
- If date is missing: use "date_unknown"
- If amount has 1 decimal (e.g. 12.5): add zero 12.50
- If post is fewer than 10 characters: return category "Other"
- If field is ambiguous between 2 categories: use the first one mentioned

Principle 06 - Make the model self-check before responding

A simple trick that significantly improves output reliability: tell the model to verify its answer before it writes it. This is especially valuable for structured outputs like JSON, tables, or strict-format summaries.

✓ validation checklist

Before you output anything, verify:
1. merchant_name: not empty
2. date: matches YYYY-MM-DD exactly
3. total: equals sum of item prices × quantities
4. JSON: no trailing commas

If any check fails: correct it before responding

Before you output anything, verify:
1. merchant_name: not empty
2. date: matches YYYY-MM-DD exactly
3. total: equals sum of item prices × quantities
4. JSON: no trailing commas

If any check fails: correct it before responding

Before you output anything, verify:
1. merchant_name: not empty
2. date: matches YYYY-MM-DD exactly
3. total: equals sum of item prices × quantities
4. JSON: no trailing commas

If any check fails: correct it before responding



Author: Ana Carolina Sanchotene Silva, Quality & Security Engineer at Zylon
Published: May 2026
Ana Carolina Sanchotene Silva is a Quality & Security Engineer focused on managing customer journey risk in AI-powered products. Her work brings together QA, security, compliance, and AI evaluation to identify where prompts, workflows, and model behavior can break down in real customer interactions. With experience in regulated environments, she understands how product reliability, security, and trust intersect across the customer journey.

Veröffentlicht am

Geschrieben von

Ana Carolina Sanchotene Silva