Category 04 / Back-Office Support

The structured data layer your pipeline actually runs on.

Not generic BPO. Not data-entry sweatshops. This is the structured work that feeds annotation queues, normalises raw input, and keeps live AI pipelines fed with clean, schema-conformant data — by trained operators, not general-purpose staff.

ISO 27001 Certified · Cert No. 452AGI102121
Scope
PipelinePre-annotation · post-annotation · live
Operators
TrainedNot generic data-entry
Output
Your schemaCSV · JSON · custom
Throughput
ContinuousScales with your pipeline

Operating model

Not generic BPO. Data ops for AI.

01 / Structured
Every output matches your target schema — not a generic format we hand back
02 / Trained
Operators trained on your specific data type, not rotated from a general pool
03 / Measured
Sample-based QA on every batch, error rates reported weekly against the floor
04 / Secure
ISO 27001 ISMS, access-controlled environments for sensitive documents

Use cases

The work that makes your pipeline work.

Real scenarios we run today. Each one sits inside someone else’s AI stack — the invisible but critical layer that decides whether the rest of the pipeline is worth running.

Case 01

Pre-annotation data prep.

Converting raw scans, PDFs, or messy CSVs into structured records ready for labelling. Normalised fields, consistent identifiers, deduped at source — so the annotation team isn’t paid to clean data.

OutputSchema-conformant JSON / CSV bundles, versioned per batch
Case 02

Document digitisation at scale.

OCR cleanup, field validation, metadata tagging for archive corpora. Regulated paper input — forms, contracts, ledgers — turned into training-ready structured records inside a secured environment.

OutputDigitised records + per-document provenance trail
Case 03

Live pipeline feeding.

Continuous structured input into your production AI system — the human fallback layer. We handle the stream that doesn’t stop: inbound documents, exception items, content entering the model in real time.

OutputContinuous delivery against throughput & latency SLAs
Case 04

Schema migration.

Restructuring legacy databases into modern training-ready formats. Field mapping, type coercion, reconciliation across sources. Documented, reversible, versioned — not a one-shot script.

OutputMigrated dataset + mapping spec + reconciliation report
Case 05

Data enrichment.

Adding missing fields, normalising formats, deduplication at scale. Entity resolution, attribute completion, canonicalisation against an authoritative reference — the boring work that makes the model’s work possible.

OutputEnriched records with provenance for each added field
Case 06

Exception handling.

The judgment-call layer that catches what RPA and classifiers miss. Low-confidence items escalated to trained operators, resolved, and folded back into the pipeline with a written decision record for retraining.

OutputResolved exceptions + decision log for model feedback

How we deliver

Structured work. Structured process.

Four stages on every engagement. The edge cases get surfaced in Intake; nothing reaches Delivery that hasn’t been through Calibration.

Stage 01 / Intake

Profile the input.

We read your source data as it really is — formats, volumes, quality, edge cases. Dirty files, not idealised samples. The intake profile becomes the working document for the rest of the engagement.

Input profile + edge-case log
Stage 02 / Schema

Define the output.

We co-write the target output structure with your data lead. Types, cardinalities, required fields, nullability, the edge-case rules — documented and versioned before any operator touches the data.

Versioned schema + convention guide
Stage 03 / Calibration

Train on your data.

Operators run a calibration batch on your sample data. QA thresholds set against that batch, not a generic baseline. Anyone not hitting the floor doesn’t get production access — no exceptions.

Calibration report + QA thresholds
Stage 04 / Delivery

Continuous output.

Continuous delivery against written SLAs. Weekly quality reporting, escalation paths defined on day one, rework within SLA if we miss the floor. The pipeline stays fed — that’s the only metric that matters.

Weekly QA reports + exception log

Non-negotiables

Quality you can trace.

Back-office work earns its premium by being measurable. Every record is attributable; every batch has a QA footprint; every exception has a written resolution.

01 / Quality

Sampled, traced, reworked.

The quality footprint is built in, not bolted on. No batch ships without a QA pass; no output ships without per-operator attribution.

  • Sample-based QAStatistical sampling on every batch, sized to the batch — not a fixed token audit.
  • Error rate weeklyReported against the written floor; trend lines published with every delivery.
  • Rework within SLAIf we miss the floor, we re-run the affected slice on our time — no renegotiation.
  • Versioned schemasEvery schema change is dated, diffed, and reflected in the output manifest.
02 / Security

ISO 27001, honestly operated.

The certification isn’t a logo. It maps to actual operating controls — access, premises, data residency, exception handling — that get audited on their own cadence.

  • ISO 27001 ISMSCert No. 452AGI102121 — covers the annotation floor and back-office operations.
  • NDA standardSigned on every engagement before the first sample moves. Individual operator NDAs on sensitive accounts.
  • Access-controlledRole-based access, audited logins, clean-desk environments for regulated document work.
  • Data residencyNo data leaves the Nairobi environment without an explicit, written sign-off per engagement.

What we’ve delivered

Measurable by the batch. Honest by the number.

2M+
Records processed

Structured records produced against client schemas across back-office engagements to date.

99%+
Accuracy floor

Written threshold on structured-entry engagements. Sample-based QA on every batch; rework within SLA.

24–72h
Typical turnaround

Batch turnaround window on steady-state engagements. Throughput scales with the pipeline, not the backlog.

15+
Formats handled

PDF, scan, XLSX, CSV, JSON, image, and client-custom formats — handed in, handed out to your schema.

ISO 27001 Certified / Cert No. 452AGI102121 / Access-controlled · Nairobi

Scope with us

Send us a sample batch. We’ll return a scoped estimate.

We look at your input data, your target schema, your volume. Then we propose a team, an SLA, a price. No cold quotes, no volume-only pricing — the scope decides the number.

Clean input is the precondition for everything else.