§ 02 — AI Data Services / Vol. 11

The data layer
beneath frontier AI.

We've been training AI models for global tech giants since 2018 — ride-hailing platforms, social networks, e-commerce, foundation-model labs, and short-form video. Marano operates as a closed, NDA-bound studio: every annotator is recruited, screened, and audited under our own QA spine.

In production since
2018
Specialist annotators
1,200+
Active language pairs
40+
Domains served
7 verticals
§ 01 — Position

We didn't pivot into AI data when localization slowed down. We've been doing this work, in parallel, since 2018.

Our linguists, voice talents, and domain reviewers were already the operational backbone of high-stakes localization work. When global tech labs needed RLHF data, voice corpora, and content-moderation taxonomies, we adapted the same workforce — under the same QA spine — to the AI training brief.

That continuity matters. We don't crowdsource. We don't onboard anonymous workers behind a platform UI. Every name on a payroll we issue an NDA for has been calibrated, qualified, and audited — and most have been with us for years.

§ 02 — Service pillars

Four lines of work.

01Capability

Frontier Model Alignment

Reinforcement Learning from Human Feedback (RLHF), red-teaming, prompt-response pair authoring, instruction-tuning datasets, and human-preference comparison ranking for foundation-model labs.

  • RLHF preference comparisons
  • Supervised fine-tuning (SFT) data
  • Red-teaming & adversarial prompts
  • Constitutional / safety alignment data
  • Long-context evaluation sets
02Capability

High-Fidelity Annotation

Image, video, document, and 3D point-cloud annotation for computer vision, autonomy, content moderation, and document AI workloads — backed by a 4-stage QA pipeline.

  • 2D / 3D bounding boxes
  • Polygon & semantic segmentation
  • Keypoint & landmark labelling
  • Object tracking across frames
  • OCR & document layout parsing
  • Content moderation taxonomies
03Capability

Studio-Grade Voice Data

Multi-speaker, multi-domain speech collection inside our own ISO-grade booths. From neutral TTS corpora to spontaneous conversation, code-switching, and paralinguistic emotion sets.

  • Scripted TTS corpora (24 / 48 kHz)
  • Spontaneous & semi-spontaneous dialogue
  • Code-switching & multilingual speech
  • Emotion, prosody & paralinguistic tags
  • Far-field & noisy-environment capture
  • Speaker-balanced demographic sampling
04Capability

Multilingual Text & Evaluation

Machine-translation post-editing benchmarks, LLM output evaluation, multilingual NLU/NLG datasets, and culture-aware content review across 40+ language pairs.

  • MTPE benchmarking (BLEU/COMET-aligned)
  • LLM output rating & critique
  • Multilingual NLU intent / entity sets
  • Toxicity, bias & policy classification
  • Localized factuality verification
§ 03 — Capabilities matrix

Modality × language × scale.

Five modality families, each delivered at production volume. Per program we publish throughput targets, IAA bands, and reviewer composition before kick-off so the customer can model unit economics before signing.

Text · NLP

RLHF pairs, SFT data, prompt-response sets, MTPE evaluation, NLU intents, toxicity / policy classification

Coverage

40+ pairs, SEA & EA depth

Scale

Up to 100K samples / week

Image & Video

2D / 3D bounding boxes, polygon & semantic segmentation, keypoints, object tracking, content-moderation taxonomies

Coverage

Locale-aware moderation in 12+ markets

Scale

Up to 500K frames / month

Audio & Voice

TTS corpora, spontaneous dialogue, code-switching, emotion / paralinguistic tagging, wake-word & command sets

Coverage

TH, ID, VI, MS, TL, JA, KO, ZH, EN + dialects

Scale

Up to 1,000 speakers / quarter

3D & Sensor Fusion

LiDAR / radar / camera fusion ground-truth, ADAS / AV scene labels, point-cloud segmentation, driver-monitoring

Coverage

Spec language: EN / ZH; cabin-voice in 9 langs

Scale

Per program; pilots from 10K frames

Document AI

OCR, layout parsing, table extraction, redaction taxonomies, multi-language document classification

Coverage

Latin, CJK, Thai, Arabic scripts

Scale

Up to 250K pages / month

§ 04 — Workflow

Six steps from brief to delivery.

Every Marano AI program follows the same operational rhythm. Customers can plug in at any stage — brief design, calibration, production, or QA-only — without forcing a workflow rewrite.

Annotation pipeline abstract — overlapping vector polygons on a dark surface
Fig. 03Annotation rooms — closed network
  1. 01

    Brief & guideline ingestion

    Customer labelling guidelines, taxonomies, and edge-case decks are ingested by senior leads. We rewrite ambiguous instructions into testable rules before annotators touch a single sample.

  2. 02

    Workforce calibration

    Domain-screened annotators take blind qualification rounds against a gold-standard set. Pass rates are reported per cohort; only calibrated reviewers enter production.

  3. 03

    Production at scale

    Work is dispatched in batches sized to the program's throughput target. We hold daily stand-ups with team leads and weekly calibration syncs with the customer.

  4. 04

    Multi-stage QA

    Dual annotation, senior adjudication, and statistical sampling run on every batch. Inter-annotator agreement (IAA / Cohen's κ) is published per delivery.

  5. 05

    Spec evolution

    Edge-case rulings are looped back into the live guideline. Annotators are re-calibrated whenever the spec moves, so historical batches stay consistent with current intent.

  6. 06

    Delivery & traceability

    Customer-facing dashboards expose batch status, IAA, and reviewer identity (pseudonymous) so the customer can audit any sample back to its annotator and adjudicator.

§ 05 — Quality spine

Four-stage QA, by default.

Sampling rates, IAA reporting, and audit cadence are configurable per program. We publish per-batch agreement metrics so customers can monitor data drift in real time.

  1. 01

    Linguist Onboarding

    Domain-screened annotators sign NDAs, complete project-specific calibration, and pass blind qualification rounds before touching production data.

  2. 02

    Dual Annotation

    Every sensitive sample is independently annotated by two specialists. Disagreements are routed to senior reviewers, never silently averaged.

  3. 03

    Senior Adjudication

    Domain leads adjudicate edge cases against the customer's labelling guideline and feed clarifications back into the live spec.

  4. 04

    Statistical QA

    Sampled audits with inter-annotator agreement (IAA / Cohen's κ) reporting per batch, plus customer-facing dashboards for traceability.

§ 06 — Voice for AI

Studio-grade voice data. In our own booths.

We share voice infrastructure with our broadcast dubbing operation. The same booths that record anime and theatrical titles capture multi-speaker AI corpora — at 24 / 48 kHz, with demographic and dialectal balancing across speaker pools.

  • Scripted TTS corpora (24 / 48 kHz)
  • Spontaneous & semi-spontaneous dialogue
  • Code-switching & multilingual capture
  • Emotional / paralinguistic tagging
  • Far-field & noisy-environment scenarios
  • Speaker-balanced demographic sampling
  • Dialectal coverage across SEA / EA
  • Wake-word & command-set datasets
See Marano Studios
Marano voice studio
Fig. 04Voice booth A — 48 kHz capture
§ 07 — Case spotlights

Three programs, anonymized.

Most of our AI work runs under partner-LSP NDAs. We can't name the customer here, but the verticals, scope, and operational outcomes are real. Detailed program references available under a mutual NDA.

Case 01NDA

Global ride-hailing platform

RLHF & safety alignment data for an in-app conversational assistant covering Southeast Asian markets.

Scope

2M+ preference pairs, TH / ID / VI / EN, 18-month rolling engagement under partner LSP NDA.

Outcome

IAA ≥0.82 sustained across batches; spec drift caught and rolled back twice without customer escalation.

Case 02NDA

Foundation-model lab (frontier LLM)

Multilingual LLM evaluation — factuality, fluency, and culture-aware critique across SEA + EA pairs.

Scope

30K+ structured critiques per quarter, 9 target languages, senior reviewers from media, legal, and medical pools.

Outcome

Customer's published model cards cite the SEA evaluation set our linguists contributed to.

Case 03NDA

Short-video / social platform

Content-moderation taxonomy refinement and policy-aware video segmentation for a global UGC product.

Scope

500K+ video clips reviewed, 12 locales, dual annotation + senior adjudication on every sensitive sample.

Outcome

Reviewer attrition <8% / year; policy changes propagated to live spec inside one calibration cycle.

§ 08 — Tooling & security

We plug into your stack.

We do not require customers to migrate to a Marano-owned platform. We staff inside the customer's preferred tooling — managed platforms, open-source labelling tools, or custom pipelines — and run our QA spine on top.

Customer-owned platforms

We staff and manage workforces inside Scale AI, Appen, Toloka, Lionbridge, RWS / TrainAI and other partner-managed environments.

Open-source labelling tools

CVAT, Label Studio, Doccano, Prodigy, V7, Encord — configured per project, with reviewer roles mirrored to our internal hierarchy.

Custom pipelines & APIs

Direct ingest from customer S3 / GCS / Azure Blob, signed-URL workflows, and webhook delivery for batched outputs.

Data security posture

Closed network annotation rooms, mandatory NDAs, device-level controls, and per-program access logs. We can operate inside the customer's VDI when required.

The hardest part of frontier AI work isn't the labelling — it's making sure the same person who labelled batch zero is still making the same call on batch three hundred. That's what our QA spine is for.

Senior AI Programs Lead — Marano, internal note
§ 09 — Partner roster

AI work we've contributed to.

Selected engagements delivered directly or through Leading Language Service Providers under strict NDA. Detailed program scopes available on request.

UberAI data services via partner
AlibabaAI data services via LSP partner
TencentAI data services via LSP partner
§ 10 — Engage

Send us a hard AI brief.

Multilingual, multi-domain, production-volume work where the quality bar is non-negotiable. Pilots can start within two weeks.