AI-powered data collection

2026

How I designed the steps before full AI Automation

Before you can automate a system, someone has to unify it. Nobody had. I did.

I unified 30+ disconnected tools, designed the human-in-the-loop layer and one structure future automation could actually build on.

Problem area

The first time I sat with a DataOps contractor, she had 6 browser tabs open. She was moving values between them with keyboard shortcuts. No validation. No confirmation. Just copy, paste, switch tab, paste again.

When I asked if she double-checked her entries, she said: "I check if I have time."

She never had time.

The moment I understood what was actually broken

The Challenge

Everyone wanted automation.
No one asked if we had the foundation for it.

30+ tools. 0 common language

"You can't automate chaos. You can't scale what isn't unified. The business was asking for a ceiling, but I realized we were missing the floor."

The moment I understood what was actually broken

The Solution I proposed

Most people waited for better AI. I built what makes AI better.

I proposed Review Mode built into the collection flow. The AI pre-fills. The human reviews.

Pattern 01 — The Inversion

Every tool before this started here — blank table, no context, no errors surfaced. Submit available from the first second.

Pattern 02 — The Sequence

The researcher always knows exactly what's left.

Pattern 04 — Arrival State

The Architecture

A loop that teaches itself to need less of manual work.

Here's what I never said in any meeting: every improvement I made brought the full automation scenario closer. Every schema I standardized, every correction loop I built — all of it was teaching the model to need less human input over time. The researchers who helped me understand this workflow were, in doing so, helping me build the thing that would eventually reduce their role.

I resolved this by being honest. Not in a deck. In individual conversations. I showed them what the AI kept getting wrong. I made the case that their judgment, made systematic, was more valuable than their data entry — that the future version of their job was more interesting than the current one. Some believed me. Some didn't. I still don't know if I was right.

Every resolved error is structured, logged, and fed back into the model.

Researchers wouldn't be replaced. They'd be integrated — their judgment made systematic, at scale, feeding the model that would one day need less of it.

The Complexity

I designed for two states of AI maturity

AI Confidence score determines which mode

Mode When it triggers What the user sees Human role

Review

High confidence, non-sensitive Streamlined queue, AI-resolved items, undo available Spot-check only — minimal friction

Quality Check

Medium confidence, AI flag

raised AI suggestion + confidence rationale + accept/override Confirm or correct — one decision per flag

Edit

Low confidence or sensitive data Full data surface, source docs, AI rationale inline Full ownership — researcher takes

the wheel

Before I touched a single screen, I had to map what "unification" actually required. None of this was in the brief. I found it by asking one question at a time.

AI Confidence score determines which mode

The primary complexity lies in volume and variability of rows, not multi-period comparison

  • Design should prioritize:

    • speed

    • clarity

    • and low interaction cost per row

Design decision 02

Enforced Quality Check before submission

Business users often sent or approved payments in a hurry. The confirmation page missed key details they didn’t always check, leading to errors and rework. I redesigned it to clearly show account, amount, and recipient, and added a final double-check screen to save time and prevent costly mistakes.

Design decision 01

Three concepts.
Two were wrong for the same reason.

Review Mode had one job: help researchers resolve errors without losing context. That single constraint eliminated two of three directions immediately.

What I Created

4 things I built that will outlast this project.

The Resistance

The hardest work wasn't the data. Five people who thought this was a bad idea.

I expected structural complexity. I didn't expect how much of my energy would go into the people around the work.

Retrospective

The platform that needed
a trust layer became
the company's data spine.

Business owners spent less time managing transfer and felt more confident using mobile banking on the go. The new flow became one of the main reasons users upgraded to premium plans.

“Before the update, I used to double-check every transfer on desktop because I didn’t trust the app. Now it’s the opposite. I send everything from my phone. It’s fast, clear, and I know exactly what’s happening.”

Impact

The last time I watched someone use the platform we shipped,
they had one tab open.

I still think about the contractor who asked if we were going to replace her. The platform made her faster. It also made her more replaceable — which was always the point, even when nobody said it out loud. I made peace with that by believing the honest version of the work is better than the dishonest version. You design for what's coming. You try to make the transition humane. You don't always get to know if you succeeded.

What I'd do differently, and what I'd

defend

End with what this taught you about designing AI-assisted operational systems at scale.

What I'd do differently

Push for structured contractor sessions

earlier — not on the UI, but on where the

confidence threshold actually felt

trustworthy in practice. The threshold was

validated against Data Ops accuracy

numbers. The subjective experience of

"this feels safe enough to auto-process" is

a different question I could have tested

with real users before locking the model.

What I'd defend

Insisting on the undo + re-queue logic

under timeline pressure. It reads like a

nice-to-have. It's the mechanism that

makes the quality model honest — the

system acknowledges when the AI was

wrong, and routes it back to a human,

without breaking the workflow. I'm most

confident about that decision in retrospect.