Back to blog
    Getting Doctors to Label Data: Change Management for AI Data Preparation
    change-managementdomain-expertsadoptionlabelingsegment:enterprise

    Getting Doctors to Label Data: Change Management for AI Data Preparation

    Domain experts have the knowledge AI needs, but most labeling tools weren't built for them. Here's how to design adoption programs that get doctors, lawyers, and engineers labeling data willingly.

    EErtas Team·

    You need a radiologist to label 500 chest X-rays. A construction engineer to classify 1,200 specification clauses. A corporate lawyer to annotate 800 contract provisions. These domain experts have the knowledge your AI model needs — the pattern recognition that took them a decade to develop, encoded in every labeling decision they make.

    There's one problem: they didn't sign up for this. They're busy. They have patients, projects, and cases. And the last time someone asked them to "help with the AI project," they spent 45 minutes trying to log into a web application before giving up.

    Getting domain experts to participate in data labeling is not a technology problem. It's a change management problem. And the solution requires understanding why they resist, designing around their constraints, and demonstrating value that matters to them.

    The Resistance Patterns

    Domain expert resistance to data labeling follows four predictable patterns. Identifying which pattern you're facing determines which intervention works.

    "I'm Too Busy"

    This is the most common response, and it's usually genuine. A cardiologist seeing 25 patients per day does not have a spare hour for data labeling. A construction project manager overseeing three active sites has no slack in their schedule.

    The mistake is asking for an hour. The solution is asking for 15-20 minutes. Twenty minutes per day — the length of a coffee break — produces 15-30 labeled examples depending on complexity. Over a month, that's 300-600 examples from a single expert. Across three experts, that's nearly 1,800 examples — enough to fine-tune a model for many classification tasks.

    The math works, but only if the tool makes those 20 minutes productive. If the expert spends 10 minutes logging in and navigating to their queue, you've lost half the session. If the tool requires explanation every time, you've lost the rest.

    "This Isn't My Job"

    This objection reflects a role boundary concern. Doctors are hired to treat patients. Lawyers are hired to advise clients. Labeling data feels like IT's problem, not theirs.

    The reframe: they are not labeling data for IT. They are teaching the AI to do what they do, which directly improves the tools they use. The radiologist labeling chest X-rays is training the AI that will pre-screen their cases and flag the urgent ones. The lawyer annotating contracts is building the system that will handle routine clause review so they can focus on complex negotiations.

    This reframe only works if it's true. If the domain expert's labeling effort feeds into a model they'll never use, the pitch rings hollow. The connection between their effort and their benefit must be concrete and visible.

    "The Tool Is Too Complicated"

    Domain experts are intelligent but not technical. A surgeon who performs laparoscopic procedures has fine motor skills and spatial reasoning that most software developers lack. But ask that surgeon to set up a Python virtual environment, install dependencies, and launch a Jupyter notebook for labeling — and they'll close their laptop.

    The tool must be as simple as the expert's existing workflow tools. If they review documents in a PDF viewer, the labeling interface should feel like a PDF viewer with annotation buttons. If they dictate clinical notes, the labeling interface should accept voice input. If they work on an iPad between patients, the interface must work on a tablet.

    The benchmark is this: can the expert start labeling within 60 seconds of opening the application, with zero training? If not, the tool is too complicated for adoption.

    "I Don't Trust How This Data Will Be Used"

    This is the most serious objection, and it's especially common in healthcare and legal contexts. Experts worry about:

    • Patient data leaving the hospital network
    • Client communications being exposed
    • Their professional judgment being used to replace them
    • Their name being attached to model outputs they can't control

    Addressing this requires transparency, not reassurance. Show them exactly how the data flows: where it's stored (on-premise, never leaves their network), who has access (named individuals, not "the AI team"), how their labels are used (to train a model for a specific purpose, documented in writing), and what happens to the data after the project (retained under their organization's data retention policy, not repurposed).

    For regulated industries, have the compliance team review and approve the data handling process before approaching domain experts. Presenting a compliance-approved workflow eliminates the expert's concern and demonstrates organizational seriousness.

    The Six-Part Adoption Framework

    1. Make the Value Visible

    Before asking anyone to label anything, demonstrate what the AI does now (poorly) and what it could do with their input (well). Show a concrete before-and-after.

    For a clinical text classification model: "Right now, the model correctly categorizes 67% of clinical notes. With 500 labeled examples from physicians, we expect to reach 91%. That means 24% fewer misrouted notes, which saves approximately 3 hours of administrative correction per day across the department."

    Specific numbers matter. "It'll be better" doesn't motivate a busy professional. "It'll save your department 3 hours daily" does.

    2. Minimize Friction

    Every step between "I have 15 minutes" and "I'm labeling" must be eliminated or automated.

    No portals. Don't require a separate login, a VPN connection, or a browser bookmark. The application should be on their desktop or home screen.

    No setup. No Python, no terminal, no environment configuration. Click the icon, see the queue, start labeling.

    No context switching. Ideally, labeling happens within the tool they already use — or the labeling tool mimics the document viewer they're familiar with.

    No waiting. The queue should load instantly. If there's a 5-second loading spinner, busy professionals interpret it as "this tool wastes my time" and close it.

    Immediate save. Every label should persist the moment it's applied. If the expert is interrupted (which happens constantly in clinical and field settings), no work should be lost.

    3. Integrate Into Existing Workflow

    The ideal: labeling happens as a natural extension of work the expert is already doing.

    A pathologist reviewing tissue samples for diagnosis can simultaneously label those samples for the AI model — the diagnostic decision IS the label. A lawyer reviewing contracts for a client can annotate clause types as they read — the review process generates labels as a byproduct.

    This "labeling as byproduct" approach produces the highest adoption rates because it doesn't feel like additional work. The expert is doing their job; the labeling interface captures their professional judgment as they exercise it.

    Where integration isn't possible, the next best option is adjacency — labeling happens immediately before or after a related task, using the same documents the expert is already working with.

    4. Time-Box Sessions

    Twenty minutes per day is the sweet spot. It's short enough to fit into gaps between meetings, rounds, or site visits. It's long enough to produce 15-30 labels per session. And it's psychologically manageable — a commitment of "20 minutes" feels minor, while "label 500 examples" feels overwhelming.

    Set a timer in the interface. When 20 minutes elapse, show the session summary and close the labeling queue. This prevents burnout and sets the expectation that limited time commitment is not just acceptable — it's the design.

    Some experts will want to continue beyond 20 minutes. Let them. But don't expect it.

    5. Show Impact

    After the first week of labeling, show the expert what their contributions produced. A dashboard showing:

    • Examples labeled this week: 87
    • Model accuracy before their labels: 67%
    • Model accuracy after incorporating their labels: 74%
    • Estimated accuracy after target examples reached: 91%

    This creates a feedback loop that sustains motivation. The expert sees their professional knowledge directly improving a system. This is meaningful in a way that "thanks for your contribution" emails are not.

    Update the dashboard weekly. Stale dashboards signal that nobody is paying attention to their effort.

    6. Address Data Concerns Proactively

    Don't wait for experts to raise data handling questions. Address them in the first conversation:

    • "Your labels are stored on [specific server name] within the hospital network. Nothing leaves the building."
    • "Access is limited to [names of 3-4 people]. Here's the access control list."
    • "The model trained on your labels will be used for [specific purpose]. It will not be sold, shared, or repurposed."
    • "You can request deletion of your contributed labels at any time."

    Put this in writing. For regulated industries, have the institutional review board or compliance committee sign off.

    Success Metrics

    Track these to measure whether your adoption program is working:

    Participation rate. Percentage of invited domain experts who label at least once per week. Target: 70%+ after the first month. Below 50% indicates friction or motivation issues.

    Session duration. Average time per labeling session. Target: 15-25 minutes. Significantly shorter suggests the tool is frustrating; significantly longer suggests potential burnout risk.

    Label quality over time. Inter-annotator agreement between domain experts. Should be stable or improving week over week. Declining quality suggests fatigue or unclear guidelines.

    Voluntary return rate. Percentage of experts who return to label without being reminded. This is the strongest signal of genuine adoption versus compliance.

    Time to first label. How long from opening the application to submitting the first label. Target: under 90 seconds. This measures friction, not expert speed.

    What Organizations Get Wrong

    Treating it as a technology deployment. Installing a labeling tool and sending a login link is not adoption. Adoption requires the same change management that any organizational initiative requires: executive sponsorship, clear communication, training, and ongoing support.

    Asking for too much, too soon. "We need 5,000 labels by end of month" pressures experts to rush, producing low-quality labels that damage model training. Start with a modest target (200-300 labels), demonstrate value, then increase.

    Forgetting to close the loop. If experts label 500 examples and never hear what happened, they won't label 500 more. Show them the model's improvement. Invite them to test the improved model. Make their contribution tangible.

    Ertas Data Suite is designed for domain experts who are not data scientists. The desktop application runs locally — no portals, no browser tabs, no Python. Documents display in a familiar reading interface with labeling controls that require no training. Sessions are time-boxed with automatic save, so interrupted work is never lost. All data stays on the organization's infrastructure, with role-based access controls that compliance teams can audit directly.


    Your data is the bottleneck — not your models.

    Ertas Data Suite turns unstructured enterprise files into AI-ready datasets — on-premise, air-gapped, with full audit trail. One platform replaces 3–7 tools.

    Further Reading

    Turn unstructured data into AI-ready datasets — without it leaving the building.

    On-premise data preparation with full audit trail. No data egress. No fragmented toolchains. EU AI Act Article 30 compliance built in.

    Keep reading