
Why Your AI Project Is Stalling — It's Not the Model
Most failed AI projects blame the model when the real failure was at the data stage. Here's why data preparation is where enterprise AI projects actually stall.
Your AI project is behind schedule. The team has evaluated three foundation models, benchmarked fine-tuning approaches, and set up GPU infrastructure. But six months in, you're still cleaning data. The model hasn't seen a single training example yet.
This isn't unusual. It's the pattern. And the root cause isn't the model, the team, or the timeline — it's that data preparation was treated as a preliminary step rather than the core of the project.
The Pattern
Here's how enterprise AI projects typically unfold:
Month 1-2: Model-first planning. The team evaluates models, compares architectures, sets up training infrastructure. Exciting, visible progress. Leadership gets demos of what the model could do with good training data.
Month 3: Data reality check. The team turns to the training data and discovers: the documents are in 12 different formats. 40% are scanned with poor OCR quality. There's no labeling schema defined. The domain experts who need to label data are booked on other projects. Nobody knows what PII is in the dataset.
Month 4-5: Data firefighting. Custom scripts are written for parsing. A labeling tool is set up. Domain experts squeeze in labeling time between their actual jobs. Quality issues surface — the OCR output is garbled, the labeling categories are ambiguous, the initial dataset is too small. The timeline slips.
Month 6+: Decision point. The project is over budget and behind schedule. Leadership asks whether to continue or shelve it. The model gets blamed. "Maybe we need a different approach." In reality, the data was never ready.
Why This Keeps Happening
Data Prep Is Invisible Work
Model training produces visible outputs: loss curves, benchmark scores, generated text. Data preparation produces... cleaned data. It doesn't demo well. It's hard to show progress on. Leadership can't see the difference between raw data and prepared data in a status update.
This visibility gap means data prep gets under-resourced. Teams know it matters but can't articulate its value in the terms that secure budget and attention.
The 60-80% Statistic Isn't Internalized
Every ML practitioner has heard that 60-80% of ML project time goes to data preparation. But project plans don't reflect this. A six-month AI project with a one-month data prep allocation is planning for failure.
The statistic persists because data preparation is genuinely hard — not because teams are inefficient. Document diversity, quality issues, labeling complexity, compliance requirements, and domain expertise needs all contribute real, irreducible effort.
Domain Experts Are Treated as Optional
The people who know whether a legal clause is "favorable" or a medical note indicates a specific condition are not the people building the AI pipeline. Domain experts are brought in late, given tools they can't use (Python-based annotation environments), and expected to label data as a side task.
The result: proxy labeling by ML engineers who guess at domain-specific categories, or extended timelines while domain experts are gradually onboarded to developer tools.
Tooling Fragmentation
The typical enterprise data preparation setup involves 3-7 disconnected tools: a parser, a cleaner, a labeler, a quality scorer, an export script. Each tool has its own interface, its own data format, and its own learning curve. Integration between tools is custom code that breaks when any tool updates.
This fragmentation multiplies the effort. Every boundary between tools is a place where data gets lost, formats get mangled, and audit trails break.
What Actually Fixes This
1. Budget Data Prep Honestly
If your AI project is six months, budget four months for data preparation. This isn't pessimism — it's realism. The model training, evaluation, and deployment will take 1-2 months if the data is ready.
2. Staff for Data Prep, Not Just Modeling
Data preparation needs different skills than model training. You need people who understand document processing, data quality, annotation workflows, and compliance — not just people who can write PyTorch training loops.
3. Involve Domain Experts from Day One
Don't bring the cardiologist in at month four. Involve domain experts from the start — in defining the labeling schema, in reviewing early data quality, in establishing what "good" training data looks like for the use case.
This means giving them tools they can actually use. Desktop applications with visual interfaces, not Jupyter notebooks and CLI tools.
4. Use Unified Tooling
Replace the 3-7 tool chain with a single platform that handles the full pipeline. Not because one tool is better at each individual stage — but because the integration cost of maintaining multiple tools exceeds the benefit of best-in-class at each stage.
5. Make Data Prep Visible
Report on data preparation progress the same way you report on model performance. Number of documents ingested, cleaning completion percentage, labeling progress, quality scores. Make the work visible to leadership so it gets the resources it needs.
The Uncomfortable Truth
There's no shortcut around data preparation. No model — no matter how large, how well-architectured, or how expensive — can compensate for poor training data. GPT-4 and Claude didn't become capable by running clever algorithms on mediocre data. They became capable because the training data was enormous, carefully curated, and rigorously quality-controlled.
Enterprise AI operates on the same principle, just at smaller scale. The quality of your AI output is bounded by the quality of your training data. Everything else — model selection, hyperparameter tuning, infrastructure optimization — is secondary.
If your AI project is stalling, look at the data first. That's almost certainly where the problem is.
Platforms like Ertas Data Suite exist because this problem is structural — fragmented tools, inaccessible interfaces, and missing audit trails create compounding delays. A unified, on-premise platform that handles the full pipeline and puts domain experts in control of labeling addresses the root cause, not the symptoms.
Turn unstructured data into AI-ready datasets — without it leaving the building.
On-premise data preparation with full audit trail. No data egress. No fragmented toolchains. EU AI Act Article 30 compliance built in.
Keep reading

The Data Preparation Gap: Why ML Teams Spend 80% of Time Before Training Starts
Why the 60-80% data preparation statistic persists — fragmented tooling, domain expert exclusion, missing audit trails, and the structural underinvestment in purpose-built data prep tools.

EU AI Act Article 10 vs. Article 30: What Your Data Team Needs to Know
A detailed comparison of EU AI Act Articles 10 and 30 — the two most critical provisions for AI training data governance, documentation, and compliance.

EU AI Act Compliance Timeline: What's Due by August 2026
A clear timeline of EU AI Act enforcement dates, what's already in effect, what's coming in August 2026, and what enterprises need to have in place for training data compliance.