
Ertas vs Modal Labs: Which Is Better for Agencies Fine-Tuning Client Models?
Comparing Ertas and Modal Labs for AI agency fine-tuning workflows. Covers the GUI vs code divide, multi-client management, cost predictability, and GGUF deployment.
Modal Labs is what you use when you want to write Python decorators and get auto-scaling GPUs. Ertas is what you use when you want to upload a CSV and click Train. They address the same underlying need — GPU compute for fine-tuning — but for completely different users.
For AI agencies specifically, the choice matters beyond personal preference. The platform you choose affects your team's ability to onboard non-ML members, manage multiple client models, predict monthly costs, and deliver GGUF files that clients can deploy locally.
What Modal Labs Is
Modal is serverless GPU infrastructure. The development experience is Python-first: you write a training script, add Modal decorators, and Modal handles spinning up GPU instances, installing dependencies, managing storage, and auto-scaling.
A typical Modal fine-tuning setup looks like this:
import modal
app = modal.App("fine-tune-llm")
image = modal.Image.debian_slim().pip_install("unsloth", "torch", "transformers")
@app.function(gpu="A100", timeout=3600, image=image)
def fine_tune(training_data_path: str):
from unsloth import FastLanguageModel
# your training code here
model, tokenizer = FastLanguageModel.from_pretrained(
model_name="unsloth/llama-3-8b-instruct",
max_seq_length=2048,
)
# ... training loop ...
model.save_pretrained("./output")
@app.local_entrypoint()
def main():
fine_tune.remote("s3://your-bucket/train.jsonl")
This is genuinely powerful. You get A100 access in minutes, your code runs exactly as written, and you pay only for what you use. If you are an ML engineer who wants total control over the training loop, Modal is excellent.
If you are not an ML engineer — or if you need your non-technical account manager to run training jobs for client X without understanding PyTorch — Modal does not work.
What Ertas Is
Ertas is a visual fine-tuning platform. The workflow is: upload JSONL dataset → configure training in the UI → train on cloud GPUs → export GGUF → run locally. No Python, no YAML, no understanding of gradient accumulation required.
For agencies specifically, Ertas adds client-labeled projects (each client's datasets and models are isolated), team seats (up to 15 on Agency Pro), concurrent job slots (up to 8), and per-client LoRA adapter management.
Comparison Table
| Dimension | Ertas | Modal Labs |
|---|---|---|
| Interface | Visual web UI | Python code |
| Setup time | ~2 min | ~60 min (environment + code) |
| ML expertise required | None | Significant |
| Team access | Role-based seats (5 on Agency) | API key sharing / Python code access |
| Client project management | Built-in, labeled per client | Build it yourself |
| GGUF export | One-click | Manual (llama.cpp convert.py) |
| Local deployment support | Yes (Ollama/LM Studio) | No (you handle it) |
| Cost model | Monthly subscription, fixed | Per GPU-second, variable |
| Cost predictability | High | Low (depends on job length/GPU type) |
| Concurrent jobs | Up to 8 (Agency Pro) | Unlimited (expensive) |
| Custom training loops | No | Yes (full Python) |
| Experiment tracking | Visual canvas, side-by-side | None built-in (add Weights & Biases) |
| Dataset tools | Validation, synthesis, eval | Bring your own |
| Who it's for | Agencies, non-ML builders | ML engineers who want GPU compute |
The Agency Perspective
Agencies have five requirements that differ from individual ML engineers:
1. Non-ML team members need to participate. Your account manager needs to upload client data and run a retrain. Your project manager needs to see training results. Neither should need to understand Python decorators or Hugging Face data collators. Ertas works for them; Modal does not.
2. Client data must be isolated. When you are fine-tuning models for 8 different clients, commingling their data in a shared Python environment is a serious risk. Ertas has client-labeled projects with explicit isolation. Modal has whatever you build.
3. Costs need to be predictable. Agency billing runs on fixed retainers. If your underlying cost for a month of model maintenance ranges from $40 to $400 depending on how many jobs ran, margin planning becomes impossible. Ertas Agency at $69.50/month is fixed. Modal's per-second billing is not.
4. GGUF output is usually required. Most agency clients want to run the model on their own infrastructure, not pay API costs indefinitely. One-click GGUF export in Ertas vs manual conversion from Modal checkpoints is a significant operational difference.
5. Multiple concurrent jobs. When you have 10 clients with staggered retraining schedules, you need to run jobs simultaneously. Ertas Agency supports 3 concurrent jobs (Agency Pro: 8). Modal supports unlimited concurrency — but you build the queue management yourself.
The Cost Math for an Agency
Scenario: a small AI agency managing 8 active clients, each requiring 1-2 training runs per month plus evaluation.
Modal Labs cost estimate:
- 8 clients × 1.5 training runs × ~90 minutes on A100 = ~18 GPU-hours/month
- A100 on Modal: ~$3/hour
- Total compute: ~$54/month
- Plus: developer time to write and maintain training scripts (ongoing)
- Plus: building project management, evaluation tooling, GGUF conversion scripts
Ertas Agency cost:
- $69.50/month (Early Bird)
- Includes 400 credits (comfortably covers 8 clients × 1.5 runs)
- Project management, evaluation, GGUF export included
- Non-ML team members can operate it
On raw compute cost, Modal might be cheaper. But the total cost of ownership — including developer time to build and maintain the training infrastructure — shifts the comparison significantly.
| Cost Category | Modal Labs | Ertas Agency |
|---|---|---|
| Monthly platform/compute | ~$40-80 | $69.50 (fixed) |
| Dev hours/month to maintain | 3-5 hrs @ $75/hr | 0 |
| Project management layer | Build yourself | Included |
| GGUF conversion tooling | Build yourself | Included |
| Team seat access | API key only | 5 seats |
| Total effective monthly cost | ~$265-455 | $69.50 |
Same Fine-Tuning Job: Side-by-Side
Task: Fine-tune Qwen 2.5 7B on 600 customer support examples for a client.
Modal Labs workflow:
- Write training script using Unsloth + Modal decorators (~4 hours for first client, 1 hour for each subsequent client with reusable code)
- Upload data to S3 or Modal volume
- Run
modal run train.py --data-path s3://... - Wait 45-90 minutes
- Download checkpoint from Modal volume
- Run llama.cpp convert.py to export GGUF (~30 min including setup)
- Test GGUF with Ollama
- Total: 45-90 min training + 2-3 hours setup (first time) + 1 hour GGUF conversion
Ertas workflow:
- Upload JSONL in browser
- Select Qwen 2.5 7B from model dropdown
- Configure training settings visually
- Click Train
- Wait 45-90 minutes (same underlying GPU time)
- Review evaluation metrics in UI
- Click Export GGUF
- Download and load in Ollama
- Total: 45-90 min training + 15 minutes active work
When Modal Labs Wins
- Your team consists of ML engineers who want to write custom training loops
- You need exotic architectures (custom attention patterns, multi-modal, etc.)
- You want to integrate directly with Weights & Biases or other ML tooling
- You have existing PyTorch code you want to scale up
- You are doing research that requires modifying the training loop
When Ertas Wins
- You are running an agency with non-ML team members
- You need client data isolation and project management
- You need GGUF output for local client deployment
- You want predictable monthly costs for agency billing
- You need 3+ concurrent jobs without managing a queue yourself
- You want to onboard a new client in 30 minutes, not 3 hours
Ship AI that runs on your users' devices.
Ertas early bird pricing starts at $14.50/mo — locked in for life. Plans for builders and agencies.
Further Reading
- Best AI Fine-Tuning Platforms in 2026 — Full multi-platform comparison
- White-Label AI Platform for Agencies — Delivering branded AI to agency clients
- Agency AI Cost Reduction — The economics of fine-tuned models for agencies
- Manage Multiple Fine-Tuned Models — Operational workflows for multi-client model management
- Ertas vs Unsloth vs Axolotl 2026 — How Ertas compares to DIY fine-tuning tools
Ship AI that runs on your users' devices.
Early bird pricing starts at $14.50/mo — locked in for life. Plans for builders and agencies.
Keep reading

Ertas vs Replicate for Fine-Tuning: Cost, Workflow, and GGUF Export Compared
Side-by-side comparison of Ertas and Replicate for fine-tuning language models. Covers workflow, pricing, GGUF export, data privacy, and when to choose each platform.

Ertas vs HuggingFace AutoTrain: Visual Fine-Tuning Without the YAML Configs
Comparing Ertas and HuggingFace AutoTrain for no-code LLM fine-tuning. Covers workflow UX, GGUF export, local deployment, pricing, and dataset format differences.

Ertas vs Together AI: Fine-Tuning Costs, Local Deployment, and Data Privacy
Comparing Ertas and Together AI for fine-tuning language models. Covers per-token vs flat-cost inference, data privacy, local deployment, and when each platform wins.