Back to blog
    FedRAMP, ITAR, and Air-Gapped AI: Data Prep Without Cloud Exposure
    fedrampitarair-gappedgovernmentdefensecompliancedata-preparationsegment:enterprise

    FedRAMP, ITAR, and Air-Gapped AI: Data Prep Without Cloud Exposure

    How FedRAMP authorization, ITAR restrictions, and air-gapped requirements shape AI data preparation for government and defense — and why native desktop apps solve the compliance problem.

    EErtas Team·

    Government and defense organizations face a unique intersection of compliance requirements that eliminates most AI data preparation tools from consideration. FedRAMP governs cloud service usage, ITAR restricts technical data handling, and classified environments require air-gapped operation. Together, these frameworks create a narrow path for AI adoption — one that runs through on-premise, offline-capable tooling.

    FedRAMP: When Cloud Is Theoretically Possible

    The Federal Risk and Authorization Management Program (FedRAMP) provides a standardized approach to security assessment for cloud services used by federal agencies. There are three authorization levels:

    • FedRAMP Low: For systems with low-impact data (public websites, non-sensitive information)
    • FedRAMP Moderate: For systems handling CUI and most government operations
    • FedRAMP High: For systems supporting the most sensitive unclassified government workloads

    The Reality for AI Data Preparation

    Even with FedRAMP authorization, cloud-based data preparation faces practical barriers:

    Authorization timeline: Obtaining FedRAMP authorization typically takes 12-18 months and costs $1-3 million. Most AI data preparation vendors haven't pursued it because the government market alone doesn't justify the investment.

    Data classification gaps: FedRAMP covers unclassified systems. Any data preparation involving classified information is out of scope — and classified data is often where the most valuable AI training data lives.

    Continuous monitoring burden: FedRAMP authorized systems require ongoing monitoring, annual assessments, and incident reporting. For a data preparation tool that an agency might use for a single project, the overhead may not justify the approach.

    The practical result: Most government AI teams skip the FedRAMP question entirely and process data on-premise, on accredited systems they already control.

    ITAR: When Data Can't Cross Borders

    The International Traffic in Arms Regulations restrict the export of defense articles and technical data. For AI data preparation, ITAR creates specific constraints:

    What's ITAR-Controlled

    • Technical data directly related to defense articles (weapons systems, military vehicles, satellites)
    • Software specifically designed for defense applications
    • Technical data derived from defense-related research

    What ITAR Means for Data Preparation

    • No foreign national access: Data preparation tools processing ITAR data cannot be accessible to non-US persons. This includes cloud services with data centers staffed by non-US personnel.
    • No foreign servers: ITAR data cannot be stored or processed on servers outside the United States — even US-owned servers in foreign data centers.
    • Vendor employee restrictions: If the data preparation vendor has non-US employees who might access the system (for support, debugging, updates), the tool may be disqualified.
    • Derivative data inherits controls: AI training data derived from ITAR-controlled documents is itself ITAR-controlled.

    The Practical Impact

    ITAR effectively requires that AI data preparation for defense technical data happens on systems physically located in the US, operated by US persons, with no cloud connectivity that could expose data to foreign access.

    A native desktop application installed on an accredited workstation satisfies ITAR requirements by design. There's no network exposure, no cloud backend, no foreign server risk.

    Air-Gapped Environments: When Offline Is Mandatory

    Classified networks (SIPRNet for Secret, JWICS for Top Secret/SCI) are physically air-gapped from the internet. There is no network path between these systems and the public internet.

    What "True Air-Gapped" Means for Tooling

    Many tools claim "offline capability" but actually require:

    • License verification: Periodic phone-home to a license server (fails air-gapped)
    • Telemetry: Usage analytics sent to the vendor (fails air-gapped)
    • Auto-updates: Automatic update checks (fails air-gapped)
    • Cloud AI features: "AI-assisted" features that call cloud APIs (fails air-gapped)
    • Container registries: Docker-based tools that pull images from registries (fails air-gapped)

    True air-gapped operation means the tool works identically whether or not a network connection exists. Every dependency, every model, every feature must be bundled in the installation package.

    Docker vs. Native Desktop in Air-Gapped Environments

    Docker-based tools (Label Studio, many open-source ML tools) face specific challenges in air-gapped environments:

    • Image transport: Container images must be exported, transferred via physical media (DVD, secure USB), and imported — a manual process for every update
    • Dependency chains: Docker images pull from multiple registries; ensuring all layers are included in the air-gapped package is error-prone
    • Networking complexity: Docker's networking model adds configuration complexity and security surface area
    • Container orchestration: K8s-based tools require additional infrastructure that may not exist on classified networks

    Native desktop applications (installed via a single binary or installer package) avoid all of these issues. The application installs like any other approved software — no containers, no orchestration, no networking configuration.

    The Convergent Solution

    FedRAMP complexity, ITAR restrictions, and air-gapped requirements all point to the same architecture: a native desktop application that runs entirely on local infrastructure with no cloud dependencies.

    This is why Ertas Data Suite is built as a Tauri 2.0 native desktop application:

    • No cloud connectivity required: Installs and runs without internet access
    • No Docker/K8s: Single application binary, no container infrastructure
    • Local AI inference: AI-assisted features use Ollama/llama.cpp, running models locally
    • Complete audit trail: Every operation logged locally with operator attribution
    • US-person controllable: No foreign server exposure, no vendor telemetry

    For government and defense organizations, the question isn't whether to process AI training data on-premise — the compliance frameworks have already answered that. The question is which on-premise tool best fits the security requirements while still being usable by the analysts who understand the data.

    The intersection of FedRAMP, ITAR, and air-gapped requirements narrows the field considerably. Purpose-built native desktop applications that handle the full data preparation pipeline — from document ingestion through labeling and export — are the architecture that satisfies all three.

    Turn unstructured data into AI-ready datasets — without it leaving the building.

    On-premise data preparation with full audit trail. No data egress. No fragmented toolchains. EU AI Act Article 30 compliance built in.

    Keep reading