Back to blog
    Your User's Data Leaves Their Phone on Every AI Request
    privacydata securitymobile AIGDPRHIPAAcompliancesegment:mobile-builder

    Your User's Data Leaves Their Phone on Every AI Request

    Every cloud AI API call sends user data to a third-party server. What that means for privacy, compliance, user trust, and your app's long-term viability.

    EErtas Team·

    When your app calls an AI API, the user's input is sent to a server operated by a third party. Their message, their question, their text, their context. It leaves the device, crosses the internet, and lands on infrastructure you do not control.

    For a weather app, this is probably fine. For a health app, a finance app, a note-taking app, or anything that touches personal information, it is a serious problem.

    What Actually Gets Sent

    A typical chat API request includes:

    {
      "model": "gpt-4o-mini",
      "messages": [
        {"role": "system", "content": "You are a health assistant..."},
        {"role": "user", "content": "I've been having chest pain for 3 days..."},
        {"role": "assistant", "content": "...previous AI response..."},
        {"role": "user", "content": "Should I see a cardiologist?"}
      ]
    }
    

    That payload contains:

    • The user's health concern (PHI under HIPAA)
    • The conversation history (accumulated personal context)
    • The system prompt (your app's proprietary logic)

    Every turn of conversation sends the full history. By turn 5, the API has received the user's complete interaction, not just the latest message.

    RAG Context Makes It Worse

    If your app uses retrieval-augmented generation, the request also includes retrieved documents. In a finance app, that might be the user's transaction history. In a legal app, confidential case details. In a note-taking app, their private notes.

    The more context you provide for quality responses, the more private data you transmit.

    The Privacy Policies You Inherit

    When you send user data to an AI provider, their data handling policies apply:

    OpenAI: As of 2026, API data is not used for model training by default (you must opt in). But data is retained for 30 days for abuse monitoring. OpenAI staff may review flagged content.

    Anthropic: Similar policy. API data is not used for training. Retained for safety monitoring with possible human review of flagged content.

    Google Gemini: For the paid API (not the free consumer product), data is not used for training. Retention policies vary by agreement.

    The common thread: your users' data sits on third-party servers for days to weeks, potentially reviewed by humans you do not know, under policies that can change with an updated Terms of Service.

    The Compliance Problem

    HIPAA (Healthcare)

    If your app handles protected health information (PHI), sending it to an AI provider requires a Business Associate Agreement (BAA). OpenAI and Anthropic offer BAAs for enterprise tiers, but:

    • BAAs are only available on enterprise plans (not standard API access)
    • The BAA covers data handling, not the fact that data transits public internet
    • Auditors increasingly flag cloud AI as a compliance risk even with BAAs

    GDPR (European Users)

    GDPR requires a lawful basis for processing personal data and for transferring it outside the EU. Sending European users' data to US-based AI providers raises transfer questions under Schrems II.

    Even with Standard Contractual Clauses (SCCs), the DPA (Data Protection Authority) may challenge whether the transfer is adequately protected. The risk is real: fines under GDPR can reach 4% of global annual revenue.

    SOC 2 / ISO 27001

    If your company holds SOC 2 or ISO 27001 certification, third-party data processing must be documented and audited. Every AI API provider becomes a vendor in your supply chain that auditors will scrutinize.

    App Store Privacy Labels

    Apple requires privacy nutrition labels. If your app sends user data to an AI API, you must disclose:

    • Data linked to the user
    • Data used for tracking (if any analytics are attached)
    • Third-party data sharing

    Users see this before downloading. Apps with extensive data sharing disclosures see lower install rates compared to privacy-friendly alternatives.

    User Trust

    Privacy concerns are growing among mobile users:

    • 87% of consumers say they would not do business with a company if they had concerns about its data practices (Cisco Consumer Privacy Survey, 2025)
    • 76% of smartphone users have declined an app permission due to privacy concerns (Pew Research)
    • App uninstall rates increase when users discover unexpected data sharing through news coverage or privacy reports

    For apps in sensitive categories (health, finance, personal productivity), privacy is a competitive differentiator. "Your data never leaves your device" is a marketing message that resonates.

    The On-Device Alternative

    On-device AI eliminates the privacy problem structurally. The model runs on the user's phone. The user's input is processed locally. No data is transmitted.

    FactorCloud APIOn-Device
    User data sent to third partyEvery requestNever
    HIPAA complianceRequires BAA + enterprise tierCompliant by architecture
    GDPR transfer concernsYes (US-based providers)None (data stays on device)
    App Store privacy label"Data shared with third parties""Data not collected"
    Vendor data retention30 days (typical)None
    Human review riskPossible (flagged content)None
    Policy change riskTerms can change unilaterallyN/A

    The Technical Implementation

    On-device inference means the AI model runs directly on the user's hardware. The inference loop is:

    1. User types input
    2. Input is tokenized on-device
    3. Model generates response on-device
    4. Response is displayed

    No network call. No external server. No data in transit. The user's data exists only in their app's local storage and the device's RAM during inference.

    The Business Case

    Beyond regulatory compliance, on-device AI has a concrete business impact:

    Lower friction in sensitive markets: Health, finance, and legal apps that can truthfully state "your data never leaves your device" face fewer user objections and lower churn.

    Simpler compliance posture: No BAAs to negotiate, no data processing agreements, no GDPR transfer assessments. The compliance surface area shrinks dramatically.

    Reduced liability: If user data is never transmitted, there is no data in transit to intercept, no third-party breach that exposes your users, and no vendor policy change that creates a retroactive compliance issue.

    App Store advantage: A clean privacy label on the App Store and Play Store builds trust before the user even opens the app.

    Making the Transition

    1. Audit your current data flow. What user data is included in each API call? Map every field.
    2. Assess regulatory exposure. Which regulations apply to your user base and data types?
    3. Collect training data. Your existing API interactions (with appropriate consent) are your training dataset.
    4. Fine-tune a domain model. Use a platform like Ertas to fine-tune a small model (1-3B parameters) on your specific task.
    5. Deploy on-device. Export as GGUF, integrate llama.cpp, run locally.
    6. Update your privacy policy. The best part: you can remove third-party data sharing disclosures.

    The privacy improvement is not just a compliance checkbox. It is a user-facing feature that builds trust and reduces churn.

    Ship AI that runs on your users' devices.

    Early bird pricing starts at $14.50/mo — locked in for life. Plans for builders and agencies.

    Ship AI that runs on your users' devices.

    Early bird pricing starts at $14.50/mo — locked in for life. Plans for builders and agencies.

    Keep reading