It supports GGUF & MLX models or connect to OpenAI-compatible servers. Free Plan includes most features like local models, iCloud Sync, Markdown conversion and 25+ tools.
Privacy AI introduces **MCP Marketplace**, a one-tap hub to browse and install ready-to-use MCP servers. Powered by [MCPRouter](https://mcprouter.co/), each server is auto-configured. No need to manual setup. So you can instantly add integrations for automation, image generation, data analysis, and more.
Now with **OpenAI Responses API** support: run **stateless multi-turn conversations** without resending full history. Use **conversation objects**, **prompt templates**, and **native prompt caching** for faster, cheaper repeats. The built-in Protocol Inspector shows every request, response, and SSE event in real time for easy debugging.
**Interleaved Thinking** is supported for models like **Minimax M2**, **Claude Sonnet 4**, and **OpenAI GPT-5 (Thinking)**. When enabled, the model preserves reasoning across turns, enabling planning, verification, and self-correction—turning simple Q&A into agentic, multi-step workflows.
**Memory** lets your AI remember across models. Add, edit, and sync personal context that stays consistent for local or cloud models. Your memory belongs to you—not a provider. Memory is free for local models and included in Pro for cloud models.
A **Free Plan** covers offline GGUF/MLX models, iCloud sync, Markdown conversion, and built-in tools. Cloud models, MCP Marketplace, and custom API providers require a subscription.
Privacy AI is a full-featured AI client that works your way: run models fully offline, connect to self-hosted servers, or use cloud providers—all in one app.
### Why Privacy AI
Most AI apps lock you in. Privacy AI gives you freedom: local execution for privacy, self-hosted for compliance, or cloud APIs for performance. You choose where models run, how memories persist, and how data is handled.
### Who It’s For
• Developers and infra builders (GGUF, MLX, OpenRouter, custom APIs)
• Privacy-conscious professionals in law, healthcare, finance, research
• Power users exploring automation and model performance
• Teams needing compliance or internal-only workflows
### What It Does
**Run Any Model, Anywhere**
• Offline on iPhone, iPad, or Mac
• Load GGUF/MLX from HuggingFace or iCloud
• Connect to OpenAI, Anthropic, Gemini, DeepSeek, or your own servers
• Switch between local and remote mid-chat without losing context
• Supports Apple’s on-device Foundation Model (iOS 26)
**Use Tools Seamlessly**
• Full MCP with integrated Marketplace
• Call tools like search, stocks, arXiv, HealthKit, JavaScript—even offline
• Inspect and replay tool calls in chat
**Automate Privately**
• Siri, Shortcuts, and Share Extension
• Offline TTS + transcription
• OpenAI protocol TTS (gpt-4o-tts, gpt-4o-mini-tts)
**Files On-Device**
• Convert PDFs, Office, EPUB, HTML, YouTube, audio
• OCR + Markdown conversion
• Export to Markdown, PDF, HTML, EPUB, JSON
**Images & Editing**
• Generate locally or via OpenRouter (e.g., Gemini 2.5 Flash)
• Support for gpt-image-1 and FLUX.1-dev
• Annotate with Apple Pencil or touch
**Control & Extend**
• “Natural Talk” voice UI
• Live token/context indicator
• Custom prompts, tools, preferences
• iCloud sync for models, chats, settings
**Built for Apple**
• Optimized for M-series Macs and iPad Pro
• Requires iOS 18.6+
• Deep Siri/Shortcuts + Share Extension integration
### No Account. No Surprises.
No forced sign-ups. Choose local, self-hosted, or cloud. Your data and memories stay under your control.
It’s more than a chatbot—it’s a professional-grade AI IDE in your pocket.
Your model. Your device. Your data.
Terms of Use: https://privacyai.acmeup.com/docs/policy/tos.html
Privacy Policy: https://privacyai.acmeup.com/docs/policy/privacy.html
I've tried just about every AI client on the App Store, and this is the one. I stumbled on this app when it had zero downloads and found an absolute gem. This isn't just another chat wrapper; it's a full-on AI agent for your phone. You can plug in APIs for basically any model (Gemini, DeepSeek, etc.) and it has tools that let the AI actually do things—manage your calendar, get directions, search your contacts, the list goes on. The "Super Siri" feature that lets you use your own AI model with a voice command is a complete game-changer.Now, full disclosure, the app is BRAND new, so it's a little rough around the edges. You might hit a few quirks or a random crash here and there as the developer irons things out. But honestly, that's totally expected for something this ambitious. The potential here is just off the charts.Giving this 5 stars for the vision alone and for what it can already do.
Developer Response
Thank you for trying Privacy AI when it only had a few downloads. It means a lot that you see what we’re building, not just another chat app, but a full AI workspace and agent for mobile devices. We’re glad you tried features like broad API support and the “Super Siri” voice command. Both are designed to put you in control of your AI on your own terms. Our philosophy is simple: AI should run under users’ control, free from Big Tech lock-in, and work exactly the way users need it to. The app is still new and not perfect. But we’re shipping fast and improving every week. Thank you for the five stars and for recognizing the vision. Reviews like yours help more people discover Privacy AI and motivate us to make it even better.
Features
- OpenAI Responses API Support
Privacy AI now supports the OpenAI Responses API, OpenAI’s next-generation protocol enabling stateless multi-turn conversations without resending full chat history on each request. The API also introduces conversation objects for automatic context management, prompt templates with variable substitution, and native prompt caching for faster, cheaper repeated queries. It’s fully compatible with OpenAI and any third-party provider implementing the Responses API standard (including LM Studio). The built-in Protocol Inspector displays every request, response, and SSE stream in real time—ideal for debugging and verifying custom endpoints.
- MCP Marketplace (Pro)
A new marketplace powered by [MCPRouter](https://mcprouter.co/) lets you browse and install pre-configured MCP servers in one tap. Each integration—automation, image generation, time zones, data analysis, and more—is auto-configured with endpoints, headers, and authentication ready to go. No manual setup required.
- llama.cpp Engine Update (b6962)
Major CPU optimization for iOS and macOS with ARM64 chunking and Flash Attention improvements. Adds support for vision models like Qwen3-VL, CogVLM, and Janus Pro, offering better multimodal reasoning with reduced memory use on mobile devices.
- Interleaved Thinking Support
New support for interleaved thinking in models such as Minimax M2, Claude Sonnet 4, and OpenAI GPT-5 (Thinking series). When enabled, the model preserves internal reasoning across turns, enabling self-reflection, planning, and retry logic for complex multi-step tasks. This brings truly agentic behavior—turning one-shot answers into autonomous workflows.
- Multi-Image Selection
Select up to 10 photos simultaneously from the picker. All selected images are processed sequentially and attached automatically, streamlining multi-image analysis or comparison.
Improvements
- Simplified MCP Authorization
MCP settings now include a dedicated *Authorization* toggle with a secure API-key field. Privacy AI formats Bearer tokens automatically, eliminating manual header composition and reducing connection errors.
- Improved Protocol Diagnostics
HTTP 404/429 errors now display clearer explanations—showing whether a model is rate-limited or temporarily unavailable—and include actionable recovery guidance.
- File Analysis UX
New real-time progress indicators during OCR or document parsing give clear feedback through every stage of file analysis.
- Collapsible “Similar Models” View
The Similar Models section now collapses by default for cleaner layout and can be expanded instantly to compare prices and alternatives.
Bug Fixes
- macOS Stability
Fixed a crash when opening Model Settings on Mac; the Model Detail view now loads reliably across all devices.
- Local Model Cleanup
Resolved a memory leak caused by failed GGUF or MLX model loads. Failed retries no longer leave stale allocations or cause crashes.
- Super Siri Integration
Restored full Siri compatibility across MLX, GGUF, Apple Intelligence, and all remote providers.
- Dynamic Model Pricing
Model pricing now refreshes instantly whenever you switch models in the Model View.
Version 1.4.5
The developer, AcmeUp Inc., indicated that the app’s privacy practices may include handling of data as described below. For more information, see the developer’s privacy policy .
Data Used to Track You
The following data may be used to track you across apps and websites owned by other companies:
Identifiers
Usage Data
Data Not Linked to You
The following data may be collected but it is not linked to your identity:
Identifiers
Usage Data
Privacy practices may vary, for example, based on the features you use or your age. Learn More
The developer indicated that this app supports the following accessibility features. Learn More
Supported Features
Larger Text
Dark Interface
Differentiate Without Color Alone
Sufficient Contrast
Information
Seller
AcmeUp Inc.
Size
1.8 GB
Category
Productivity
Compatibility
Requires iOS 18.6 or later.
iPhone Requires iOS 18.6 or later.
iPad Requires iPadOS 18.6 or later.
Mac Requires macOS 15.6 or later and a Mac with Apple M1 chip or later.