
Private LLM - Local AI Chat 12+
Numen Technologies Limited
-
- 5,99 €
Screenshot
Descrizione
Meet Private LLM: Your Secure, Offline AI Assistant for macOS
Private LLM brings advanced AI capabilities directly to your iPhone, iPad, and Mac—all while keeping your data private and offline. With a one-time purchase and no subscriptions, you get a personal AI assistant that works entirely on your device.
Key Features:
- Local AI Functionality: Interact with a sophisticated AI chatbot without needing an internet connection. Your conversations stay on your device, ensuring complete privacy.
- Wide Range of AI Models: Choose from various open-source LLM models like Llama 3.2, Llama 3.1, Google Gemma 2, Microsoft Phi-3, Mistral 7B, and StableLM 3B. Each model is optimized for iOS and macOS hardware using advanced OmniQuant quantization, which offers superior performance compared to traditional RTN quantization methods.
- Siri and Shortcuts Integration: Create AI-driven workflows without writing code. Use Siri commands and Apple Shortcuts to enhance productivity in tasks like text parsing and generation.
- No Subscriptions or Logins: Enjoy full access with a single purchase. No need for subscriptions, accounts, or API keys. Plus, with Family Sharing, up to six family members can use the app.
- AI Language Services on macOS: Utilize AI-powered tools for grammar correction, summarization, and more across various macOS applications in multiple languages.
- Superior Performance with OmniQuant: Benefit from the advanced OmniQuant quantization process, which preserves the model's weight distribution for faster and more accurate responses, outperforming apps that use standard quantization techniques.
Supported Model Families:
- DeepSeek R1 Distill based models
- Phi-4 14B model
- Llama 3.3 70B based models
- Llama 3.2 based models
- Llama 3.1 based models
- Llama 3.0 based models
- Google Gemma 2 based models
- Qwen 2.5 based models (0.5B to 32B)
- Qwen 2.5 Coder based models (0.5B to 32B)
- Google Gemma 3 1B based models
- Solar 10.7B based models
- Yi 34B based models
For a full list of supported models, including detailed specifications, please visit privatellm.app/models.
Private LLM is a better alternative to generic llama.cpp and MLX wrappers apps like Enchanted, Ollama, LLM Farm, LM Studio, RecurseChat, etc on three fronts:
1. Private LLM uses a significantly faster mlc-llm based inference engine.
2. All models in Private LLM are quantised using the state of the art OmniQuant quantization algorithm, while competing apps use naive round-to-nearest quantization.
3. Private LLM is a fully native app built using C++, Metal and Swift, while many of the competing apps are bloated and non-native Electron JS based apps.
Please note that Private LLM only supports inference with text based LLMs.
Private LLM has been specifically optimized for Apple Silicon Macs.Private LLM for macOS delivers the best performance on Macs equipped with the Apple M1 or newer chips. Users on older Intel Macs without eGPUs may experience reduced performance. Please note that although the app nominally works on Intel Macs, we've stopped adding support for new models on Intel Macs due to performance issues associated with Intel hardware.
Novità
Versione 1.9.10
- Support for the new Qwen3 4B Instruct 2507 model (on Apple Silicon Macs with 16GB or more RAM)
- Minor bug fixes and updates
Valutazioni e recensioni
Missing Shortcuts repository and extensions
I purchased also for the idea of having a repository for shortcuts and extensions to lunch the Private LLM directly from Finder or inside a document. It's incomplete without that functionality. It would be nice to have the following "extensions" directly in Finder:
- analyze the document
- summerize document
- chat with the document
and inside the file
- translate
- summerize the text
- rephase
- elaborate
Last but not least it would be nice to add the DocsGPT model and the possibility of dragging an drop one or more document directly in the Private LLM app.
Risposta dello sviluppatore ,
Thanks for the review! We don't ship any shortcuts with the app, but you can download popular user contributed shortcuts from our website (privatellm.app). Also, we have macOS services for bulletizing text, grammar correction, rephrasing and shortening text. You can right click on editable text within any app and access them from the services sub-menu. Additionally, users can also assign keyboard shortcuts to these services from System Settings -> Keyboard -> Keyboard Shortcuts -> Services -> Text. Also, users can add their own services for translation, or anything else they'd like to wish for by creating a macOS Shortcut that uses Private LLM and checking the checkbox within the Shortcuts app to make it accessible from the Services menu. Finally, We're working on implementing Retrieval augmented generation (RAG). As you know, given our commitment to privacy, we need to implement completely on-device and there are some technical challenges that need to be overcome before we can ship that feature. If you have any feature requests, please feel free to email us (support@numen.ie) or DM us on Twitter (@private_llm) or join our Discord.
Unfortunately, useless. No RAG, no unpublished models. Good only for newbies.
I use LMStudio, anythingLLM, Ollama and open web ui on daily basis. I have a M3MAX/64GB and unfortunately this software doesn’t have anything of useful: 1) if you need more control on models and parameters, use LMStudio; 2) if you need something to analyze your own documents use Dot (it’s free); 3) if you need something to transcribe use Whisper; 4) if you need something to Upscale images use Upscaly; 5) If you need to create images use DiffusionBee. Unfortunately this chatbot is completely useless since you have to stick to their models and nothing else. And the models are freely available on Hugginface via LMStudio or Ollama. So it is 6 bucks wasted. Sorry devs, but someone has to say this. This app is useful only if you don’t want to read anything online to make it completely for free.
Risposta dello sviluppatore ,
Dear shaice, it looks like you aren't following your own advice to do some reading online. Please go to Youtube, Google, or any other search engine, and search for "LM Studio vs Private LLM", "Ollama vs Private LLM", etc. While you're at it, you might also want to go read the FAQ on our website (privatellm.app/faq). What you'll learn, might change your mind. If not, go to reportaproblem.apple.com and request a refund. One more thing: two of the three VC funded llama.cpp wrappers that you're shilling are bloated electron.js apps (LMStudio, anythingLLM), the third does not have a UI (Ollama). Private LLM is a fully native SwiftUI app, with deep macOS and iOS integrations. Also, none of the apps you're shilling run on iOS and iPadOS, Private LLM does.
Dare la possibilità di caricare un modello personale
Your application is paid, but it should allow users to upload their own personal model via file. I use Llama3, 4Q, which works everywhere, while your model does not function correctly as it responds without understanding. You have only provided Llama3, 3Q. We request that you add a feature to select and upload a GGUF file that we have already downloaded.
Risposta dello sviluppatore ,
One of the reasons why our app is far better than other GGUF based apps is because we quantize models with the state of the art OmniQuant algorithm. It takes many hours of NVIDIA GPU time to quantize each model with OmniQuant, while it only takes a few minutes of CPU time to RTN quantize GGUF models. But the resulting OmniQuant quantized models are superior to RTN quantized models, in terms of perplexity. You'll find more details about this on our website (privatellm.app). If you have any requests for models, please write to us and we'll quantize and include the option to download it in an upcoming release.
Privacy dell’app
Lo sviluppatore, Numen Technologies Limited, ha indicato che le procedure per la tutela della privacy dell’app potrebbero includere il trattamento dei dati descritto di seguito. Per scoprire di più, consulta l’informativa sulla privacy dello sviluppatore.
Dati non raccolti
Lo sviluppatore non raccoglie alcun dato da quest’app.
Le procedure per la tutela della privacy possono variare, per esempio, in base alle funzioni che usi o alla tua età. Scopri di più
Informazioni
- Fornitore
-
Numen Technologies Limited si identifica come operatore commerciale per questa app e ha confermato che il prodotto o il servizio è conforme alla legge dell’Unione europea.
- Numero DUNS
- 985782161
- Indirizzo
- WORK HUB, 77 CAMDEN STREET LOWER, DUBLIN N/A D02XE80 Irlanda
- Numero di telefono
- +353 831846817
- support@numen.ie
- Dimensione
- 1,3 GB
- Categoria
- Utility
- Compatibilità
-
- iPhone
- Richiede iOS 17.0 o versioni successive e un dispositivo con chip A12 Bionic o versioni successive.
- iPad
- Richiede iPadOS 17.0 o versioni successive e un dispositivo con chip A12 Bionic o versioni successive.
- Mac
- Richiede macOS 14.0 o versioni successive.
- Apple Vision
- Richiede visionOS 1.0 o versioni successive e un dispositivo con chip A12 Bionic o versioni successive.
- Lingue
-
Inglese
- Età
- 12+ Raro/moderato contenuto adatto a un pubblico adulto Raro/moderato contenuto horror/paura Raro/moderato contenuto medico/terapeutico
- Copyright
- © 2024 Numen Technologies Limited
- Prezzo
- 5,99 €
Supporto
-
In famiglia
Se “In famiglia” è stato attivato, fino a sei membri della famiglia potranno usare l’app.