Private LLM - Local AI Chatbot 12+

Numen Technologies Limited

    • 299,00 ₽

Снимки экрана

Описание

Meet Private LLM: Your Secure, Offline AI Assistant for macOS

Private LLM brings advanced AI capabilities directly to your iPhone, iPad, and Mac—all while keeping your data private and offline. With a one-time purchase and no subscriptions, you get a personal AI assistant that works entirely on your device.

Key Features:

- Local AI Functionality: Interact with a sophisticated AI chatbot without needing an internet connection. Your conversations stay on your device, ensuring complete privacy.

- Wide Range of AI Models: Choose from various open-source LLM models like Llama 3.2, Llama 3.1, Google Gemma 2, Microsoft Phi-3, Mistral 7B, and StableLM 3B. Each model is optimized for iOS and macOS hardware using advanced OmniQuant quantization, which offers superior performance compared to traditional RTN quantization methods.

- Siri and Shortcuts Integration: Create AI-driven workflows without writing code. Use Siri commands and Apple Shortcuts to enhance productivity in tasks like text parsing and generation.

- No Subscriptions or Logins: Enjoy full access with a single purchase. No need for subscriptions, accounts, or API keys. Plus, with Family Sharing, up to six family members can use the app.

- AI Language Services on macOS: Utilize AI-powered tools for grammar correction, summarization, and more across various macOS applications in multiple languages.

- Superior Performance with OmniQuant: Benefit from the advanced OmniQuant quantization process, which preserves the model's weight distribution for faster and more accurate responses, outperforming apps that use standard quantization techniques.

Supported Model Families:

- Llama 3.3 70B
- Llama 3.2 Based Models
- Llama 3.1 Based Models
- Phi-3 Based Models
- Google Gemma 2 Based Models
- Mixtral 8x7B Based Models
- CodeLlama 13B Based Models
- Solar 10.7B Based Models
- Mistral 7B Based Models
- StableLM 3B Based Models
- Yi 6B Based Models
- Yi 34B Based Models
- Qwen 2.5 Based Models (0.5B to 32B)
- Qwen 2.5 Coder Based Models (0.5B to 32B)

For a full list of supported models, including detailed specifications, please visit privatellm.app/models.

Private LLM is a better alternative to generic llama.cpp and MLX wrappers apps like Ollama, LLM Farm, LM Studio, RecurseChat, etc on three fronts:
1. Private LLM uses a faster mlc-llm based inference engine.
2. All models in Private LLM are quantised using the state of the art OmniQuant quantization algorithm, while competing apps use naive round-to-nearest quantization.
3. Private LLM is a fully native app built using C++, Metal and Swift, while many of the competing apps are (bloated) Electron based apps.

Optimized for Apple Silicon Macs with the Apple M1 chip or later, Private LLM for macOS delivers the best performance. Users on older Intel Macs without eGPUs may experience reduced performance.

Что нового

Версия 1.9.5

- Support for downloading 16 new models (varies by device capacity).
- Three new Llama 3.3 based uncensored models: EVA-LLaMA-3.33-70B-v0.0, Llama-3.3-70B-Instruct-abliterated and L3.3-70B-Euryale-v2.3.
- Hermes-3-Llama-3.2-3B and Hermes-3-Llama-3.1-8B models.
- FuseChat-Llama-3.2-1B-Instruct, FuseChat-Llama-3.2-3B-Instruct, FuseChat-Llama-3.1-8B-Instruct, FuseChat-Qwen-2.5-7B-Instruct and FuseChat-Gemma-2-9B-Instruct models.
- FuseChat-Llama-3.2-1B-Instruct also comes with an unquantized variant.
- EVA-D-Qwen2.5-1.5B-v0.0, EVA-Qwen2.5-7B-v0.1, EVA-Qwen2.5-14B-v0.2 and EVA-Qwen2.5-32B-v0.2 models.
- Llama-3.1-8B-Lexi-Uncensored-V2 model
- Improved LaTeX rendering
- Stability improvements and bug fixes.

Thank you for choosing Private LLM. We are committed to continue improving the app and to making it more useful for you. For support requests and feature suggestions, please feel free to email us at support@numen.ie, or tweet us @private_llm. If you enjoy the app, leaving an App Store is a great way to support us.

Оценки и отзывы

4,9 из 5
Оценок: 7

Оценок: 7

AVZ000 ,

Crash in new version

The app is great, however the version 1.9.1 (84) crash while loading the model: "Nous-Hermes 2 SOLAR 10.7B 4-bit OmniQuant"

Ответ разработчика ,

Thanks for the review! We're aware of this issue and it was caused due to a bug in an upstream tokenizer library that we use. We've managed to fix it and the update with the fix (v1.9.2) is currently under App Store review. The update should be out within a day.

ParamonSaurik ,

Шикарно

Добавьте выбор любой модели - и будет вообще супер. Огромное спасибо за вашу работу. Отлично работает на iphone и mac. Единственное, что представленные модели очень плохо работают с русским языком. Еще раз спасибо за то, что вы делаете)

George An. ,

Poor support of iOS and several other problems

First, I have an iPhone 14 Pro, and the app has layout issues. There's no option to save temperature for next launch. Opening llama 3 will crash the app. Shortcuts functionality is poor, as it doesn't allow custom system prompts only queries. Also, executing a shortcut function opens the app instead of quietly running in the background. Fixing these issues would instantly make this review a 5-star one

Ответ разработчика ,

Thanks for the review! Unfortunately, the Llama 3 8B model just barely fits in the iPhone 14 Pro's 6GB RAM. And this doesn't work if there are other memory hungry apps running in the background. We recommend using devices with 8GB or more RAM to run it. We decided to let users try it and decide for themselves, based on the feedback we got from our beta testers. I'd recommend trying any of the Llama 2 7B or Mistral 7B models. Also, the new Phi-3-Mini model. Regarding executing shortcut functions in the background, this is an iOS limitation. For performance reasons, iOS and tvOS do not let background apps use the GPU. macOS does not have this limitation, and consequently, the macOS version of the app runs perfectly fine in the background when invoked from Shortcuts.

События

Конфиденциальность приложения

Разработчик Numen Technologies Limited указал, что в соответствии с политикой конфиденциальности приложения данные могут обрабатываться так, как описано ниже. Подробные сведения доступны в политике конфиденциальности разработчика.

Сбор данных не ведется

Разработчик не ведет сбор данных в этом приложении.

Конфиденциальные данные могут использоваться по-разному в зависимости от вашего возраста, задействованных функций или других факторов. Подробнее

Поддерживается

  • Семейный доступ

    При включенном Семейном доступе приложением могут пользоваться до шести членов «Семьи».

Вам может понравиться

MLC Chat
Производительность
Local Brain
Производительность
YourChat
Производительность
PocketGPT: Private AI
Производительность
Hugging Chat
Производительность
PocketPal AI
Производительность