Desktop App¶
The desktop app is the best fit for long-term Prompt Optimizer usage, especially when you need local models, custom gateways, or internal enterprise APIs.
Why We Recommend The Desktop App First¶
- No browser CORS or mixed-content restrictions
- Easier access to
http://localhost, LAN services, and internal APIs - Shares the main workspace capabilities with the web app
- Keeps data inside the local application directory for easier backup
- Built-in update checks are available in release builds, although actual update availability still depends on platform and distribution channel
Downloads¶
GitHub Releases: https://github.com/linshenkx/prompt-optimizer/releases
The current desktop product name is PromptOptimizer. Build artifacts follow this naming pattern:
PromptOptimizer-<version>-<os>-<arch>.<ext>
Common examples:
- Windows installer:
PromptOptimizer-<version>-win-x64.exe - Windows archive:
PromptOptimizer-<version>-win-x64.zip - macOS disk image:
PromptOptimizer-<version>-mac-arm64.dmg - Linux AppImage:
PromptOptimizer-<version>-linux-x64.AppImage
Installation¶
Windows¶
- Download the
.exeinstaller. - Double-click to install.
- If SmartScreen appears, choose "More info" and continue.
macOS¶
- Download the
.dmg. - Drag the app into
Applications. - If the first launch is blocked, allow it in "System Settings -> Privacy & Security".
Linux¶
- Download the
.AppImageor.zip. - If you use the AppImage build, run:
chmod +x PromptOptimizer-<version>-linux-x64.AppImage
- Then launch the file.
Recommended First-Time Setup¶
- Open the app.
- Go to Model Management.
- Configure your primary text models.
- If you use Ollama, prefer the built-in
Ollamaprovider directly. - If you use LM Studio, an enterprise gateway, or another OpenAI-compatible endpoint, use
Custom. - If you plan to use image modes, configure the relevant image models as well.
- Enter the workspace that matches your task.
Example: Connect A Local Model¶
Ollama¶
Provider: Ollama
Base URL: http://localhost:11434/v1
Model: qwen2.5:7b
API Key: usually can be left empty
LM Studio¶
Provider: Custom
Base URL: http://localhost:1234/v1
Model: the model name currently exposed by LM Studio
API Key: any non-empty string
When To Still Use Custom¶
Custom is usually a better fit when your service is one of the following:
- LM Studio
- vLLM / OneAPI / a self-hosted gateway
- An OpenAI-compatible endpoint inside an enterprise network
- A setup where you need a special Base URL or extra request parameters
For localhost, private-network, and LAN OpenAI-compatible endpoints, the desktop app can route requests directly where appropriate. This avoids common proxy or browser-network misrouting issues while still allowing public HTTPS providers to use the normal network path.
Data And Logs¶
The desktop app uses the standard OS application-data directory, and the folder name follows the product name PromptOptimizer.
Common locations:
Windows: %APPDATA%\PromptOptimizer\
macOS: ~/Library/Application Support/PromptOptimizer/
Linux: ~/.config/PromptOptimizer/
If you mainly need logs, they are typically under:
%APPDATA%\PromptOptimizer\logs\
When You Probably Do Not Need The Desktop App¶
If you only need:
- occasional access
- public HTTPS model APIs
- quick temporary usage across devices
then the hosted site or a self-deployed web app is usually enough.
FAQ¶
Does the desktop app match the Chinese documentation feature set?¶
Yes. The current desktop app reuses the core application workspaces and most of the same features. The Chinese documentation reflects the current implementation.
Is the desktop app always faster than the web app?¶
Not always, but it is usually more stable when you depend on local models, custom endpoints, or more complex network environments.
Can desktop app data be migrated?¶
Yes. Prefer using Data Management inside the app for export and import instead of editing files directly inside the application directory.