Model Selection
Switch between cloud and local models based on your needs. Use cloud models (Claude, GPT, Gemini) for faster responses and advanced reasoning. Use local models for complete privacy and offline capability.
Model selector dropdown
LLM Runtime Modal
Click the active model control (for example the model name in the bottom toolbar) to open the LLM runtime modal. There you can enter API keys if needed, switch models, and open the full catalog of local and cloud-hosted models served through Pieces.
Cloud models from OpenAI, Anthropic, and Google (and others) are listed on the Cloud Models page. For on-device options and privacy, use local models; the full catalog for the Desktop App is in supported local and cloud models.
Browse and Download Local Models
Open the LLM runtime modal, open All Models, then scroll to find local models. Select a model to download it on demand through PiecesOS; once downloaded, you can run it entirely on your device.
Chat Appearance and Defaults
In the LLM runtime area, open the Settings gear to set a chat accent color and choose whether LTM context is on by default for new chats.
You can also use cmd+shift+t (macOS) or ctrl+shift+t (Windows/Linux) to toggle the Desktop App Dark/Bright theme.
For detailed model configuration and management, see Configuration > Models.