Skip to content

Lemonade

Lemonade Server lets you run local LLMs on your PC’s NPU and GPU. Free, private, and OpenAI-compatible.

  1. Install Lemonade Server - Download from lemonade-server.ai
  2. Install a model - Use the Lemonade GUI or CLI to download a model

Setting up Lemonade in Multi

  1. Open the Multi panel → Settings (gear icon)
  2. Click Add Profile
  3. Select Lemonade as the provider
  4. Set the base URL (default: http://localhost:13305/api/v1)
  5. Select your model
  6. Save
OptionDescription
Base URLLemonade server URL (default: http://localhost:13305/api/v1)
Model IDThe model to use (e.g., Llama-3.2-1B-Instruct-Hybrid)

Free - Lemonade Server is open-source and runs on your own hardware.