Overview
Utter can run fully local in two stages:- Local transcription: Choose an on-device speech model (Parakeet) for voice-to-text.
- Local AI processing: Point Utter to a local OpenAI-compatible server (LM Studio or Ollama) for post-processing.
Stage 1: Use Local Transcription Models
To make transcription fully local, switch the default transcription model to a Parakeet model.Change the Default Transcription Model
At the top of Default Models, select a Parakeet model for Transcription. Built-in models are marked with a MacBook icon in the model list.

When a Parakeet model is selected, transcription stays on-device, requires no internet, and remains private.
Stage 2: Use Local AI Processing (LM Studio or Ollama)
Utter can also run AI post-processing locally by connecting to a local OpenAI-compatible server. Two popular options are Ollama and LM Studio. The steps below walk through LM Studio.Option A: LM Studio (Recommended Walkthrough)
Enable Power User Mode and Open the Developer Tab
Open LM Studio, enable Power User Mode, then open the Developer tab.

Load a Model and Start the Server
Make sure a model is loaded, enable the server, then copy the Model ID and Base URL from the Developer tab.

Open Utter Custom Models
In Utter, go to Settings > Advanced Settings > Custom Models, then choose Custom OpenAI Compatible as the provider.

Add the Local Model
Click Add Custom Model. Paste the Model ID and Base URL, then click Test to verify the connection and Save.

