Dyad LogoDyad

Local models

Use local LLMs

Overview

You can use Dyad with local LLMs through Ollama, a free and open-source tool that makes it easy to run open-source language models directly on your device.

Running local LLMs can be resource-intensive, especially with larger or more advanced models. Smaller models may have trouble following Dyad's instructions and could produce less useful responses.

How to Use

Once Ollama is running on your device, open the model picker in Dyad, select Local Models, and you'll see the available models from Ollama.

Troubleshooting

If no models appear in the local models list, ensure that Ollama is running—specifically at http://localhost:11434, which is the default address.

On this page