From ee92ca3e1d1f95ce5372a395d19b6a5ed11afaeb Mon Sep 17 00:00:00 2001 From: Jeffrey Morgan Date: Tue, 5 Aug 2025 13:09:10 -0700 Subject: [PATCH] docs: add docs for Ollama Turbo (#11687) --- docs/turbo.md | 107 ++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 107 insertions(+) create mode 100644 docs/turbo.md diff --git a/docs/turbo.md b/docs/turbo.md new file mode 100644 index 0000000000..8c45d39ceb --- /dev/null +++ b/docs/turbo.md @@ -0,0 +1,107 @@ +# Turbo + +> ⚠️ Turbo is preview + +Ollama’s [Turbo](https://ollama.com/turbo) is a new way to run open-source models with acceleration from datacenter-grade hardware. + +Currently, the following models are available in Turbo: + +- `gpt-oss:20b` +- `gpt-oss:120b` + +## Get started + +### Ollama for macOS & Windows + +Download Ollama + +- Select a model such as `gpt-oss:20b` or `gpt-oss:120b` +- Click on **Turbo**. You’ll be prompted to create an account or sign in + +### Ollama’s CLI + +- [Sign up](https://ollama.com/signup) for an Ollama account +- Add your Ollama key [to ollama.com](https://ollama.com/settings/keys). + + On macOS and Linux: + + ```shell + cat ~/.ollama/id_ed25519.pub + ``` + + On Windows: + + ``` + type "%USERPROFILE%\.ollama\id_ed25519.pub" + ``` + +- Then run a model setting `OLLAMA_HOST` to `ollama.com`: + ```shell + OLLAMA_HOST=ollama.com ollama run gpt-oss:120b + ``` + +### Ollama’s Python library + +- Download Ollama's [Python library](https://github.com/ollama/ollama-python) +- [Sign up](https://ollama.com/signup) for an Ollama account +- Create an API key by visiting https://ollama.com/settings/keys + +```python +from ollama import Client + +client = Client( + host="https://ollama.com", + headers={'Authorization': ''} +) + +messages = [ + { + 'role': 'user', + 'content': 'Why is the sky blue?', + }, +] + +for part in client.chat('gpt-oss:120b', messages=messages, stream=True): + print(part['message']['content'], end='', flush=True) +``` + +### Ollama’s JavaScript library + +- Download Ollama's [JavaScript library](https://github.com/ollama/ollama-js) +- [Sign up](https://ollama.com/signup) for an Ollama account +- Create an API key by visiting https://ollama.com/settings/keys + +```typescript +import { Ollama } from 'ollama'; + +const ollama = new Ollama({ + host: 'https://ollama.com' + headers: { + Authorization: "Bearer " + } +}); + +const response = await ollama.chat({ + model: 'deepseek-r1:671b', + messages: [{ role: 'user', content: 'Explain quantum computing' }], + stream: true +}); + +for await (const part of response) { + process.stdout.write(part.message.content) +} +``` + +### Community integrations + +Turbo mode is also compatible with several community integrations. + +#### Open WebUI + +- Go to **settings** → **Admin settings** → **Connections** +- Under **Ollama API,** click **+** +- For the **URL** put `https://ollama.com` +- For the **API key,** create an API key on https://ollama.com/settings/keys and add it. +- Click **Save** + +Now, if you navigate to the model selector, Turbo models should be available under **External**.