mirror of
https://github.com/ollama/ollama.git
synced 2025-04-11 05:09:45 +02:00
docs: add tokenize and detokenize api
This commit is contained in:
parent
e60db349b7
commit
24613df094
79
docs/api.md
79
docs/api.md
@ -13,6 +13,8 @@
|
||||
- [Push a Model](#push-a-model)
|
||||
- [Generate Embeddings](#generate-embeddings)
|
||||
- [List Running Models](#list-running-models)
|
||||
- [Tokenize Text](#tokenize-text)
|
||||
- [Detokenize Text](#detokenize-text)
|
||||
|
||||
## Conventions
|
||||
|
||||
@ -1485,6 +1487,83 @@ A single JSON object will be returned.
|
||||
}
|
||||
```
|
||||
|
||||
## Tokenize Text
|
||||
|
||||
Tokenize text using a model
|
||||
|
||||
```shell
|
||||
POST /api/tokenize
|
||||
```
|
||||
|
||||
##### Parameters
|
||||
|
||||
- `model`: name of model to generate tokens from
|
||||
- `prompt`: text to generate tokens for
|
||||
|
||||
##### Advanced parameters:
|
||||
|
||||
- `options`: additional model parameters listed in the documentation for the [Modelfile](./modelfile.md#valid-parameters-and-values) such as `temperature`
|
||||
- `keep_alive`: controls how long the model will stay loaded into memory following the request (default: `5m`)
|
||||
|
||||
### Examples
|
||||
|
||||
#### Request
|
||||
|
||||
```shell
|
||||
curl -X POST http://localhost:11434/api/tokenize -d '{
|
||||
"model": "llama3.1:8b",
|
||||
"prompt": "Why the sky is blue?"
|
||||
}'
|
||||
```
|
||||
|
||||
#### Response
|
||||
|
||||
```json
|
||||
{
|
||||
"model": "llama3.1:8b",
|
||||
"tokens": [10445,279,13180,374,6437,30]
|
||||
}
|
||||
```
|
||||
|
||||
## Detokenize Text
|
||||
|
||||
Detokenize text using a model
|
||||
|
||||
```shell
|
||||
POST /api/detokenize
|
||||
```
|
||||
|
||||
#### Parameters
|
||||
|
||||
- `model`: name of model to generate text from
|
||||
- `tokens`: list of tokens to generate text from
|
||||
|
||||
##### Advanced parameters:
|
||||
|
||||
- `options`: additional model parameters listed in the documentation for the [Modelfile](./modelfile.md#valid-parameters-and-values) such as `temperature`
|
||||
- `keep_alive`: controls how long the model will stay loaded into memory following the request (default: `5m`)
|
||||
|
||||
### Examples
|
||||
|
||||
#### Request
|
||||
|
||||
```shell
|
||||
curl -X POST http://localhost:11434/api/detokenize -d '{
|
||||
"model": "llama3.1:8b",
|
||||
"tokens": [10445,279,13180,374,6437,30]
|
||||
}'
|
||||
```
|
||||
|
||||
#### Response
|
||||
|
||||
```json
|
||||
{
|
||||
"model": "llama3.1:8b",
|
||||
"text": "Why the sky is blue?"
|
||||
}
|
||||
```
|
||||
|
||||
|
||||
## Generate Embedding
|
||||
|
||||
> Note: this endpoint has been superseded by `/api/embed`
|
||||
|
Loading…
x
Reference in New Issue
Block a user