Jeffrey Morgan 76cb60d496 wip go engine
Co-authored-by: Patrick Devine <pdevine@sonic.net>
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 00:04:06 -04:00
2023-07-06 16:34:44 -04:00
2023-07-01 17:54:29 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:18:40 -04:00
2023-07-02 17:34:46 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-06 16:34:44 -04:00
2023-07-02 20:33:23 -04:00
2023-07-06 16:34:44 -04:00

Ollama

A fast runtime for large language models, powered by llama.cpp.

Note: this project is a work in progress. Certain models that can be run with ollama are intended for research and/or non-commercial use only.

Install

Using pip:

pip install ollama

Using docker:

docker run ollama/ollama

Quickstart

To run a model, use ollama run:

ollama run orca-mini-3b

You can also run models from hugging face:

ollama run huggingface.co/TheBloke/orca_mini_3B-GGML

Or directly via downloaded model files:

ollama run ~/Downloads/orca-mini-13b.ggmlv3.q4_0.bin

Documentation

Description
Get up and running with Llama 2, Mistral, and other large language models.
Readme MIT 155 MiB
Languages
Go 93.5%
C 2.5%
Shell 1.2%
TypeScript 1%
PowerShell 0.7%
Other 1%