Kim Hallberg

10h ago

Ollama Turbo now in preview

Introduced with the release of Ollama's support for @GPT OSS, is Turbo; Ollama's privacy-first datacenter-grade cloud inference service.

Whilst it's currently in preview, the service costs $20/m, and has both hourly and daily limits. Usage-based pricing will be available soon. So far, the service only has gpt-oss-12b and gpt-oss-120b models, and works with Ollama's App, CLI, and API.

Chris Messina

2yr ago

Ollama - The easiest way to run large language models locally

Run Llama 2 and other models on macOS, with Windows and Linux coming soon. Customize and create your own.