diff --git a/packages/web/astro.config.mjs b/packages/web/astro.config.mjs index 612d4fb8c..110c8ce91 100644 --- a/packages/web/astro.config.mjs +++ b/packages/web/astro.config.mjs @@ -224,7 +224,7 @@ export default defineConfig({ "zh-CN": "使用", "zh-TW": "使用", }, - items: ["tui", "cli", "web", "ide", "zen", "share", "github", "gitlab"], + items: ["go", "tui", "cli", "web", "ide", "zen", "share", "github", "gitlab"], }, { diff --git a/packages/web/src/content/docs/go.mdx b/packages/web/src/content/docs/go.mdx new file mode 100644 index 000000000..9c95c0ded --- /dev/null +++ b/packages/web/src/content/docs/go.mdx @@ -0,0 +1,147 @@ +--- +title: Go +description: Low cost subscription for open coding models. +--- + +import config from "../../../config.mjs" +export const console = config.console +export const email = `mailto:${config.email}` + +OpenCode Go is a low cost subscription that gives you reliable access to popular open coding models, with the goal of making AI coding accessible to more people. + +:::note +OpenCode Go is currently in beta. +::: + +Go works like any other provider in OpenCode. You subscribe to OpenCode Go and +get your API key. It's **completely optional** and you don't need to use it to +use OpenCode. + +--- + +## Background + +Open models have gotten really good. They now reach performance close to +proprietary models for coding tasks. And because many providers can serve them +competitively, they are usually far cheaper. + +However, getting reliable, low latency access to them can be difficult. Providers +vary in quality and availability. + +:::tip +We tested a select group of models and providers that work well with OpenCode. +::: + +To fix this, we did a couple of things: + +1. We tested a select group of open models and talked to their teams about how to + best run them. +2. We then worked with a few providers to make sure these were being served + correctly. +3. Finally, we benchmarked the combination of the model/provider and came up + with a list that we feel good recommending. + +OpenCode Go gives you access to these models for **$10/month**. + +--- + +## How it works + +OpenCode Go works like any other provider in OpenCode. + +1. You sign in to **OpenCode Zen**, subscribe to Go, and + copy your API key. +2. You run the `/connect` command in the TUI, select `OpenCode Go`, and paste + your API key. +3. Run `/models` in the TUI to see the list of models available through Go. + +:::note +Only one member per workspace can subscribe to OpenCode Go. +::: + +The current list of models includes: + +- **Kimi K2.5** +- **GLM-5** +- **MiniMax M2.5** + +The list of models may change as we test and add new ones. + +--- + +## Usage limits + +Go includes generous usage limits with three tiers: + +- **5 hour limit** — $4 worth of usage +- **Weekly limit** — $10 worth of usage +- **Monthly limit** — $20 worth of usage + +To give you an idea of the limit in terms of tokens, $20 roughly gets you: + +- 69 million GLM 5 tokens +- 121 million Kimi K2.5 tokens +- 328 million MiniMax M2.5 tokens + +You can view your current usage in the **console**. + +:::tip +If you hit a usage limit, you can continue using the free models available. +::: + +Usage limits may change as we learn from early usage and feedback. + +--- + +### Pricing + +Below are the prices **per 1M tokens**. + +| Model | Input | Output | Cached Read | +| ------------ | ----- | ------ | ----------- | +| GLM 5 | $1.00 | $3.20 | $0.20 | +| Kimi K2.5 | $0.60 | $3.00 | $0.10 | +| MiniMax M2.5 | $0.30 | $1.20 | $0.03 | + +--- + +### Use balance after limits + +If you also have credits on your Zen balance, you can enable the **Use balance** +option in the console. When enabled, Go will fall back to your Zen balance +after you've reached your usage limits instead of blocking requests. + +--- + +## Endpoints + +You can also access Go models through the following API endpoints. + +| Model | Model ID | Endpoint | AI SDK Package | +| ------------ | ------------ | ------------------------------------------------ | --------------------------- | +| GLM 5 | glm-5 | `https://opencode.ai/zen/go/v1/chat/completions` | `@ai-sdk/openai-compatible` | +| Kimi K2.5 | kimi-k2.5 | `https://opencode.ai/zen/go/v1/chat/completions` | `@ai-sdk/openai-compatible` | +| MiniMax M2.5 | minimax-m2.5 | `https://opencode.ai/zen/go/v1/messages` | `@ai-sdk/anthropic` | + +The [model id](/docs/config/#models) in your OpenCode config +uses the format `opencode-go/`. For example, for Kimi K2.5, you would +use `opencode-go/kimi-k2.5` in your config. + +--- + +## Privacy + +The plan is designed primarily for international users, with models hosted in the US, EU, and Singapore for stable global access. + +Contact us if you have any questions. + +--- + +## Goals + +We created OpenCode Go to: + +1. Make AI coding **accessible** to more people with a low cost subscription. +2. Provide **reliable** access to the best open coding models. +3. Curate models that are **tested and benchmarked** for coding agent use. +4. Have **no lock-in** by allowing you to use any other provider with OpenCode as well.