BazaarLinkBazaarLink
登入
I

Mercury 2

Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving >1,000 tokens/sec on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost. Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output. Built for coding workflows where latency compounds, real-time voice/search, and agent loops. OpenAI API compatible. Read more in the [blog post](https://www.inceptionlabs.ai/blog/introducing-mercury-2).

定價資訊

輸入價格
$0.2500
/ 百萬 tokens
輸出價格
$0.7500
/ 百萬 tokens
上下文視窗
128K
tokens
供應商Iinception
模態text->text
發布日期2026年3月
Model IDinception/mercury-2

快速開始

只需更改 base_url 即可透過 BazaarLink API 使用 Mercury 2:

Python
from openai import OpenAI

client = OpenAI(
    base_url="https://bazaarlink.ai/api/v1",
    api_key="sk-bl-YOUR_API_KEY",
)

response = client.chat.completions.create(
    model="inception/mercury-2",
    messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)
TypeScript
import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://bazaarlink.ai/api/v1",
  apiKey: "sk-bl-YOUR_API_KEY",
});

const response = await client.chat.completions.create({
  model: "inception/mercury-2",
  messages: [{ role: "user", content: "Hello!" }],
});
console.log(response.choices[0].message.content);

為什麼透過 BazaarLink 使用 Mercury 2?

  • 美元計費(台幣報價)+ 統一發票 — 台灣團隊無需海外刷卡
  • OpenAI 相容 API — 無需改寫程式碼
  • 自動故障轉移 — 同模型多供應商備援
  • 中文客服支援 — 在地團隊即時協助
立即試用← 查看所有模型

Frequently Asked Questions

What is Mercury 2?

Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving >1,000 tokens/sec on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost. Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output. Built for coding workflows where latency compounds, real-time voice/search, and agent loops. OpenAI API compatible. Read more in the [blog post](https://www.inceptionlabs.ai/blog/introducing-mercury-2).

How much does the Mercury 2 API cost?

Mercury 2 costs $0.2500 per 1K input tokens and $0.7500 per 1K output tokens when accessed through BazaarLink.

How do I use Mercury 2 with the OpenAI SDK?

Set base_url to "https://bazaarlink.ai/api/v1" and use model ID "inception/mercury-2". All OpenAI SDK methods (chat.completions, embeddings, streaming) work without code changes.

What is the context window for Mercury 2?

Mercury 2 supports a context window of 128,000 tokens.

Is Mercury 2 available for free?

Mercury 2 is a paid model. BazaarLink offers free trial credits on registration so you can test it without a credit card.

Support
Support
Hi! How can we help you?
Send a message and we'll get back to you soon.
Mercury 2 API Pricing & Documentation — inception