Mercury 2
Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving >1,000 tokens/sec on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost. Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output. Built for coding workflows where latency compounds, real-time voice/search, and agent loops. OpenAI API compatible. Read more in the [blog post](https://www.inceptionlabs.ai/blog/introducing-mercury-2).
定價資訊
快速開始
只需更改 base_url 即可透過 BazaarLink API 使用 Mercury 2:
from openai import OpenAI
client = OpenAI(
base_url="https://bazaarlink.ai/api/v1",
api_key="sk-bl-YOUR_API_KEY",
)
response = client.chat.completions.create(
model="inception/mercury-2",
messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://bazaarlink.ai/api/v1",
apiKey: "sk-bl-YOUR_API_KEY",
});
const response = await client.chat.completions.create({
model: "inception/mercury-2",
messages: [{ role: "user", content: "Hello!" }],
});
console.log(response.choices[0].message.content);為什麼透過 BazaarLink 使用 Mercury 2?
- ✓美元計費(台幣報價)+ 統一發票 — 台灣團隊無需海外刷卡
- ✓OpenAI 相容 API — 無需改寫程式碼
- ✓自動故障轉移 — 同模型多供應商備援
- ✓中文客服支援 — 在地團隊即時協助
Frequently Asked Questions
What is Mercury 2?
Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving >1,000 tokens/sec on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost. Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output. Built for coding workflows where latency compounds, real-time voice/search, and agent loops. OpenAI API compatible. Read more in the [blog post](https://www.inceptionlabs.ai/blog/introducing-mercury-2).
How much does the Mercury 2 API cost?
Mercury 2 costs $0.2500 per 1K input tokens and $0.7500 per 1K output tokens when accessed through BazaarLink.
How do I use Mercury 2 with the OpenAI SDK?
Set base_url to "https://bazaarlink.ai/api/v1" and use model ID "inception/mercury-2". All OpenAI SDK methods (chat.completions, embeddings, streaming) work without code changes.
What is the context window for Mercury 2?
Mercury 2 supports a context window of 128,000 tokens.
Is Mercury 2 available for free?
Mercury 2 is a paid model. BazaarLink offers free trial credits on registration so you can test it without a credit card.