⚡Preview capacity — responses may take a minute while we
bring dedicated GPU inference online.
The deliberation protocol is already production-grade; latency here reflects shared-tenant inference, not the architecture. Self-hosted and reserved-capacity deployments run at native speed.
Option details
Claims
Reasoning
MVP Preview
nsed:mid
377B parameters · 31B active · mixture of models
Multiple models fused into one via the
NSED algorithm.
Paste your API token above to start chatting.
Request AccessI already have a token — let me inDon't show again
dismiss ×
Slower. Smarter.
Answers take 30 seconds to a few minutes.
Best for complex, non-trivial questions.
Nudge mid-stream or pick an early option if you've seen enough.
dismiss ×
Resume previous session?
Your last deliberation may still be running or recently completed.