The API gateway that lets you control exactly who sees your data. Route to any model. Set your privacy level per request.
Route to the world's best models
Northern Inference infrastructure runs in AWS ca-central-1 (Montreal). Data at rest stays in Canada. We provide transparent chain-of-custody documentation showing exactly which entities handle your data at each privacy tier, so your security team can make informed decisions.
Inference runs in ca-central-1 (Montreal) and Azure Canada East. Data at rest stays in Canadian data centres. PIPEDA jurisdiction applies.
Every API request includes a clear record of which entities processed your data. You choose your privacy tier per request, from customer-hosted hardware (only you see data) to cloud providers with Canadian data residency.
Select your privacy level per request. Customer-hosted hardware for sensitive workloads, Canadian cloud for general use, or provider-default routes when Canadian data residency is not required.
Opt-in automatic PII detection and substitution. Names, emails, phone numbers, and other identifiers are replaced with realistic fakes before leaving your infrastructure, then restored in responses. Full audit log available.
Inherits AWS Canada's Cloud Service Provider attestations (SOC 2 Type II, ISO 27001, PCI-DSS). ITSG-33 PBMM control mapping in progress. Full status and roadmap on our Trust page.
Transparent per-token billing with no hidden costs. Clear invoicing in CAD. Easy to forecast and budget for departmental spending.
Drop-in replacement for existing AI tools and prototypes. Government developers can adopt without rewriting existing integrations.
Priority access and dedicated onboarding for government and public sector organizations.
Northern Inference uses the industry-standard chat completions API format. Point any compatible SDK to our endpoint, add a privacy tier, and you're done. No code changes needed.
Set your privacy tier on every API call. From zero-knowledge customer-hosted inference to provider-default global access, you control the trade-off.
Your hardware, your premises, open-source models. NI tunnels API requests into your machine; only you ever see prompts and completions. The strongest privacy guarantee we offer. Phase 5 of our roadmap.
AWS Nitro Enclaves with cryptographic isolation. Data encrypted with KMS before entering the enclave; AWS itself cannot read inside. Currently CPU-only (small models); GPU support pending Blackwell GA. Phase 5 of our roadmap.
AWS Bedrock and Azure OpenAI with Canadian data residency. Providers cannot train on your data. Infrastructure hosted by US-headquartered companies in Canadian data centres (ca-central-1, Canada East) under contractual data protection.
Broader model access through NI-routed upstream providers outside the Canadian-residency boundary. This can include direct APIs, Bedrock in US regions, Azure GlobalStandard, Vertex US regions, and other provider-default routes.
Tier 3: Canadian cloud. AWS Bedrock or Azure OpenAI in ca-central-1. Provider cannot train on your data.
All models available through the NI API with transparent per-token pricing. Choose by provider, tier, or data residency.
One API for every model. Switch providers with a single parameter.
Uses the industry-standard chat completions format. Change one URL and your existing code works with 100+ models across providers.
Anthropic, OpenAI, Google, Meta, Mistral, Cohere. Access the best model for each task through one gateway.
See the provider cost and our fee separately on every request. Thinking tokens billed at the rate shown per model. No expiring credits. No surprise overages.
Set your privacy tier on every API call. Route sensitive prompts through your own hardware, casual queries through cloud providers.
Names, emails, and identifiers are replaced with realistic fakes before reaching the model, then restored in responses. Opt-in per API key or per request.
Live dashboard with per-model, per-team cost breakdowns. Budget alerts and hard limits built in.
Infrastructure in Canada (ca-central-1). PIPEDA jurisdiction. Transparent data custody for organizations that need it.
Northern Inference is developer infrastructure, not a consumer chatbot.
| Northern Inference | Venice.ai | OpenRouter | Direct APIs | |
|---|---|---|---|---|
| Per-request privacy control | ||||
| Canadian data residency (per request, Tier 3) | ||||
| Transparent pricing | ||||
| Standard API format | ||||
| Per-request custody trail | ||||
| PII substitution | ||||
| No crypto required |
See the provider's cost and our routing fee separately on every request. No monthly minimums, no expiring credits, and thinking tokens are priced openly per model.
Early access members get priority onboarding + founding member pricing
Be among the first to use Northern Inference.
Each referral moves you closer to the front of the line.