April 24, 2026
Model Catalog Updates- DeepSeek-V3.1 moved to Global Model Catalog: Now served via proxy (US region) instead of EU-hosted infrastructure
- gemma-3-12b-it added: Google’s Gemma 3 12B model now available via Global Model Catalog (Japan). This is the first vision-capable model on Infercom — see Vision guide
- DeepSeek-V3.2 context length corrected: Updated from 8k to 32k tokens
- EU-hosted models: MiniMax-M2.5 and gpt-oss-120b remain as the sovereign EU options
- Updated Vision guide with gemma-3-12b-it examples and sovereignty notice
- Added Capabilities column to model overview tables
April 6, 2026
New Documentation: Agentic Coding- Added Agentic Coding section with guides for 9 AI coding tools
- Terminal tools: Aider, OpenCode, Goose, Warp
- IDE tools: Cursor, Cline, Continue, Windsurf
- Advanced: Claude Code via ccproxy
- Tool comparison guide - Recommendations for choosing the right tool
April 3, 2026
Global Model Catalog Streamlined- Reduced Global Model Catalog to 2 models:
Meta-Llama-3.3-70B-Instruct(128k) andDeepSeek-V3.2(8k) - 8 models deprecated: Qwen3-32B, Qwen3-235B, DeepSeek-R1-0528, DeepSeek-R1-Distill-Llama-70B, DeepSeek-V3-0324, DeepSeek-V3.1-Terminus, Llama-4-Maverick-17B-128E-Instruct, Meta-Llama-3.1-8B-Instruct
- See deprecations for migration guidance
- Fixed MiniMax-M2.5 context length to 160k (was incorrectly listed as 164k)
March 12, 2026
Model Update: MiniMax M2.5 replaces Meta-Llama-3.3-70B-Instruct- MiniMax M2.5 is now available as an EU-hosted model on Infercom’s sovereign infrastructure in Germany
- Meta-Llama-3.3-70B-Instruct has been deprecated and removed from the platform. See deprecations for migration guidance
- Rate limits updated for MiniMax M2.5
February 27, 2026
Global Model Catalog Launch- 9 new models available through the Global Model Catalog, including DeepSeek-R1-0528, Qwen3-235B, Llama-4-Maverick, and more
- EU-hosted and globally-routed models are now clearly separated in the model overview
- Rate limits updated for all 12 models across Free and Developer tiers
- API
sn_metadatanow returnsis_externalandregionfields for all models
February 14, 2026
New Documentation- Added Performance & Latency guide covering connection pooling, response performance metadata, streaming optimization, and best practices for minimizing latency
February 5, 2026
Documentation for Global Model Catalog- Added documentation for the Global Model Catalog, covering EU-hosted and globally-routed models
- Documented how to identify model regions via the API (
sn_metadata.region) and the Playground - Updated API reference with
sn_metadataschema and?verbose=truequery parameter - Reviewed and qualified sovereignty claims across documentation to reflect Global Model Catalog
January 23, 2026
Model Update- Updated DeepSeek model from
DeepSeek-V3-0324-cbtoDeepSeek-V3.1with expanded 128k context window - The previous model has been deprecated and is listed on the deprecations page
- Updated rate limits for all models
January 13, 2026
New- Hugging Face Spaces Chat Demo - Interactive chat with model selection and real-time performance metrics
- Documentation cleanup and link fixes
December 30, 2025
Updated the model catalog to reflect currently available models on Infercom Inference Service. The model list and rate limits documentation have been updated accordingly.November 17, 2025
We are pleased to announce the launch of Infercom’s documentation portal. This comprehensive documentation has been created based on SambaNova’s documentation (dated October 7, 2025) and adapted for Infercom’s EU sovereign AI inference platform. Key features- Complete API reference documentation with OpenAI-compatible endpoints.
- Developer guides for integrating with Infercom’s inference platform.
- Model catalog and configuration documentation for all supported models.
- Platform architecture and deployment guides.
- Usage examples in Python and TypeScript, powered by OpenAI-compatible SDKs.