Caricamento interfaccia
Cost-effective enterprise AI powered by Mistral Large 3 with 675B total parameters and 41B active parameters through granular Mixture-of-Experts architecture. Post-trained on 2,500 Claude Opus 4.5 outputs for exceptional instruction-following. NVFP4 quantization enables single-node deployment on H100s or A100s.
Shannon Lite 1.6 is built on Mistral Large 3, a state-of-the-art multimodal granular Mixture-of-Experts model designed from the ground up for reliability, long-context comprehension, and production-grade performance. The instruct post-trained version is fine-tuned for chat, agentic, and instruction-based use cases.
Granular MoE architecture with 39B active parameters per forward pass
Integrated multimodal encoder for image analysis and visual understanding
Extended context for comprehensive document understanding and RAG
English, French, Spanish, German, Chinese, Japanese, Korean, Arabic, and more
Shannon Lite 1.6 leverages NVIDIA's NVFP4 (4-bit floating point) quantization technology to dramatically reduce memory requirements while preserving model quality. Deploy frontier-class AI on accessible GPU infrastructure without multi-node complexity.
NVFP4 quantization reduces memory footprint by approximately 4x compared to BF16, enabling deployment on fewer GPUs and dramatically lowering TCO for enterprise AI.
Deploy the full 675B parameter model on a single node of H100s or A100s. No complex multi-node orchestration, reduced networking overhead, simplified operations.
Advanced quantization techniques maintain model performance across reasoning, instruction-following, and multimodal tasks with minimal quality degradation.
Shannon Lite 1.6 has been meticulously post-trained using 2,500 carefully curated outputs from Claude Opus 4.5, Anthropic's most capable model. This knowledge distillation approach captures advanced reasoning patterns, nuanced instruction interpretation, and superior response quality.
Built on Mistral's state-of-the-art Instruct model (version 2512) in BF16 precision. This foundation provides frontier-level capabilities engineered for production-grade assistants, retrieval-augmented systems, scientific workloads, and complex enterprise workflows.
Post-trained on 2,500 high-quality outputs from Claude Opus 4.5, capturing Anthropic's most advanced reasoning capabilities. The curated dataset focuses on complex instruction-following, nuanced understanding, and high-quality response generation across diverse domains.
Advanced NVIDIA FP4 quantization applied post-training to reduce memory footprint while maintaining model quality. Calibrated specifically for the post-trained weights to preserve the Claude Opus 4.5 knowledge transfer and instruction-following capabilities.
Comprehensive evaluation across instruction-following benchmarks, reasoning tasks, and real-world enterprise scenarios. Validated for consistent cross-domain behavior, stable outputs, and reliable performance in production environments.
Shannon Lite 1.6 with NVFP4 quantization enables cost-effective deployment on industry-standard NVIDIA GPU configurations, making frontier AI accessible for enterprise deployments without requiring expensive multi-node clusters.
Optimal performance with Hopper architecture and HBM3 memory
Proven reliability on Ampere architecture GPUs
Fully managed deployment with zero infrastructure
Shannon Lite 1.6 delivers frontier capabilities inherited from Mistral Large 3 and enhanced through Claude Opus 4.5 post-training, optimized for production workloads across diverse enterprise scenarios.
Integrated 2.5B parameter vision encoder enables image analysis, visual question answering, and document understanding with images.
Native support for 12+ languages including English, French, Spanish, German, Italian, Portuguese, Dutch, Chinese, Japanese, Korean, and Arabic.
Best-in-class agentic features with native function calling and structured JSON output for autonomous tool use and workflow automation.
Strong adherence and support for system prompts, enabling precise behavioral control and consistent persona maintenance.
Extended context window for comprehensive document understanding, extended conversations, and retrieval-augmented generation (RAG).
Built-in function calling support with reliable JSON output for seamless integration with external tools, APIs, and services.
With powerful long-context performance, stable and consistent cross-domain behavior, Shannon Lite 1.6 excels across diverse enterprise and research scenarios.
Process and analyze extensive documents, contracts, reports, and research papers with the 256K context window
Power daily-driver AI assistants with reliable, consistent responses and strong instruction-following
State-of-the-art tool use and function calling for autonomous task execution and workflow automation
Complex enterprise workflows requiring frontier AI capabilities with consistent, reliable outputs
Code generation, debugging, documentation, and software development assistance across multiple languages
Research assistance, literature review, scientific workload processing, and hypothesis generation
Optimal performance for RAG systems with reliable context integration and accurate retrieval synthesis
Global enterprise applications requiring consistent quality across 12+ supported languages
Choose the right Shannon model for your needs. Shannon Lite offers cost-effective enterprise deployment, while Shannon Pro provides maximum capability with advanced chain-of-thought reasoning and Skills support.
| Feature | Shannon Lite 1.6 | Shannon Pro 1.6 |
|---|---|---|
| Base Model | Mistral Large 3 (675B) | Mistral Large 3 (675B) |
| Active Parameters | 41B (Granular MoE) | 41B (Granular MoE) |
| Precision | NVFP4 (4-bit) | Full BF16 (16-bit) |
| Post-Training Data | 2,500 Claude Opus 4.5 outputs | KIMI K2 Thinking Traces |
| Post-Training Method | Supervised Fine-Tuning | GRPO (Group Relative Policy Optimization) |
| Reasoning Mode | Standard | Chain-of-Thought Traces |
| Skills Support | - | Pro Only Native Skills |
| Deployment | H100/A100 (Single Node) | B200/H200 (FP8) |
| Best For | Cost-Effective Enterprise AI | Maximum Capability + Reasoning |
Frontier AI capabilities with cost-effective NVFP4 quantization. Deploy on H100 or A100 infrastructure for enterprise-grade performance at accessible cost.