DeepSeek-V3
DeepSeek has launched DeepSeek-V3, which processes 60 tokens per second, features 671 billion parameters, and maintains open-source compatibility. Pricing changes will occur after February 8, 2024, with future updates planned.
Read original articleDeepSeek has announced the release of DeepSeek-V3, marking a significant advancement in their API capabilities. The new version boasts a processing speed of 60 tokens per second, which is three times faster than its predecessor, DeepSeek-V2. It features 671 billion mixture of experts (MoE) parameters, with 37 billion activated parameters, and has been trained on 14.8 trillion high-quality tokens. The API remains compatible with previous versions, and the models and research papers are fully open-source. Pricing for the API will remain the same as V2 until February 8, after which new rates will apply: $0.27 per million tokens for cache misses, $0.07 for cache hits, and $1.10 for output. DeepSeek emphasizes its commitment to open-source principles and aims to bridge the gap between open and closed models in the AI landscape. Future updates are expected to include multimodal support and other innovative features, reinforcing DeepSeek's mission to advance inclusive artificial general intelligence (AGI).
- DeepSeek-V3 is three times faster than V2, processing 60 tokens per second.
- The new model includes 671 billion MoE parameters and is trained on 14.8 trillion tokens.
- API pricing will remain unchanged until February 8, 2024, after which new rates will be implemented.
- DeepSeek is committed to open-source development and aims to enhance the AI community.
- Future updates will introduce multimodal support and additional features.
Related
Cerebras Inference now 3x faster: Llama3.1-70B breaks 2,100 tokens/s
Cerebras has enhanced its Inference platform, achieving a threefold speed increase for the Llama 3.1-70B model, now processing 2,100 tokens per second, benefiting various industries with real-time AI applications.
DeepSeek v2.5 – open-source LLM comparable to GPT-4o, but 95% less expensive
DeepSeek launched DeepSeek-V2.5, an advanced open-source model with a 128K context length, excelling in math and coding tasks, and offering competitive API pricing for developers.
32k context length text embedding models
Voyage AI released two embedding models, voyage-3 and voyage-3-lite, enhancing retrieval quality, reducing costs, and supporting a 32K-token context length, outperforming OpenAI's models in various domains.
DeepSeek v3 beats Claude sonnet 3.5 and way cheaper
DeepSeek-V3 is a 671 billion parameter language model that excels in benchmarks, particularly math and coding tasks, utilizing advanced training strategies and supporting various hardware for local deployment.
DeepSeek-V3, ultra-large open-source AI, outperforms Llama and Qwen on launch
Chinese AI startup DeepSeek launched DeepSeek-V3, a 671 billion parameter model outperforming major competitors. It features cost-effective training, innovative architecture, and is available for testing and commercial use.
Related
Cerebras Inference now 3x faster: Llama3.1-70B breaks 2,100 tokens/s
Cerebras has enhanced its Inference platform, achieving a threefold speed increase for the Llama 3.1-70B model, now processing 2,100 tokens per second, benefiting various industries with real-time AI applications.
DeepSeek v2.5 – open-source LLM comparable to GPT-4o, but 95% less expensive
DeepSeek launched DeepSeek-V2.5, an advanced open-source model with a 128K context length, excelling in math and coding tasks, and offering competitive API pricing for developers.
32k context length text embedding models
Voyage AI released two embedding models, voyage-3 and voyage-3-lite, enhancing retrieval quality, reducing costs, and supporting a 32K-token context length, outperforming OpenAI's models in various domains.
DeepSeek v3 beats Claude sonnet 3.5 and way cheaper
DeepSeek-V3 is a 671 billion parameter language model that excels in benchmarks, particularly math and coding tasks, utilizing advanced training strategies and supporting various hardware for local deployment.
DeepSeek-V3, ultra-large open-source AI, outperforms Llama and Qwen on launch
Chinese AI startup DeepSeek launched DeepSeek-V3, a 671 billion parameter model outperforming major competitors. It features cost-effective training, innovative architecture, and is available for testing and commercial use.