DeepSeek Coder 7B Base
accounts/fireworks/models/deepseek-coder-7b-base
Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese. Deepseek Coder 6.7B Base is a 6.7B parameter model with Multi-Head Attention trained on 2 trillion tokens by employing a window size of 16K and an extra fill-in-the-blank task
On-demand deployments allow you to use DeepSeek Coder 7B Base on dedicated GPUs with Fireworks' high-performance serving stack with high reliability and no rate limits.