About 251,000 results
Open links in new tab
  1. [2401.14196] DeepSeek-Coder: When the Large Language Model

    Jan 25, 2024 · Our extensive evaluations demonstrate that DeepSeek-Coder not only achieves state-of-the-art performance among open-source code models across multiple benchmarks but also surpasses existing closed-source models like Codex and GPT-3.5.

  2. DeepSeek-Coder: When the Large Language Model Meets …

    Our extensive evaluations demonstrate that DeepSeek-Coder not only achieves state-of-the-art performance among open-source code models across multiple benchmarks but also surpasses existing closed-source models like Codex and GPT-3.5.

  3. GitHub - deepseek-ai/DeepSeek-Coder: DeepSeek Coder: Let …

    DeepSeek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese. We provide various sizes of the code model, ranging from 1B to 33B versions.

  4. Deepseek Coder: A new line of high quality coding models!

    DeepSeek Coder comprises a series of code language models trained on both 87% code and 13% natural language in English and Chinese, with each model pre-trained on 2T tokens. We provide various sizes of the code model, ranging from 1B to 33B versions.

  5. GitHub - deepseek-ai/DeepSeek-Coder-V2: DeepSeek-Coder

    We present DeepSeek-Coder-V2, an open-source Mixture-of-Experts (MoE) code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. Specifically, DeepSeek-Coder-V2 is further pre-trained from an intermediate checkpoint of DeepSeek-V2 with additional 6 trillion tokens.

  6. DeepSeek-Coder: Code Smarter” | by Jyoti Dabass, Ph.D ...

    Jan 29, 2025 · The DeepSeek-Coder model consists of the following components: Encoder 📝: The encoder takes in a sequence of tokens (e.g., code, natural language, or a mix of both) and outputs a continuous representation of the input sequence. Decoder 📢: The decoder generates a sequence of tokens based on the output of the encoder.

  7. Comparing DeepSeek Models: R1 vs. V2 vs. Coder - Medium

    Jan 29, 2025 · Overview: DeepSeek Coder is a specialized AI model designed for software development, rivaling models like Codex and GitHub Copilot. Why It Matters: DeepSeek Coder enhances productivity by...

  8. deepseek-ai/deepseek-coder-6.7b-base - Hugging Face

    Aug 16, 2024 · Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese. We provide various sizes of the …

  9. deepseek-ai/DeepSeek-Coder-V2-Base · Hugging Face

    We present DeepSeek-Coder-V2, an open-source Mixture-of-Experts (MoE) code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. Specifically, DeepSeek-Coder-V2 is further pre-trained from an intermediate checkpoint of DeepSeek-V2 with additional 6 trillion tokens.

  10. DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source …

    Feb 24, 2025 · DeepSeek-Coder-V2 offers improved efficiency in code generation, completion, and chat-based interactions with several model variants meant for various use cases. Its model variants are discussed in this article along with comprehensive instruction on running DeepSeek-Coder-V2 locally.

  11. Some results have been removed
Refresh