Today, we are thrilled to announce that DeepSeek R1 distilled Llama and Qwen designs are available through Amazon Bedrock Marketplace and Amazon SageMaker JumpStart. With this launch, you can now release DeepSeek AI's first-generation frontier design, DeepSeek-R1, together with the distilled variations varying from 1.5 to 70 billion specifications to develop, experiment, and properly scale your generative AI ideas on AWS.
In this post, we demonstrate how to begin with DeepSeek-R1 on Amazon Bedrock Marketplace and SageMaker JumpStart. You can follow similar steps to release the distilled versions of the designs too.
Overview of DeepSeek-R1
DeepSeek-R1 is a large language model (LLM) developed by DeepSeek AI that uses reinforcement learning to improve thinking abilities through a multi-stage training procedure from a DeepSeek-V3-Base foundation. A crucial identifying feature is its reinforcement learning (RL) step, which was utilized to fine-tune the design's responses beyond the standard pre-training and tweak process. By including RL, DeepSeek-R1 can adapt more effectively to user feedback and objectives, eventually enhancing both importance and clearness. In addition, DeepSeek-R1 employs a chain-of-thought (CoT) approach, suggesting it's geared up to break down complex inquiries and reason through them in a detailed way. This guided reasoning process enables the model to produce more accurate, transparent, and detailed responses. This model integrates RL-based fine-tuning with CoT abilities, aiming to generate structured actions while concentrating on interpretability and user interaction. With its wide-ranging capabilities DeepSeek-R1 has recorded the industry's attention as a versatile text-generation model that can be incorporated into numerous workflows such as agents, logical reasoning and information interpretation tasks.
DeepSeek-R1 uses a Mixture of Experts (MoE) architecture and is 671 billion parameters in size. The MoE architecture permits activation of 37 billion criteria, wavedream.wiki making it possible for efficient inference by routing questions to the most relevant professional "clusters." This technique allows the model to specialize in different issue domains while maintaining general efficiency. DeepSeek-R1 requires a minimum of 800 GB of HBM memory in FP8 format for inference. In this post, we will utilize an ml.p5e.48 xlarge circumstances to release the model. ml.p5e.48 xlarge features 8 Nvidia H200 GPUs supplying 1128 GB of GPU memory.
DeepSeek-R1 distilled models bring the reasoning abilities of the main R1 model to more efficient architectures based upon popular open designs like Qwen (1.5 B, 7B, 14B, and wiki.myamens.com 32B) and Llama (8B and 70B). Distillation refers to a process of training smaller, more efficient designs to simulate the behavior and reasoning patterns of the bigger DeepSeek-R1 design, utilizing it as a teacher model.
You can release DeepSeek-R1 design either through SageMaker JumpStart or Bedrock Marketplace. Because DeepSeek-R1 is an emerging design, we recommend releasing this model with guardrails in place. In this blog, we will use Amazon Bedrock Guardrails to introduce safeguards, prevent damaging content, and assess designs against key safety criteria. At the time of composing this blog, for DeepSeek-R1 implementations on SageMaker JumpStart and Bedrock Marketplace, Bedrock Guardrails supports only the ApplyGuardrail API. You can develop multiple guardrails tailored to different usage cases and use them to the DeepSeek-R1 design, improving user experiences and standardizing security controls throughout your generative AI applications.
Prerequisites
To release the DeepSeek-R1 design, you need access to an ml.p5e instance. To inspect if you have quotas for P5e, open the Service Quotas console and under AWS Services, pick Amazon SageMaker, and validate you're using ml.p5e.48 xlarge for endpoint use. Make certain that you have at least one ml.P5e.48 xlarge circumstances in the AWS Region you are deploying. To ask for a limitation boost, develop a limitation boost request and connect to your account team.
Because you will be releasing this design with Amazon Bedrock Guardrails, make certain you have the correct AWS Identity and [forum.batman.gainedge.org](https://forum.batman.gainedge.org/index.php?action=profile
1
DeepSeek R1 Model now Available in Amazon Bedrock Marketplace And Amazon SageMaker JumpStart
percyovens7386 edited this page 4 months ago