From e975062c42acae4df5a36c6d348e302021f7a185 Mon Sep 17 00:00:00 2001 From: yixing1992 <34238811+yixing1992@users.noreply.github.com> Date: Mon, 16 Jun 2025 14:34:28 +0800 Subject: [PATCH] Update README.md for Huawei Ascend NPU support modes --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 7ac034c..e94a77f 100644 --- a/README.md +++ b/README.md @@ -235,7 +235,7 @@ DeepSeek-V3 can be deployed locally using the following hardware and open-source 5. **vLLM**: Support DeepSeek-V3 model with FP8 and BF16 modes for tensor parallelism and pipeline parallelism. 6. **LightLLM**: Supports efficient single-node or multi-node deployment for FP8 and BF16. 7. **AMD GPU**: Enables running the DeepSeek-V3 model on AMD GPUs via SGLang in both BF16 and FP8 modes. -8. **Huawei Ascend NPU**: Supports running DeepSeek-V3 on Huawei Ascend devices. +8. **Huawei Ascend NPU**: Supports running DeepSeek-V3 on Huawei Ascend devices in both INT8 and BF16. Since FP8 training is natively adopted in our framework, we only provide FP8 weights. If you require BF16 weights for experimentation, you can use the provided conversion script to perform the transformation.