diff --git a/README.md b/README.md index 0a8c9f0..4025032 100644 --- a/README.md +++ b/README.md @@ -215,7 +215,7 @@ print(result) import torch from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig -model_name = "deepseek-ai/DeepSeek-V2-Chat-RL" +model_name = "deepseek-ai/DeepSeek-V2-Chat" tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) # `max_memory` should be set based on your devices max_memory = {i: "75GB" for i in range(8)}