diff --git a/README.md b/README.md
index 8e2320e..e20e49c 100644
--- a/README.md
+++ b/README.md
@@ -153,8 +153,8 @@ pip install -r requirements.txt -e .
 import torch
 from transformers import AutoModelForCausalLM
 
-from deepseek_vlm.models import VLChatProcessor, MultiModalityCausalLM
-from deepseek_vlm.utils.io import load_pil_images
+from deepseek_vl.models import VLChatProcessor, MultiModalityCausalLM
+from deepseek_vl.utils.io import load_pil_images
 
 
 # specify the path to the model
diff --git a/cli_chat.py b/cli_chat.py
index 6a75a1b..c639542 100644
--- a/cli_chat.py
+++ b/cli_chat.py
@@ -148,7 +148,8 @@ def chat(args, tokenizer, vl_chat_processor, vl_gpt, generation_config):
 
                 sys.stdout.write("\n")
                 sys.stdout.flush()
-                conv.messages[-1][-1] = answer
+                conv.update_last_message(answer)
+                # conv.messages[-1][-1] = answer
 
 
 def main(args):
diff --git a/deepseek_vl/models/image_processing_vlm.py b/deepseek_vl/models/image_processing_vlm.py
index 984117e..42aa7c4 100644
--- a/deepseek_vl/models/image_processing_vlm.py
+++ b/deepseek_vl/models/image_processing_vlm.py
@@ -150,7 +150,6 @@ class VLMImageProcessor(BaseImageProcessor):
         return [3, self.image_size, self.image_size]
 
 
-# AutoConfig.register("deepseek_vlm", VLMImageProcessorConfig)
 AutoImageProcessor.register(VLMImageProcessorConfig, VLMImageProcessor)