from transformers import PretrainedConfig,LlamaConfig from typing import List class InfMLLMChatConfig(LlamaConfig): def __init__( self, image_size=336, vision_tower="openai/clip-vit-large-patch14-336", **kwargs ): self.image_size = image_size self.vision_tower = vision_tower super().__init__(**kwargs)