| from transformers import PretrainedConfig,LlamaConfig | |
| from typing import List | |
| class InfMLLMChatConfig(LlamaConfig): | |
| def __init__( | |
| self, | |
| image_size=336, | |
| vision_tower="openai/clip-vit-large-patch14-336", | |
| **kwargs | |
| ): | |
| self.image_size = image_size | |
| self.vision_tower = vision_tower | |
| super().__init__(**kwargs) | |