Commit
·
86744eb
1
Parent(s):
c89a453
update
Browse files- app.py +2 -2
- models/conversation.py +33 -1
- models/mllava/__init__.py +1 -0
- models/mllava/modeling_llava.py +8 -3
- models/mllava/processing_llava.py +124 -10
- models/mllava/utils.py +99 -35
app.py
CHANGED
|
@@ -5,8 +5,8 @@ import time
|
|
| 5 |
from PIL import Image
|
| 6 |
from models.mllava import MLlavaProcessor, LlavaForConditionalGeneration, chat_mllava, MLlavaForConditionalGeneration
|
| 7 |
from typing import List
|
| 8 |
-
processor = MLlavaProcessor.from_pretrained("TIGER-Lab/Mantis-
|
| 9 |
-
model = LlavaForConditionalGeneration.from_pretrained("TIGER-Lab/Mantis-
|
| 10 |
|
| 11 |
@spaces.GPU
|
| 12 |
def generate(text:str, images:List[Image.Image], history: List[dict], **kwargs):
|
|
|
|
| 5 |
from PIL import Image
|
| 6 |
from models.mllava import MLlavaProcessor, LlavaForConditionalGeneration, chat_mllava, MLlavaForConditionalGeneration
|
| 7 |
from typing import List
|
| 8 |
+
processor = MLlavaProcessor.from_pretrained("TIGER-Lab/Mantis-8B-siglip-llama3")
|
| 9 |
+
model = LlavaForConditionalGeneration.from_pretrained("TIGER-Lab/Mantis-8B-siglip-llama3")
|
| 10 |
|
| 11 |
@spaces.GPU
|
| 12 |
def generate(text:str, images:List[Image.Image], history: List[dict], **kwargs):
|
models/conversation.py
CHANGED
|
@@ -10,6 +10,7 @@ class SeparatorStyle(Enum):
|
|
| 10 |
MPT = auto()
|
| 11 |
PLAIN = auto()
|
| 12 |
LLAMA_2 = auto()
|
|
|
|
| 13 |
MFuyu = auto()
|
| 14 |
|
| 15 |
|
|
@@ -30,6 +31,7 @@ class Conversation:
|
|
| 30 |
def get_prompt(self):
|
| 31 |
messages = self.messages
|
| 32 |
if len(messages) > 0 and type(messages[0][1]) is tuple:
|
|
|
|
| 33 |
messages = self.messages.copy()
|
| 34 |
init_role, init_msg = messages[0].copy()
|
| 35 |
init_msg = init_msg[0].replace("<image>", "").strip()
|
|
@@ -39,7 +41,6 @@ class Conversation:
|
|
| 39 |
messages.insert(1, (self.roles[1], "Received."))
|
| 40 |
else:
|
| 41 |
messages[0] = (init_role, "<image>" + init_msg)
|
| 42 |
-
|
| 43 |
if self.sep_style == SeparatorStyle.SINGLE:
|
| 44 |
ret = self.system + self.sep
|
| 45 |
for role, message in messages:
|
|
@@ -89,6 +90,15 @@ class Conversation:
|
|
| 89 |
else:
|
| 90 |
ret += ""
|
| 91 |
ret = ret.lstrip(self.sep)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
elif self.sep_style == SeparatorStyle.MFuyu:
|
| 93 |
seps = [self.sep, self.sep2]
|
| 94 |
ret = self.system + "\n"
|
|
@@ -393,6 +403,25 @@ conv_mllava_v1_mmtag = Conversation(
|
|
| 393 |
version="v1_mmtag",
|
| 394 |
)
|
| 395 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 396 |
|
| 397 |
default_conversation = conv_mfuyu_v1
|
| 398 |
conv_templates = {
|
|
@@ -409,6 +438,9 @@ conv_templates = {
|
|
| 409 |
"llava_v1": conv_llava_v1,
|
| 410 |
"v1_mmtag": conv_llava_v1_mmtag,
|
| 411 |
"llava_llama_2": conv_llava_llama_2,
|
|
|
|
|
|
|
|
|
|
| 412 |
|
| 413 |
"mpt": conv_mpt,
|
| 414 |
}
|
|
|
|
| 10 |
MPT = auto()
|
| 11 |
PLAIN = auto()
|
| 12 |
LLAMA_2 = auto()
|
| 13 |
+
LLAMA_3 = auto()
|
| 14 |
MFuyu = auto()
|
| 15 |
|
| 16 |
|
|
|
|
| 31 |
def get_prompt(self):
|
| 32 |
messages = self.messages
|
| 33 |
if len(messages) > 0 and type(messages[0][1]) is tuple:
|
| 34 |
+
|
| 35 |
messages = self.messages.copy()
|
| 36 |
init_role, init_msg = messages[0].copy()
|
| 37 |
init_msg = init_msg[0].replace("<image>", "").strip()
|
|
|
|
| 41 |
messages.insert(1, (self.roles[1], "Received."))
|
| 42 |
else:
|
| 43 |
messages[0] = (init_role, "<image>" + init_msg)
|
|
|
|
| 44 |
if self.sep_style == SeparatorStyle.SINGLE:
|
| 45 |
ret = self.system + self.sep
|
| 46 |
for role, message in messages:
|
|
|
|
| 90 |
else:
|
| 91 |
ret += ""
|
| 92 |
ret = ret.lstrip(self.sep)
|
| 93 |
+
elif self.sep_style == SeparatorStyle.LLAMA_3:
|
| 94 |
+
ret = self.system + self.sep
|
| 95 |
+
for role, message in messages:
|
| 96 |
+
if message:
|
| 97 |
+
if type(message) is tuple:
|
| 98 |
+
message, _, _ = message
|
| 99 |
+
ret += f"<|start_header_id|>{role}<|end_header_id|>\n\n" + message + self.sep
|
| 100 |
+
else:
|
| 101 |
+
ret += f"<|start_header_id|>{role}<|end_header_id|>\n\n"
|
| 102 |
elif self.sep_style == SeparatorStyle.MFuyu:
|
| 103 |
seps = [self.sep, self.sep2]
|
| 104 |
ret = self.system + "\n"
|
|
|
|
| 403 |
version="v1_mmtag",
|
| 404 |
)
|
| 405 |
|
| 406 |
+
conv_mllava_v1 = Conversation(
|
| 407 |
+
system="A chat between a curious human and an artificial intelligence assistant. "
|
| 408 |
+
"The assistant gives helpful, detailed, and polite answers to the human's questions.",
|
| 409 |
+
roles=("USER", "ASSISTANT"),
|
| 410 |
+
version="v1",
|
| 411 |
+
messages=(),
|
| 412 |
+
offset=0,
|
| 413 |
+
sep_style=SeparatorStyle.SINGLE,
|
| 414 |
+
sep="</s>",
|
| 415 |
+
)
|
| 416 |
+
|
| 417 |
+
conv_llama_3 = Conversation(
|
| 418 |
+
system="<|start_header_id|>system<|end_header_id|>\n\nYou are a pirate chatbot who always responds in pirate speak!",
|
| 419 |
+
roles=("user", "assistant"),
|
| 420 |
+
messages=(),
|
| 421 |
+
offset=0,
|
| 422 |
+
sep_style=SeparatorStyle.LLAMA_3,
|
| 423 |
+
sep="<|eot_id|>",
|
| 424 |
+
)
|
| 425 |
|
| 426 |
default_conversation = conv_mfuyu_v1
|
| 427 |
conv_templates = {
|
|
|
|
| 438 |
"llava_v1": conv_llava_v1,
|
| 439 |
"v1_mmtag": conv_llava_v1_mmtag,
|
| 440 |
"llava_llama_2": conv_llava_llama_2,
|
| 441 |
+
"llama_3": conv_llama_3,
|
| 442 |
+
"mllava_v1": conv_mllava_v1,
|
| 443 |
+
"mllava_v1_mmtag": conv_mllava_v1_mmtag,
|
| 444 |
|
| 445 |
"mpt": conv_mpt,
|
| 446 |
}
|
models/mllava/__init__.py
CHANGED
|
@@ -1,3 +1,4 @@
|
|
| 1 |
from .modeling_llava import LlavaForConditionalGeneration, MLlavaForConditionalGeneration
|
| 2 |
from .processing_llava import MLlavaProcessor
|
|
|
|
| 3 |
from .utils import chat_mllava
|
|
|
|
| 1 |
from .modeling_llava import LlavaForConditionalGeneration, MLlavaForConditionalGeneration
|
| 2 |
from .processing_llava import MLlavaProcessor
|
| 3 |
+
from .configuration_llava import LlavaConfig
|
| 4 |
from .utils import chat_mllava
|
models/mllava/modeling_llava.py
CHANGED
|
@@ -249,15 +249,15 @@ LLAVA_INPUTS_DOCSTRING = r"""
|
|
| 249 |
LLAVA_START_DOCSTRING,
|
| 250 |
)
|
| 251 |
class LlavaForConditionalGeneration(LlavaPreTrainedModel):
|
| 252 |
-
def __init__(self, config: LlavaConfig):
|
| 253 |
super().__init__(config)
|
| 254 |
-
self.vision_tower = AutoModel.from_config(config.vision_config)
|
| 255 |
|
| 256 |
self.multi_modal_projector = LlavaMultiModalProjector(config)
|
| 257 |
self.vocab_size = config.vocab_size
|
| 258 |
self.language_model = AutoModelForCausalLM.from_config(
|
| 259 |
config.text_config, attn_implementation=config._attn_implementation
|
| 260 |
-
)
|
| 261 |
self.pad_token_id = self.config.pad_token_id if self.config.pad_token_id is not None else -1
|
| 262 |
self.post_init()
|
| 263 |
|
|
@@ -428,6 +428,11 @@ class LlavaForConditionalGeneration(LlavaPreTrainedModel):
|
|
| 428 |
|
| 429 |
# 2. Merge text and images
|
| 430 |
if pixel_values is not None and input_ids.shape[1] != 1:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 431 |
image_outputs = self.vision_tower(pixel_values, output_hidden_states=True)
|
| 432 |
# this is not memory efficient at all (output_hidden_states=True) will save all the hidden stated.
|
| 433 |
selected_image_feature = image_outputs.hidden_states[vision_feature_layer]
|
|
|
|
| 249 |
LLAVA_START_DOCSTRING,
|
| 250 |
)
|
| 251 |
class LlavaForConditionalGeneration(LlavaPreTrainedModel):
|
| 252 |
+
def __init__(self, config: LlavaConfig, vision_tower=None, language_model=None):
|
| 253 |
super().__init__(config)
|
| 254 |
+
self.vision_tower = AutoModel.from_config(config.vision_config) if vision_tower is None else vision_tower
|
| 255 |
|
| 256 |
self.multi_modal_projector = LlavaMultiModalProjector(config)
|
| 257 |
self.vocab_size = config.vocab_size
|
| 258 |
self.language_model = AutoModelForCausalLM.from_config(
|
| 259 |
config.text_config, attn_implementation=config._attn_implementation
|
| 260 |
+
) if language_model is None else language_model
|
| 261 |
self.pad_token_id = self.config.pad_token_id if self.config.pad_token_id is not None else -1
|
| 262 |
self.post_init()
|
| 263 |
|
|
|
|
| 428 |
|
| 429 |
# 2. Merge text and images
|
| 430 |
if pixel_values is not None and input_ids.shape[1] != 1:
|
| 431 |
+
if isinstance(pixel_values, list):
|
| 432 |
+
pixel_values = torch.cat([x for x in pixel_values if x is not None], dim=0)
|
| 433 |
+
# for siglip, need to transform the pixel_values to the right data type
|
| 434 |
+
if pixel_values.dtype != self.vision_tower.dtype:
|
| 435 |
+
pixel_values = pixel_values.type(self.vision_tower.dtype)
|
| 436 |
image_outputs = self.vision_tower(pixel_values, output_hidden_states=True)
|
| 437 |
# this is not memory efficient at all (output_hidden_states=True) will save all the hidden stated.
|
| 438 |
selected_image_feature = image_outputs.hidden_states[vision_feature_layer]
|
models/mllava/processing_llava.py
CHANGED
|
@@ -16,7 +16,8 @@
|
|
| 16 |
Processor class for Llava.
|
| 17 |
"""
|
| 18 |
|
| 19 |
-
|
|
|
|
| 20 |
from typing import List, Optional, Union, Dict
|
| 21 |
|
| 22 |
# from ...feature_extraction_utils import BatchFeature
|
|
@@ -30,6 +31,9 @@ from transformers.image_utils import ImageInput
|
|
| 30 |
from transformers.processing_utils import ProcessorMixin
|
| 31 |
from transformers.tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy
|
| 32 |
from transformers.utils import TensorType
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
from PIL import Image
|
| 35 |
import logging
|
|
@@ -52,8 +56,8 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 52 |
"""
|
| 53 |
|
| 54 |
attributes = ["image_processor", "tokenizer"]
|
| 55 |
-
image_processor_class = "CLIPImageProcessor"
|
| 56 |
-
tokenizer_class = ("LlamaTokenizer", "LlamaTokenizerFast")
|
| 57 |
|
| 58 |
def __init__(self, image_processor=None, tokenizer=None):
|
| 59 |
super().__init__(image_processor, tokenizer)
|
|
@@ -109,7 +113,7 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 109 |
if i < num_images:
|
| 110 |
text[i] = t + "<image>"
|
| 111 |
text = "".join(text)
|
| 112 |
-
logger.warning("Number of <image> tokens exceeds number of images. Automatically removing extra tokens at the end of the text.")
|
| 113 |
# raise ValueError("Invalid input text. Number of <image> tokens exceeds number of images.")
|
| 114 |
texts = [text]
|
| 115 |
elif isinstance(text, list):
|
|
@@ -135,7 +139,7 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 135 |
if j < num_images:
|
| 136 |
t[j] = s + "<image>"
|
| 137 |
t = "".join(t)
|
| 138 |
-
logger.warning("Number of <image> tokens exceeds number of images. Automatically removing extra tokens at the end of the text.")
|
| 139 |
# raise ValueError("Invalid input text. Number of <image> tokens exceeds number of images.")
|
| 140 |
text[i] = t
|
| 141 |
texts = text
|
|
@@ -171,6 +175,7 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 171 |
truncation: Union[bool, str, TruncationStrategy] = None,
|
| 172 |
max_length=None,
|
| 173 |
return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
|
|
|
|
| 174 |
) -> BatchFeature:
|
| 175 |
"""
|
| 176 |
Main method to prepare for the model one or several sequences(s) and image(s). This method forwards the `text`
|
|
@@ -218,13 +223,14 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 218 |
`None`).
|
| 219 |
- **pixel_values** -- Pixel values to be fed to a model. Returned when `images` is not `None`.
|
| 220 |
"""
|
| 221 |
-
|
|
|
|
| 222 |
if images is not None:
|
| 223 |
pixel_values = self.image_processor(images, return_tensors=return_tensors)["pixel_values"] # [batch_size, num_channels, height, width], e.g. [1, 3, 336, 336]
|
| 224 |
else:
|
| 225 |
pixel_values = None
|
| 226 |
text_inputs = self.tokenizer(
|
| 227 |
-
|
| 228 |
)
|
| 229 |
# text_inputs:
|
| 230 |
# 1. input_ids: [batch_size, sequence_length], e.g. [1, 6]
|
|
@@ -259,9 +265,117 @@ class MLlavaProcessor(ProcessorMixin):
|
|
| 259 |
results = {}
|
| 260 |
assert len(model_inputs) == 1, "This method only supports a single input, but get {} inputs".format(len(model_inputs))
|
| 261 |
for k in model_inputs[0].keys():
|
| 262 |
-
if
|
| 263 |
-
results[k] =
|
| 264 |
else:
|
| 265 |
-
results[k] =
|
| 266 |
return results
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 267 |
|
|
|
|
| 16 |
Processor class for Llava.
|
| 17 |
"""
|
| 18 |
|
| 19 |
+
import os
|
| 20 |
+
import json
|
| 21 |
from typing import List, Optional, Union, Dict
|
| 22 |
|
| 23 |
# from ...feature_extraction_utils import BatchFeature
|
|
|
|
| 31 |
from transformers.processing_utils import ProcessorMixin
|
| 32 |
from transformers.tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy
|
| 33 |
from transformers.utils import TensorType
|
| 34 |
+
from transformers.processing_utils import transformers_module
|
| 35 |
+
from transformers.utils.hub import is_remote_url, download_url, cached_file, is_offline_mode
|
| 36 |
+
from transformers.utils import IMAGE_PROCESSOR_NAME
|
| 37 |
|
| 38 |
from PIL import Image
|
| 39 |
import logging
|
|
|
|
| 56 |
"""
|
| 57 |
|
| 58 |
attributes = ["image_processor", "tokenizer"]
|
| 59 |
+
image_processor_class = ("CLIPImageProcessor", "SiglipImageProcessor")
|
| 60 |
+
tokenizer_class = ("LlamaTokenizer", "LlamaTokenizerFast", "PreTrainedTokenizerFast")
|
| 61 |
|
| 62 |
def __init__(self, image_processor=None, tokenizer=None):
|
| 63 |
super().__init__(image_processor, tokenizer)
|
|
|
|
| 113 |
if i < num_images:
|
| 114 |
text[i] = t + "<image>"
|
| 115 |
text = "".join(text)
|
| 116 |
+
logger.warning(f"Number of <image> tokens: {num_image_tokens} exceeds number of images: {num_images}. Automatically removing extra tokens at the end of the text.")
|
| 117 |
# raise ValueError("Invalid input text. Number of <image> tokens exceeds number of images.")
|
| 118 |
texts = [text]
|
| 119 |
elif isinstance(text, list):
|
|
|
|
| 139 |
if j < num_images:
|
| 140 |
t[j] = s + "<image>"
|
| 141 |
t = "".join(t)
|
| 142 |
+
logger.warning(f"Number of <image> tokens: {num_image_tokens} exceeds number of images: {num_images}. Automatically removing extra tokens at the end of the text.")
|
| 143 |
# raise ValueError("Invalid input text. Number of <image> tokens exceeds number of images.")
|
| 144 |
text[i] = t
|
| 145 |
texts = text
|
|
|
|
| 175 |
truncation: Union[bool, str, TruncationStrategy] = None,
|
| 176 |
max_length=None,
|
| 177 |
return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
|
| 178 |
+
add_image_ids: bool = True,
|
| 179 |
) -> BatchFeature:
|
| 180 |
"""
|
| 181 |
Main method to prepare for the model one or several sequences(s) and image(s). This method forwards the `text`
|
|
|
|
| 223 |
`None`).
|
| 224 |
- **pixel_values** -- Pixel values to be fed to a model. Returned when `images` is not `None`.
|
| 225 |
"""
|
| 226 |
+
if add_image_ids:
|
| 227 |
+
text, images = self.preprocess_interleaved_images_and_text(text, images)
|
| 228 |
if images is not None:
|
| 229 |
pixel_values = self.image_processor(images, return_tensors=return_tensors)["pixel_values"] # [batch_size, num_channels, height, width], e.g. [1, 3, 336, 336]
|
| 230 |
else:
|
| 231 |
pixel_values = None
|
| 232 |
text_inputs = self.tokenizer(
|
| 233 |
+
text, return_tensors=return_tensors, padding=padding, truncation=truncation, max_length=max_length
|
| 234 |
)
|
| 235 |
# text_inputs:
|
| 236 |
# 1. input_ids: [batch_size, sequence_length], e.g. [1, 6]
|
|
|
|
| 265 |
results = {}
|
| 266 |
assert len(model_inputs) == 1, "This method only supports a single input, but get {} inputs".format(len(model_inputs))
|
| 267 |
for k in model_inputs[0].keys():
|
| 268 |
+
if k == "pixel_values":
|
| 269 |
+
results[k] = [inputs[k] if inputs[k] is not None else None for inputs in model_inputs]
|
| 270 |
else:
|
| 271 |
+
results[k] = torch.cat([inputs[k] for inputs in model_inputs], dim=0)
|
| 272 |
return results
|
| 273 |
+
|
| 274 |
+
@classmethod
|
| 275 |
+
def _get_arguments_from_pretrained(cls, pretrained_model_name_or_path, **kwargs):
|
| 276 |
+
args = []
|
| 277 |
+
|
| 278 |
+
cache_dir = kwargs.pop("cache_dir", None)
|
| 279 |
+
force_download = kwargs.pop("force_download", False)
|
| 280 |
+
resume_download = kwargs.pop("resume_download", False)
|
| 281 |
+
proxies = kwargs.pop("proxies", None)
|
| 282 |
+
token = kwargs.pop("token", None)
|
| 283 |
+
local_files_only = kwargs.pop("local_files_only", False)
|
| 284 |
+
revision = kwargs.pop("revision", None)
|
| 285 |
+
subfolder = kwargs.pop("subfolder", "")
|
| 286 |
+
|
| 287 |
+
from_pipeline = kwargs.pop("_from_pipeline", None)
|
| 288 |
+
from_auto_class = kwargs.pop("_from_auto", False)
|
| 289 |
+
|
| 290 |
+
user_agent = {"file_type": "processor", "from_auto_class": from_auto_class}
|
| 291 |
+
if from_pipeline is not None:
|
| 292 |
+
user_agent["using_pipeline"] = from_pipeline
|
| 293 |
+
|
| 294 |
+
if is_offline_mode() and not local_files_only:
|
| 295 |
+
logger.info("Offline mode: forcing local_files_only=True")
|
| 296 |
+
local_files_only = True
|
| 297 |
+
|
| 298 |
+
pretrained_model_name_or_path = str(pretrained_model_name_or_path)
|
| 299 |
+
is_local = os.path.isdir(pretrained_model_name_or_path)
|
| 300 |
+
if os.path.isdir(pretrained_model_name_or_path):
|
| 301 |
+
processor_file = os.path.join(pretrained_model_name_or_path, IMAGE_PROCESSOR_NAME)
|
| 302 |
+
if os.path.isfile(pretrained_model_name_or_path):
|
| 303 |
+
resolved_processor_file = pretrained_model_name_or_path
|
| 304 |
+
is_local = True
|
| 305 |
+
elif is_remote_url(pretrained_model_name_or_path):
|
| 306 |
+
processor_file = pretrained_model_name_or_path
|
| 307 |
+
resolved_processor_file = download_url(pretrained_model_name_or_path)
|
| 308 |
+
else:
|
| 309 |
+
processor_file = IMAGE_PROCESSOR_NAME
|
| 310 |
+
try:
|
| 311 |
+
# Load from local folder or from cache or download from model Hub and cache
|
| 312 |
+
resolved_processor_file = cached_file(
|
| 313 |
+
pretrained_model_name_or_path,
|
| 314 |
+
processor_file,
|
| 315 |
+
cache_dir=cache_dir,
|
| 316 |
+
force_download=force_download,
|
| 317 |
+
proxies=proxies,
|
| 318 |
+
resume_download=resume_download,
|
| 319 |
+
local_files_only=local_files_only,
|
| 320 |
+
token=token,
|
| 321 |
+
user_agent=user_agent,
|
| 322 |
+
revision=revision,
|
| 323 |
+
subfolder=subfolder,
|
| 324 |
+
_raise_exceptions_for_missing_entries=True,
|
| 325 |
+
)
|
| 326 |
+
except EnvironmentError:
|
| 327 |
+
# Raise any environment error raise by `cached_file`. It will have a helpful error message adapted to
|
| 328 |
+
# the original exception.
|
| 329 |
+
raise
|
| 330 |
+
except Exception:
|
| 331 |
+
# For any other exception, we throw a generic error.
|
| 332 |
+
raise EnvironmentError(
|
| 333 |
+
f"Can't load processor for '{pretrained_model_name_or_path}'. If you were trying to load"
|
| 334 |
+
" it from 'https://huggingface.co/models', make sure you don't have a local directory with the"
|
| 335 |
+
f" same name. Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a"
|
| 336 |
+
f" directory containing a {IMAGE_PROCESSOR_NAME} file"
|
| 337 |
+
)
|
| 338 |
+
|
| 339 |
+
# Existing processors on the Hub created before #27761 being merged don't have `processor_config.json` (if not
|
| 340 |
+
# updated afterward), and we need to keep `from_pretrained` work. So here it fallbacks to the empty dict.
|
| 341 |
+
# (`cached_file` called using `_raise_exceptions_for_missing_entries=False` to avoid exception)
|
| 342 |
+
# However, for models added in the future, we won't get the expected error if this file is missing.
|
| 343 |
+
if resolved_processor_file is None:
|
| 344 |
+
image_processor_dict = {}
|
| 345 |
+
|
| 346 |
+
try:
|
| 347 |
+
# Load processor dict
|
| 348 |
+
with open(resolved_processor_file, "r", encoding="utf-8") as reader:
|
| 349 |
+
text = reader.read()
|
| 350 |
+
image_processor_dict = json.loads(text)
|
| 351 |
+
|
| 352 |
+
except json.JSONDecodeError:
|
| 353 |
+
raise EnvironmentError(
|
| 354 |
+
f"It looks like the config file at '{resolved_processor_file}' is not a valid JSON file."
|
| 355 |
+
)
|
| 356 |
+
|
| 357 |
+
for attribute_name in cls.attributes:
|
| 358 |
+
class_name = getattr(cls, f"{attribute_name}_class")
|
| 359 |
+
if isinstance(class_name, tuple):
|
| 360 |
+
if attribute_name == "tokenizer":
|
| 361 |
+
classes = tuple(getattr(transformers_module, n) if n is not None else None for n in class_name)
|
| 362 |
+
use_fast = kwargs.get("use_fast", True)
|
| 363 |
+
if use_fast and classes[1] is not None:
|
| 364 |
+
attribute_class = classes[1]
|
| 365 |
+
else:
|
| 366 |
+
attribute_class = classes[0]
|
| 367 |
+
elif attribute_name == "image_processor":
|
| 368 |
+
image_processor_type = image_processor_dict.get("image_processor_type", None)
|
| 369 |
+
if image_processor_type is not None:
|
| 370 |
+
assert image_processor_type in class_name, f"Invalid image processor type: {image_processor_type}"
|
| 371 |
+
attribute_class = getattr(transformers_module, image_processor_type)
|
| 372 |
+
else:
|
| 373 |
+
attribute_class = getattr(transformers_module, class_name[0])
|
| 374 |
+
else:
|
| 375 |
+
raise ValueError(f"Invalid attribute name: {attribute_name}")
|
| 376 |
+
else:
|
| 377 |
+
attribute_class = getattr(transformers_module, class_name)
|
| 378 |
+
|
| 379 |
+
args.append(attribute_class.from_pretrained(pretrained_model_name_or_path, **kwargs))
|
| 380 |
+
return args
|
| 381 |
|
models/mllava/utils.py
CHANGED
|
@@ -2,7 +2,9 @@ import PIL
|
|
| 2 |
import torch
|
| 3 |
from .modeling_llava import LlavaForConditionalGeneration
|
| 4 |
from .processing_llava import MLlavaProcessor
|
| 5 |
-
from ..conversation import conv_mllava_v1_mmtag as default_conv
|
|
|
|
|
|
|
| 6 |
from typing import List, Tuple, Union, Tuple
|
| 7 |
|
| 8 |
def chat_mllava(
|
|
@@ -12,7 +14,6 @@ def chat_mllava(
|
|
| 12 |
processor:MLlavaProcessor,
|
| 13 |
max_input_length:int=None,
|
| 14 |
history:List[dict]=None,
|
| 15 |
-
stream:bool=False,
|
| 16 |
**kwargs) -> Tuple[str, List[dict]]:
|
| 17 |
"""
|
| 18 |
Chat with the Mllava model
|
|
@@ -29,7 +30,17 @@ def chat_mllava(
|
|
| 29 |
|
| 30 |
|
| 31 |
"""
|
| 32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
conv.messages = []
|
| 34 |
if history is not None:
|
| 35 |
for message in history:
|
|
@@ -38,17 +49,8 @@ def chat_mllava(
|
|
| 38 |
conv.append_message(message["role"], message["text"])
|
| 39 |
else:
|
| 40 |
history = []
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
conv.append_message(conv.roles[0], text)
|
| 44 |
-
conv.append_message(conv.roles[1], "")
|
| 45 |
-
history.append({"role": conv.roles[0], "text": text})
|
| 46 |
-
history.append({"role": conv.roles[1], "text": ""})
|
| 47 |
-
else:
|
| 48 |
-
assert history, "The history should not be empty if the text is None"
|
| 49 |
-
assert history[-1]['role'] == conv.roles[1], "The last message in the history should be the assistant, an empty message"
|
| 50 |
-
assert history[-2]['text'], "The last user message in the history should not be empty"
|
| 51 |
-
assert history[-1]['text'] == "", "The last assistant message in the history should be empty"
|
| 52 |
|
| 53 |
prompt = conv.get_prompt()
|
| 54 |
if images:
|
|
@@ -57,27 +59,89 @@ def chat_mllava(
|
|
| 57 |
images[i] = PIL.Image.open(images[i])
|
| 58 |
|
| 59 |
inputs = processor(images=images, text=prompt, return_tensors="pt", truncation=True, max_length=max_input_length)
|
| 60 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
|
| 62 |
-
if stream:
|
| 63 |
-
from transformers import TextIteratorStreamer
|
| 64 |
-
from threading import Thread
|
| 65 |
-
streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
|
| 66 |
-
kwargs["streamer"] = streamer
|
| 67 |
-
inputs.update(kwargs)
|
| 68 |
-
thread = Thread(target=model.generate, kwargs=inputs)
|
| 69 |
-
thread.start()
|
| 70 |
-
for _output in streamer:
|
| 71 |
-
history[-1]["text"] += _output
|
| 72 |
-
yield history[-1]["text"], history
|
| 73 |
-
else:
|
| 74 |
-
output_ids = model.generate(**inputs, **kwargs)
|
| 75 |
-
output_ids = output_ids[0]
|
| 76 |
-
|
| 77 |
-
# remove the input tokens
|
| 78 |
-
generated_ids = output_ids[inputs["input_ids"].shape[-1]:]
|
| 79 |
-
generated_text = processor.decode(generated_ids, skip_special_tokens=True)
|
| 80 |
|
| 81 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
|
| 83 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
import torch
|
| 3 |
from .modeling_llava import LlavaForConditionalGeneration
|
| 4 |
from .processing_llava import MLlavaProcessor
|
| 5 |
+
# from ..conversation import conv_mllava_v1_mmtag as default_conv
|
| 6 |
+
from ..conversation import conv_mllava_v1 as default_conv, conv_templates
|
| 7 |
+
|
| 8 |
from typing import List, Tuple, Union, Tuple
|
| 9 |
|
| 10 |
def chat_mllava(
|
|
|
|
| 14 |
processor:MLlavaProcessor,
|
| 15 |
max_input_length:int=None,
|
| 16 |
history:List[dict]=None,
|
|
|
|
| 17 |
**kwargs) -> Tuple[str, List[dict]]:
|
| 18 |
"""
|
| 19 |
Chat with the Mllava model
|
|
|
|
| 30 |
|
| 31 |
|
| 32 |
"""
|
| 33 |
+
if "llama-3" in model.language_model.name_or_path.lower():
|
| 34 |
+
conv = conv_templates['llama_3']
|
| 35 |
+
terminators = [
|
| 36 |
+
processor.tokenizer.eos_token_id,
|
| 37 |
+
processor.tokenizer.convert_tokens_to_ids("<|eot_id|>")
|
| 38 |
+
]
|
| 39 |
+
else:
|
| 40 |
+
conv = default_conv
|
| 41 |
+
terminators = None
|
| 42 |
+
kwargs["eos_token_id"] = terminators
|
| 43 |
+
conv = conv.copy()
|
| 44 |
conv.messages = []
|
| 45 |
if history is not None:
|
| 46 |
for message in history:
|
|
|
|
| 49 |
conv.append_message(message["role"], message["text"])
|
| 50 |
else:
|
| 51 |
history = []
|
| 52 |
+
conv.append_message(conv.roles[0], text)
|
| 53 |
+
conv.append_message(conv.roles[1], "")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
prompt = conv.get_prompt()
|
| 56 |
if images:
|
|
|
|
| 59 |
images[i] = PIL.Image.open(images[i])
|
| 60 |
|
| 61 |
inputs = processor(images=images, text=prompt, return_tensors="pt", truncation=True, max_length=max_input_length)
|
| 62 |
+
for k, v in inputs.items():
|
| 63 |
+
if v is not None:
|
| 64 |
+
if isinstance(v, torch.Tensor):
|
| 65 |
+
inputs[k] = v.to(model.device)
|
| 66 |
+
elif isinstance(v, list):
|
| 67 |
+
inputs[k] = [x.to(model.device) for x in v]
|
| 68 |
+
else:
|
| 69 |
+
raise ValueError(f"Invalid input type: {type(v)}")
|
| 70 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
|
| 72 |
+
output_ids = model.generate(**inputs, **kwargs)
|
| 73 |
+
output_ids = output_ids[0]
|
| 74 |
+
|
| 75 |
+
# remove the input tokens
|
| 76 |
+
generated_ids = output_ids[inputs["input_ids"].shape[-1]:]
|
| 77 |
+
generated_text = processor.decode(generated_ids, skip_special_tokens=True)
|
| 78 |
+
|
| 79 |
+
history.append({"role": conv.roles[0], "text": text})
|
| 80 |
+
history.append({"role": conv.roles[1], "text": generated_text})
|
| 81 |
+
|
| 82 |
+
return generated_text, history
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
def chat_mllava_stream(
|
| 86 |
+
text:str,
|
| 87 |
+
images: List[Union[PIL.Image.Image, str]],
|
| 88 |
+
model:LlavaForConditionalGeneration,
|
| 89 |
+
processor:MLlavaProcessor,
|
| 90 |
+
max_input_length:int=None,
|
| 91 |
+
history:List[dict]=None,
|
| 92 |
+
**kwargs) -> Tuple[str, List[dict]]:
|
| 93 |
+
"""
|
| 94 |
+
Chat with the Mllava model
|
| 95 |
+
Args:
|
| 96 |
+
text: str, the text to be sent to the model, where <image> will be the placeholder for the image
|
| 97 |
+
images: List[PIL.Image.Image], the images to be sent to the model, or None
|
| 98 |
+
model: LlavaForConditionalGeneration, the model to be used
|
| 99 |
+
processor: MLlavaProcessor, the processor to be used
|
| 100 |
+
max_input_length: int, the maximum input length
|
| 101 |
+
history: List[dict], list of messages in the conversation as history. Each message is a dictionary {"role": "ASSISTANT/USER", "text": "the message"}. If None, the conversation will start from scratch
|
| 102 |
+
kwargs: dict, the generation kwargs
|
| 103 |
+
Returns:
|
| 104 |
+
Tuple[str, List[dict]], the generated text and the history of the conversation
|
| 105 |
|
| 106 |
+
|
| 107 |
+
"""
|
| 108 |
+
conv = default_conv.copy()
|
| 109 |
+
conv.messages = []
|
| 110 |
+
if history is not None:
|
| 111 |
+
for message in history:
|
| 112 |
+
message["role"] = message["role"].upper()
|
| 113 |
+
assert message["role"] in conv.roles
|
| 114 |
+
conv.append_message(message["role"], message["text"])
|
| 115 |
+
else:
|
| 116 |
+
history = []
|
| 117 |
+
conv.append_message(conv.roles[0], text)
|
| 118 |
+
conv.append_message(conv.roles[1], "")
|
| 119 |
+
|
| 120 |
+
prompt = conv.get_prompt()
|
| 121 |
+
if images:
|
| 122 |
+
for i in range(len(images)):
|
| 123 |
+
if isinstance(images[i], str):
|
| 124 |
+
images[i] = PIL.Image.open(images[i])
|
| 125 |
+
|
| 126 |
+
inputs = processor(images=images, text=prompt, return_tensors="pt", truncation=True, max_length=max_input_length)
|
| 127 |
+
for k, v in inputs.items():
|
| 128 |
+
if v is not None:
|
| 129 |
+
if isinstance(v, torch.Tensor):
|
| 130 |
+
inputs[k] = v.to(model.device)
|
| 131 |
+
elif isinstance(v, list):
|
| 132 |
+
inputs[k] = [x.to(model.device) for x in v]
|
| 133 |
+
else:
|
| 134 |
+
raise ValueError(f"Invalid input type: {type(v)}")
|
| 135 |
+
|
| 136 |
+
from transformers import TextIteratorStreamer
|
| 137 |
+
from threading import Thread
|
| 138 |
+
streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
|
| 139 |
+
kwargs["streamer"] = streamer
|
| 140 |
+
inputs.update(kwargs)
|
| 141 |
+
thread = Thread(target=model.generate, kwargs=inputs)
|
| 142 |
+
thread.start()
|
| 143 |
+
history.append({"role": conv.roles[0], "text": text})
|
| 144 |
+
history.append({"role": conv.roles[1], "text": ""})
|
| 145 |
+
for _output in streamer:
|
| 146 |
+
history[-1]["text"] += _output
|
| 147 |
+
yield history[-1]["text"], history
|