add kwargs for processor
Browse files- image_processing_minicpmv.py +1 -0
- processing_minicpmv.py +5 -3
image_processing_minicpmv.py
CHANGED
@@ -359,6 +359,7 @@ class MiniCPMVImageProcessor(BaseImageProcessor):
|
|
359 |
do_pad: Optional[bool] = True, # TODO: add pad for MiniCPM-Llama3-V-2_5
|
360 |
max_slice_nums: int = None,
|
361 |
return_tensors: Optional[Union[str, TensorType]] = None,
|
|
|
362 |
) -> MiniCPMVBatchFeature:
|
363 |
if isinstance(images, Image.Image):
|
364 |
images_list = [[images]]
|
|
|
359 |
do_pad: Optional[bool] = True, # TODO: add pad for MiniCPM-Llama3-V-2_5
|
360 |
max_slice_nums: int = None,
|
361 |
return_tensors: Optional[Union[str, TensorType]] = None,
|
362 |
+
**kwargs
|
363 |
) -> MiniCPMVBatchFeature:
|
364 |
if isinstance(images, Image.Image):
|
365 |
images_list = [[images]]
|
processing_minicpmv.py
CHANGED
@@ -59,11 +59,12 @@ class MiniCPMVProcessor(ProcessorMixin):
|
|
59 |
max_slice_nums: int = None,
|
60 |
use_image_id: bool = None,
|
61 |
return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
|
|
|
62 |
) -> MiniCPMVBatchFeature:
|
63 |
|
64 |
if images is not None:
|
65 |
image_inputs = self.image_processor(images, do_pad=do_pad, max_slice_nums=max_slice_nums, return_tensors=return_tensors)
|
66 |
-
return self._convert_images_texts_to_inputs(image_inputs, text, max_slice_nums=max_slice_nums, use_image_id=use_image_id, max_length=max_length)
|
67 |
|
68 |
# Copied from transformers.models.clip.processing_clip.CLIPProcessor.batch_decode with CLIP->Llama
|
69 |
def batch_decode(self, *args, **kwargs):
|
@@ -133,10 +134,11 @@ class MiniCPMVProcessor(ProcessorMixin):
|
|
133 |
max_length=None,
|
134 |
max_slice_nums=None,
|
135 |
use_image_id=None,
|
136 |
-
return_tensors=None
|
|
|
137 |
):
|
138 |
if images is None or not len(images):
|
139 |
-
model_inputs = self.tokenizer(texts, return_tensors=return_tensors, truncation=truncation, max_length=max_length)
|
140 |
return MiniCPMVBatchFeature(data={**model_inputs})
|
141 |
|
142 |
pattern = "(<image>./</image>)"
|
|
|
59 |
max_slice_nums: int = None,
|
60 |
use_image_id: bool = None,
|
61 |
return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
|
62 |
+
**kwargs
|
63 |
) -> MiniCPMVBatchFeature:
|
64 |
|
65 |
if images is not None:
|
66 |
image_inputs = self.image_processor(images, do_pad=do_pad, max_slice_nums=max_slice_nums, return_tensors=return_tensors)
|
67 |
+
return self._convert_images_texts_to_inputs(image_inputs, text, max_slice_nums=max_slice_nums, use_image_id=use_image_id, max_length=max_length, **kwargs)
|
68 |
|
69 |
# Copied from transformers.models.clip.processing_clip.CLIPProcessor.batch_decode with CLIP->Llama
|
70 |
def batch_decode(self, *args, **kwargs):
|
|
|
134 |
max_length=None,
|
135 |
max_slice_nums=None,
|
136 |
use_image_id=None,
|
137 |
+
return_tensors=None,
|
138 |
+
**kwargs
|
139 |
):
|
140 |
if images is None or not len(images):
|
141 |
+
model_inputs = self.tokenizer(texts, return_tensors=return_tensors, truncation=truncation, max_length=max_length, **kwargs)
|
142 |
return MiniCPMVBatchFeature(data={**model_inputs})
|
143 |
|
144 |
pattern = "(<image>./</image>)"
|