finalf0 commited on
Commit
f3797ba
1 Parent(s): 81fdc3b

add kwargs for processor

Browse files
image_processing_minicpmv.py CHANGED
@@ -359,6 +359,7 @@ class MiniCPMVImageProcessor(BaseImageProcessor):
359
  do_pad: Optional[bool] = True, # TODO: add pad for MiniCPM-Llama3-V-2_5
360
  max_slice_nums: int = None,
361
  return_tensors: Optional[Union[str, TensorType]] = None,
 
362
  ) -> MiniCPMVBatchFeature:
363
  if isinstance(images, Image.Image):
364
  images_list = [[images]]
 
359
  do_pad: Optional[bool] = True, # TODO: add pad for MiniCPM-Llama3-V-2_5
360
  max_slice_nums: int = None,
361
  return_tensors: Optional[Union[str, TensorType]] = None,
362
+ **kwargs
363
  ) -> MiniCPMVBatchFeature:
364
  if isinstance(images, Image.Image):
365
  images_list = [[images]]
processing_minicpmv.py CHANGED
@@ -59,11 +59,12 @@ class MiniCPMVProcessor(ProcessorMixin):
59
  max_slice_nums: int = None,
60
  use_image_id: bool = None,
61
  return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
 
62
  ) -> MiniCPMVBatchFeature:
63
 
64
  if images is not None:
65
  image_inputs = self.image_processor(images, do_pad=do_pad, max_slice_nums=max_slice_nums, return_tensors=return_tensors)
66
- return self._convert_images_texts_to_inputs(image_inputs, text, max_slice_nums=max_slice_nums, use_image_id=use_image_id, max_length=max_length)
67
 
68
  # Copied from transformers.models.clip.processing_clip.CLIPProcessor.batch_decode with CLIP->Llama
69
  def batch_decode(self, *args, **kwargs):
@@ -133,10 +134,11 @@ class MiniCPMVProcessor(ProcessorMixin):
133
  max_length=None,
134
  max_slice_nums=None,
135
  use_image_id=None,
136
- return_tensors=None
 
137
  ):
138
  if images is None or not len(images):
139
- model_inputs = self.tokenizer(texts, return_tensors=return_tensors, truncation=truncation, max_length=max_length)
140
  return MiniCPMVBatchFeature(data={**model_inputs})
141
 
142
  pattern = "(<image>./</image>)"
 
59
  max_slice_nums: int = None,
60
  use_image_id: bool = None,
61
  return_tensors: Optional[Union[str, TensorType]] = TensorType.PYTORCH,
62
+ **kwargs
63
  ) -> MiniCPMVBatchFeature:
64
 
65
  if images is not None:
66
  image_inputs = self.image_processor(images, do_pad=do_pad, max_slice_nums=max_slice_nums, return_tensors=return_tensors)
67
+ return self._convert_images_texts_to_inputs(image_inputs, text, max_slice_nums=max_slice_nums, use_image_id=use_image_id, max_length=max_length, **kwargs)
68
 
69
  # Copied from transformers.models.clip.processing_clip.CLIPProcessor.batch_decode with CLIP->Llama
70
  def batch_decode(self, *args, **kwargs):
 
134
  max_length=None,
135
  max_slice_nums=None,
136
  use_image_id=None,
137
+ return_tensors=None,
138
+ **kwargs
139
  ):
140
  if images is None or not len(images):
141
+ model_inputs = self.tokenizer(texts, return_tensors=return_tensors, truncation=truncation, max_length=max_length, **kwargs)
142
  return MiniCPMVBatchFeature(data={**model_inputs})
143
 
144
  pattern = "(<image>./</image>)"