|
"""Processor class for VideoLLaMA3.""" |
|
|
|
import copy |
|
import importlib.util |
|
import os |
|
import os.path as osp |
|
import warnings |
|
from collections import defaultdict |
|
from typing import Any, List, Union, Dict, Optional, Tuple, TypedDict |
|
|
|
import cv2 |
|
import ffmpeg |
|
import imageio |
|
import json |
|
import numpy as np |
|
import torch |
|
import transformers |
|
from decord import VideoReader, cpu |
|
from PIL import Image |
|
from transformers.feature_extraction_utils import BatchFeature |
|
from transformers.image_utils import ImageInput |
|
from transformers.processing_utils import ProcessingKwargs, ProcessorMixin, Unpack |
|
from transformers.tokenization_utils_base import PreTokenizedInput, TextInput |
|
|
|
try: |
|
from . import image_processing_videollama3 |
|
from .image_processing_videollama3 import ( |
|
is_valid_image, is_valid_video, |
|
) |
|
except ModuleNotFoundError: |
|
spec = importlib.util.spec_from_file_location( |
|
"image_processing_videollama3", |
|
osp.join(osp.dirname(__file__), "image_processing_videollama3.py"), |
|
) |
|
image_processing_videollama3 = importlib.util.module_from_spec(spec) |
|
spec.loader.exec_module(image_processing_videollama3) |
|
is_valid_image = getattr(image_processing_videollama3, "is_valid_image") |
|
is_valid_video = getattr(image_processing_videollama3, "is_valid_video") |
|
|
|
|
|
DEFAULT_IMAGE_TOKEN = "<image>" |
|
IGNORE_INDEX = -100 |
|
|
|
|
|
Conversation = List[Dict[str, Any]] |
|
SingleImage = Union[Image.Image, np.ndarray, torch.Tensor] |
|
SingleVideo = Union[List[SingleImage], np.ndarray, torch.Tensor] |
|
BatchedImage = List[Union[SingleImage, SingleVideo]] |
|
BatchedNamedImage = List[Tuple[str, Union[SingleImage, SingleVideo]]] |
|
|
|
|
|
def _custom_import(class_name: str): |
|
try: |
|
attribute_class = getattr(transformers, class_name) |
|
except AttributeError: |
|
attribute_class = getattr(image_processing_videollama3, class_name) |
|
return attribute_class |
|
|
|
|
|
def is_named_image(image) -> bool: |
|
return isinstance(image, (list, tuple)) and \ |
|
len(image) == 2 and \ |
|
isinstance(image[0], str) and \ |
|
image[0] in ["image", "video"] and \ |
|
(is_valid_image(image[1]) or is_valid_video(image[1])) |
|
|
|
|
|
def make_batched_images(images) -> List[List[ImageInput]]: |
|
if isinstance(images, (list, tuple)) and all(is_named_image(image) for image in images): |
|
|
|
return [image[0] for image in images], [image[1] for image in images] |
|
elif isinstance(images, (list, tuple)) and all(is_valid_image(image) or is_valid_video(image) for image in images): |
|
|
|
batch = [] |
|
for image in images: |
|
if is_valid_video(image): |
|
batch.append(("video", image)) |
|
elif is_valid_image(image): |
|
batch.append(("image", image)) |
|
else: |
|
raise ValueError(f"Could not make batched images from {images}") |
|
return [x[0] for x in batch], [x[1] for x in batch] |
|
elif is_named_image(images): |
|
|
|
return [images[0]], [image[1]] |
|
elif is_valid_video(images): |
|
|
|
return ["video"], [images] |
|
elif is_valid_image(images): |
|
|
|
return ["image"], [images] |
|
|
|
raise ValueError(f"Could not make batched images from {images}") |
|
|
|
|
|
def frame_sample(duration, mode='uniform', num_frames=None, vid_fps=None, fps=None): |
|
if mode == 'uniform': |
|
assert num_frames is not None, "Number of frames must be provided for uniform sampling." |
|
if duration <= num_frames: |
|
return np.arange(duration).astype(int) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return np.linspace(0, duration-1, num_frames, dtype=int) |
|
elif mode == 'fps': |
|
assert vid_fps is not None, "FPS must be provided for FPS sampling." |
|
assert fps is not None, "FPS must be provided for FPS sampling." |
|
segment_len = min(vid_fps // fps, duration) |
|
return np.arange(segment_len // 2, duration, segment_len, dtype=int) |
|
else: |
|
raise ImportError(f'Unsupported frame sampling mode: {mode}') |
|
|
|
|
|
def load_video_from_ids(video_path, s=None, e=None, fps=None, max_frames=128, temporal_factor=1): |
|
if s is not None and e is not None: |
|
s = s if s >= 0. else 0. |
|
e = e if e >= 0. else 0. |
|
if s > e: |
|
s, e = e, s |
|
elif s == e: |
|
e = s + 1 |
|
|
|
|
|
if os.path.isdir(video_path): |
|
frame_files = sorted(os.listdir(video_path)) |
|
|
|
vid_fps = 3 |
|
num_frames_of_video = len(frame_files) |
|
elif video_path.endswith('.gif'): |
|
gif_reader = imageio.get_reader(video_path) |
|
|
|
vid_fps = 25 |
|
num_frames_of_video = len(gif_reader) |
|
else: |
|
vreader = VideoReader(video_path, ctx=cpu(0), num_threads=2) |
|
|
|
|
|
vid_fps = vreader.get_avg_fps() |
|
num_frames_of_video = len(vreader) |
|
|
|
|
|
f_start = 0 if s is None else max(int(s * vid_fps) - 1, 0) |
|
f_end = num_frames_of_video - 1 if e is None else min(int(e * vid_fps) - 1, num_frames_of_video - 1) |
|
frame_indices = list(range(f_start, f_end + 1)) |
|
|
|
duration = len(frame_indices) |
|
|
|
if fps is not None and duration / vid_fps < max_frames: |
|
sampled_frame_indices = [frame_indices[i] for i in frame_sample(duration, mode='fps', vid_fps=vid_fps, fps=fps)] |
|
else: |
|
sampled_frame_indices = [frame_indices[i] for i in frame_sample(duration, mode='uniform', num_frames=max_frames)] |
|
|
|
|
|
if os.path.isdir(video_path): |
|
frames = np.array([cv2.cvtColor(cv2.imread(os.path.join(video_path, frame_files[frame_idx])), cv2.COLOR_BGR2RGB) for frame_idx in sampled_frame_indices]) |
|
elif video_path.endswith('.gif'): |
|
frames = np.array([cv2.cvtColor(frame, cv2.COLOR_RGBA2RGB) for idx, frame in enumerate(gif_reader) if idx in sampled_frame_indices]) |
|
else: |
|
frames = vreader.get_batch(sampled_frame_indices).asnumpy() |
|
|
|
frames = frames.transpose(0, 3, 1, 2) |
|
timestamps = [x / vid_fps for x in sampled_frame_indices] |
|
|
|
if temporal_factor > 1: |
|
pad_length = temporal_factor - len(frames) % temporal_factor |
|
frames = np.concatenate([frames, frames[-1:].repeat(pad_length, axis=0)]) |
|
[timestamps.append(timestamps[-1] + 1 / fps) for _ in range(pad_length)] |
|
|
|
frames = [frame for frame in frames] |
|
|
|
return frames, timestamps |
|
|
|
|
|
class ChatTemplateKwargs(TypedDict, total=False): |
|
|
|
chat_template: Optional[str] |
|
add_system_prompt: Optional[bool] |
|
add_generation_prompt: Optional[bool] |
|
|
|
|
|
class Videollama3Qwen2ProcessorKwargs(ProcessingKwargs, ChatTemplateKwargs, total=False): |
|
|
|
chat_template_kwargs: ChatTemplateKwargs = { |
|
**ChatTemplateKwargs.__annotations__, |
|
} |
|
|
|
_defaults = { |
|
"text_kwargs": { |
|
"padding": False, |
|
}, |
|
"image_kwargs": { |
|
"merge_size": None, |
|
}, |
|
"chat_template_kwargs": { |
|
"chat_template": None, |
|
"add_system_prompt": False, |
|
"add_generation_prompt": False, |
|
}, |
|
} |
|
|
|
|
|
class Videollama3Qwen2Processor(ProcessorMixin): |
|
|
|
attributes = ["image_processor", "tokenizer"] |
|
image_processor_class = "Videollama3ImageProcessor" |
|
tokenizer_class = ("Qwen2Tokenizer", "Qwen2TokenizerFast") |
|
valid_kwargs = ["chat_template", "image_merge_size", "video_merge_size", "fps", "max_frames"] |
|
|
|
def __init__( |
|
self, |
|
image_processor=None, |
|
tokenizer=None, |
|
chat_template: str = None, |
|
image_merge_size: int = 1, |
|
video_merge_size: int = 2, |
|
fps: Optional[int] = 1, |
|
max_frames: Optional[int] = 128, |
|
): |
|
self.image_processor = image_processor |
|
self.tokenizer = tokenizer |
|
if chat_template is None: |
|
chat_template = self.tokenizer.chat_template |
|
self.chat_template = chat_template |
|
|
|
self.image_merge_size = image_merge_size |
|
self.video_merge_size = video_merge_size |
|
self.fps = fps |
|
self.max_frames = max_frames |
|
|
|
self.generation_prompt = self._infer_generation_prompt() |
|
self.generation_prompt_ids = self.tokenizer.encode(self.generation_prompt, return_tensors="pt") |
|
self.generation_prompt_length = len(self.generation_prompt_ids[0]) |
|
self.image_token_id = self.tokenizer.convert_tokens_to_ids(DEFAULT_IMAGE_TOKEN) |
|
self.eos_token_id = self.tokenizer.eos_token_id |
|
|
|
@classmethod |
|
def _get_arguments_from_pretrained(cls, pretrained_model_name_or_path, **kwargs): |
|
args = [] |
|
for attribute_name in cls.attributes: |
|
class_name = getattr(cls, f"{attribute_name}_class") |
|
if isinstance(class_name, tuple): |
|
classes = tuple(_custom_import(n) if n is not None else None for n in class_name) |
|
use_fast = kwargs.get("use_fast", True) |
|
if use_fast and classes[1] is not None: |
|
attribute_class = classes[1] |
|
else: |
|
attribute_class = classes[0] |
|
else: |
|
attribute_class = _custom_import(class_name) |
|
|
|
args.append(attribute_class.from_pretrained(pretrained_model_name_or_path, **kwargs)) |
|
return args |
|
|
|
def get_generation_prompt(self): |
|
return self.generation_prompt |
|
|
|
def get_generation_prompt_ids(self): |
|
return self.generation_prompt_ids |
|
|
|
def _infer_generation_prompt(self): |
|
pseudo_message = [{"role": "user", "content": ""}] |
|
instruction = self.apply_chat_template(pseudo_message, tokenize=False, add_generation_prompt=True) |
|
conversation = self.apply_chat_template(pseudo_message, tokenize=False, add_generation_prompt=False) |
|
return instruction.replace(conversation, "") |
|
|
|
def _get_downsampled_grid_sizes(self, image_inputs: Dict[str, Any]): |
|
grid_sizes = [] |
|
for grid_size, merge_size in zip(image_inputs.get("grid_sizes", []), image_inputs.get("merge_sizes", [])): |
|
if not torch.all(grid_size[1:] % merge_size == 0): |
|
warnings.warn(f"Grid size {grid_size} is not divisible by merge size. Some undesired errors may occur.") |
|
if grid_size[0] == 1: |
|
grid_sizes.append(grid_size[1:] / merge_size) |
|
elif grid_size[0] > 1: |
|
grid_sizes.extend([grid_size[1:] / merge_size] * grid_size[0]) |
|
return grid_sizes |
|
|
|
def _get_visual_seq_len(self, grid_size: torch.Tensor): |
|
num_tokens = int(grid_size.prod().item()) |
|
return num_tokens |
|
|
|
def load_images(self, image_path: Union[str, List[str], Image.Image, List[Image.Image]]): |
|
if isinstance(image_path, str) and os.path.isfile(image_path): |
|
|
|
images = [Image.open(image_path).convert('RGB')] |
|
elif isinstance(image_path, str) and os.path.isdir(image_path): |
|
|
|
images = [Image.open(os.path.join(image_path, f)).convert('RGB') for f in sorted(os.listdir(image_path))] |
|
elif isinstance(image_path, list) and isinstance(image_path[0], str): |
|
|
|
images = [Image.open(f).convert('RGB') for f in image_path] |
|
elif isinstance(image_path, list) and isinstance(image_path[0], Image.Image): |
|
images = [np.array(x) for x in image_path] |
|
elif isinstance(image_path, Image.Image): |
|
images = [np.array(image_path)] |
|
else: |
|
raise ValueError(f"Unsupported image path type: {type(image_path)}") |
|
return images |
|
|
|
def load_video( |
|
self, |
|
video_path: str, |
|
start_time: Optional[float] = None, |
|
end_time: Optional[float] = None, |
|
fps: Optional[float] = None, |
|
max_frames: Optional[float] = None, |
|
size: Optional[int] = None, |
|
size_divisible: int = 1, |
|
precise_time: bool = False, |
|
verbose: bool = False, |
|
temporal_factor: int = 1 |
|
): |
|
""" |
|
Load and process a video file and return the frames and the timestamps of each frame. |
|
|
|
Args: |
|
video_path (str): Path to the video file. |
|
start_time (float, optional): Start time in seconds. Defaults to None. |
|
end_time (float, optional): End time in seconds. Defaults to None. |
|
fps (float, optional): Frames per second. Defaults to None. |
|
num_frames (float, optional): Number of frames to sample. Defaults to None. |
|
size (int, optional): Size of the shortest side. Defaults to None. |
|
size_divisible (int, optional): Size divisible by this number. Defaults to 1. |
|
precise_time (bool, optional): Whether to use precise time. Defaults to False. |
|
verbose (bool, optional): Print ffmpeg output. Defaults to False. |
|
|
|
Returns: |
|
frames (List[PIL.Image]): List of frames. |
|
timestamps (List[float]): List of timestamps. |
|
""" |
|
fps = self.fps if fps is None else fps |
|
max_frames = self.max_frames if max_frames is None else max_frames |
|
|
|
if start_time is not None and end_time is not None and end_time - start_time < 1: |
|
return load_video_from_ids(video_path, start_time, end_time, fps=fps, max_frames=max_frames) |
|
if os.path.isdir(video_path): |
|
return load_video_from_ids(video_path, start_time, end_time, fps=fps, max_frames=max_frames) |
|
if video_path.endswith('.gif'): |
|
return load_video_from_ids(video_path, start_time, end_time, fps=fps, max_frames=max_frames) |
|
probe = ffmpeg.probe(video_path) |
|
duration = float(probe['format']['duration']) |
|
video_stream = next((stream for stream in probe['streams'] if stream['codec_type'] == 'video'), None) |
|
w, h = int(video_stream['width']), int(video_stream['height']) |
|
|
|
kwargs, input_kwargs, output_kwargs = {}, {}, {} |
|
do_trim = start_time is not None or end_time is not None |
|
if start_time is not None: |
|
new_start_time = max(float(video_stream['start_time']), start_time) |
|
duration -= new_start_time - start_time |
|
start_time = new_start_time |
|
else: |
|
start_time = float(video_stream['start_time']) |
|
if end_time is not None: |
|
duration = min(duration, end_time - start_time) |
|
else: |
|
duration = duration |
|
if do_trim: |
|
kwargs = {'ss': start_time, 't': duration} |
|
if precise_time: |
|
output_kwargs.update(kwargs) |
|
else: |
|
input_kwargs.update(kwargs) |
|
|
|
if size is not None: |
|
scale_factor = size / min(w, h) |
|
new_w, new_h = round(w * scale_factor), round(h * scale_factor) |
|
else: |
|
new_w, new_h = w, h |
|
new_w = new_w // size_divisible * size_divisible |
|
new_h = new_h // size_divisible * size_divisible |
|
|
|
|
|
|
|
|
|
|
|
|
|
stream = ffmpeg.input(video_path, **input_kwargs) |
|
if fps is not None: |
|
stream = ffmpeg.filter(stream, "fps", fps=fps, round="down") |
|
if new_w != w or new_h != h: |
|
stream = ffmpeg.filter(stream, 'scale', new_w, new_h) |
|
stream = ffmpeg.output(stream, "pipe:", format="rawvideo", pix_fmt="rgb24", **output_kwargs) |
|
out, _ = ffmpeg.run(stream, capture_stdout=True, quiet=not verbose) |
|
|
|
frames = np.frombuffer(out, np.uint8).reshape([-1, new_h, new_w, 3]).transpose([0, 3, 1, 2]) |
|
|
|
if fps is not None: |
|
timestamps = np.arange(start_time, start_time + duration + 1 / fps, 1 / fps)[:len(frames)] |
|
else: |
|
timestamps = np.linspace(start_time, start_time + duration, len(frames)) |
|
|
|
if max_frames is not None and len(frames) > max_frames: |
|
indices = np.linspace(0, len(frames) - 1, max_frames, dtype=int) |
|
frames = frames[indices] |
|
timestamps = timestamps[indices] |
|
|
|
if temporal_factor > 1: |
|
pad_length = temporal_factor - len(frames) % temporal_factor |
|
frames = np.concatenate([frames, frames[-1:].repeat(pad_length, axis=0)]) |
|
timestamps = np.concatenate([timestamps, timestamps[-1:].repeat(pad_length) + np.arange(1, pad_length + 1) / fps]) |
|
|
|
frames = [frame for frame in frames] |
|
timestamps = [timestamp for timestamp in timestamps] |
|
|
|
return frames, timestamps |
|
|
|
def _load_multimodal_data(self, conversation: Conversation): |
|
multimodal_info = defaultdict(list) |
|
new_conversation = [] |
|
for message in conversation: |
|
new_message = {"role": message["role"]} |
|
if not isinstance(message["content"], (list, tuple)): |
|
new_message["content"] = message["content"] |
|
new_conversation.append(new_message) |
|
continue |
|
|
|
new_contents = [] |
|
for content in message["content"]: |
|
if not isinstance(content, dict): |
|
new_contents.append(content) |
|
continue |
|
assert "type" in content, "Content must have 'type' field." |
|
if content["type"] in ["image", "video"] and content["type"] in content and isinstance(content[content["type"]], dict): |
|
|
|
load_args = content[content["type"]] |
|
data_id = json.dumps({k: v for k, v in load_args.items() if not k in ["start_time", "end_time"]}) |
|
new_content = copy.deepcopy(content) |
|
multimodal_info[data_id].append(new_content) |
|
new_contents.append(new_content) |
|
else: |
|
new_contents.append(content) |
|
|
|
new_message["content"] = new_contents |
|
new_conversation.append(new_message) |
|
|
|
for data_id, contents in multimodal_info.items(): |
|
data_type = contents[0]["type"] |
|
if data_type == "image": |
|
image = self.load_images(contents[0][data_type]["image_path"])[0] |
|
for content in contents: |
|
content["image"] = [image.copy()] |
|
|
|
elif data_type == "video": |
|
|
|
start_times = [content["video"].get("start_time", 0.) for content in contents] |
|
end_times = [content["video"].get("end_time", float("inf")) for content in contents] |
|
|
|
load_args = contents[0][data_type] |
|
start_time, end_time = min(start_times), max(end_times) |
|
if start_time > 0: |
|
load_args["start_time"] = start_time |
|
if end_time < float("inf"): |
|
load_args["end_time"] = end_time |
|
images, timestamps = self.load_video(**load_args) |
|
|
|
for content, start_time, end_time in zip(contents, start_times, end_times): |
|
cur_images, cur_timestamps = [], [] |
|
for image, timestamp in zip(images, timestamps): |
|
if start_time <= timestamp <= end_time: |
|
cur_images.append(image.copy()) |
|
cur_timestamps.append(timestamp) |
|
|
|
content[data_type] = cur_images |
|
content["num_frames"] = len(cur_images) |
|
content["timestamps"] = cur_timestamps |
|
|
|
return new_conversation |
|
|
|
def _gather_multimodal_data(self, conversation: Conversation): |
|
images = [] |
|
for message in conversation: |
|
if not isinstance(message["content"], (list, tuple)): |
|
continue |
|
for content in message["content"]: |
|
if not isinstance(content, dict): |
|
continue |
|
if content["type"] == "video": |
|
video = content["video"] |
|
assert is_valid_video(video), f"Invalid video data: {video}." |
|
images.append(("video", video)) |
|
if content["type"] == "image": |
|
image = content["image"] |
|
images.append(("image", image)) |
|
images = images if len(images) > 0 else None |
|
return images |
|
|
|
def _process_conversation_with_label( |
|
self, |
|
conversation: Conversation, |
|
image_inputs: Dict[str, Any], |
|
**kwargs, |
|
): |
|
assert kwargs.pop("return_tensors", "pt") == "pt", "Only PyTorch tensors are supported when return_labels=True." |
|
assert not "add_generation_prompt" in kwargs, "'add_generation_prompt' argument is not supported when return_labels=True." |
|
|
|
output_kwargs = self._merge_kwargs( |
|
Videollama3Qwen2ProcessorKwargs, |
|
tokenizer_init_kwargs=self.tokenizer.init_kwargs, |
|
**kwargs, |
|
) |
|
output_kwargs["chat_template_kwargs"].pop("add_generation_prompt") |
|
|
|
grid_sizes = self._get_downsampled_grid_sizes(image_inputs) |
|
text_inputs = {"input_ids": [], "labels": []} |
|
sample_types_list = [] |
|
image_idx = 0 |
|
|
|
for message_idx, message in enumerate(conversation): |
|
prompt = self.apply_chat_template( |
|
[message], |
|
tokenize=False, |
|
add_generation_prompt=False, |
|
**output_kwargs["chat_template_kwargs"], |
|
) |
|
prompt_chunks = prompt.split(DEFAULT_IMAGE_TOKEN) |
|
prompt = [] |
|
for chunk_idx in range(len(prompt_chunks) - 1): |
|
prompt.append(prompt_chunks[chunk_idx]) |
|
num_tokens = self._get_visual_seq_len(grid_sizes[image_idx]) |
|
prompt.append(DEFAULT_IMAGE_TOKEN * num_tokens) |
|
image_idx += 1 |
|
prompt.append(prompt_chunks[-1]) |
|
prompt = "".join(prompt) |
|
|
|
|
|
input_ids = self.tokenizer.encode(prompt, return_tensors="pt", **output_kwargs["text_kwargs"])[0] |
|
text_inputs["input_ids"].append(input_ids) |
|
|
|
targets = torch.full_like(input_ids, IGNORE_INDEX) |
|
sample_types = torch.full_like(input_ids, IGNORE_INDEX) |
|
if message["role"] == "assistant": |
|
targets[self.generation_prompt_length:-1] = input_ids[self.generation_prompt_length:-1].clone() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if message_idx > 0 and conversation[message_idx - 1]["role"] == "stream": |
|
targets[0] = input_ids[0] |
|
|
|
sample_types[0] = input_ids[0] |
|
|
|
text_inputs["labels"].append(targets) |
|
sample_types_list.append(sample_types) |
|
|
|
|
|
text_inputs = {k: torch.cat(v) for k, v in text_inputs.items()} |
|
sample_types = torch.cat(sample_types_list) |
|
types, counts = torch.unique(sample_types[sample_types > -1], return_counts=True) |
|
|
|
if len(types) > 0: |
|
target_num_samples = counts.amin() |
|
for type_id, type_count in zip(types, counts): |
|
if type_count > target_num_samples: |
|
indices = torch.nonzero(sample_types == type_id)[:, 0] |
|
random_selector = torch.randperm(indices.size(0))[:-target_num_samples] |
|
text_inputs["labels"][indices[random_selector]] = IGNORE_INDEX |
|
|
|
|
|
assert len(grid_sizes) == image_idx, "Number of images does not match the number of image tokens in the text." |
|
|
|
return text_inputs |
|
|
|
def _process_conversation_without_label( |
|
self, |
|
conversation: Conversation, |
|
image_inputs: Dict[str, Any], |
|
**kwargs, |
|
): |
|
output_kwargs = self._merge_kwargs( |
|
Videollama3Qwen2ProcessorKwargs, |
|
tokenizer_init_kwargs=self.tokenizer.init_kwargs, |
|
**kwargs, |
|
) |
|
prompt = self.apply_chat_template( |
|
conversation, |
|
tokenize=False, |
|
**output_kwargs["chat_template_kwargs"], |
|
) |
|
return self.process_text(prompt, image_inputs, **output_kwargs["text_kwargs"]) |
|
|
|
def _process_conversation( |
|
self, |
|
conversation: Conversation, |
|
images: Optional[Union[BatchedImage, BatchedNamedImage]] = None, |
|
return_labels: bool = False, |
|
**kwargs: Unpack[Videollama3Qwen2ProcessorKwargs], |
|
) -> BatchFeature: |
|
assert isinstance(conversation, list), "Conversation must be a list of messages." |
|
|
|
if images is None: |
|
conversation = self._load_multimodal_data(conversation) |
|
images = self._gather_multimodal_data(conversation) |
|
|
|
output_kwargs = self._merge_kwargs( |
|
Videollama3Qwen2ProcessorKwargs, |
|
tokenizer_init_kwargs=self.tokenizer.init_kwargs, |
|
**kwargs, |
|
) |
|
|
|
if images is not None: |
|
image_inputs = self.process_images(images, **output_kwargs["images_kwargs"]) |
|
else: |
|
image_inputs = {} |
|
|
|
if return_labels: |
|
text_inputs = self._process_conversation_with_label(conversation, image_inputs, **kwargs) |
|
else: |
|
text_inputs = self._process_conversation_without_label(conversation, image_inputs, **kwargs) |
|
|
|
return BatchFeature(data={**text_inputs, **image_inputs}) |
|
|
|
def _process_plain( |
|
self, |
|
text: Union[TextInput, PreTokenizedInput] = None, |
|
images: Optional[Union[BatchedImage, BatchedNamedImage]] = None, |
|
return_labels: bool = False, |
|
**kwargs: Unpack[Videollama3Qwen2ProcessorKwargs], |
|
) -> BatchFeature: |
|
if text is None: |
|
raise ValueError("You must provide 'text' or 'message'.") |
|
if return_labels: |
|
raise ValueError("return_labels is not supported for plain text processing.") |
|
|
|
output_kwargs = self._merge_kwargs( |
|
Videollama3Qwen2ProcessorKwargs, |
|
tokenizer_init_kwargs=self.tokenizer.init_kwargs, |
|
**kwargs, |
|
) |
|
|
|
if images is not None: |
|
image_inputs = self.process_images(images, **output_kwargs["images_kwargs"]) |
|
else: |
|
image_inputs = {} |
|
|
|
text_inputs = self.process_text(text, image_inputs, **output_kwargs["text_kwargs"]) |
|
|
|
return BatchFeature(data={**text_inputs, **image_inputs}) |
|
|
|
def process_images(self, images: Union[BatchedImage, BatchedNamedImage], **kwargs): |
|
modals, images = make_batched_images(images) |
|
if not "merge_size" in kwargs: |
|
kwargs["merge_size"] = [ |
|
self.image_merge_size if modal == "image" else self.video_merge_size |
|
for modal in modals |
|
] |
|
image_inputs = self.image_processor(images=images, **kwargs) |
|
image_inputs["modals"] = modals |
|
return image_inputs |
|
|
|
def process_text( |
|
self, |
|
text: TextInput, |
|
image_inputs: Dict[str, Any], |
|
**kwargs, |
|
): |
|
grid_sizes = self._get_downsampled_grid_sizes(image_inputs) |
|
|
|
kwargs.pop("padding") |
|
kwargs.pop("padding_side") |
|
|
|
image_idx = 0 |
|
while DEFAULT_IMAGE_TOKEN in text: |
|
num_tokens = self._get_visual_seq_len(grid_sizes[image_idx]) |
|
text = text.replace(DEFAULT_IMAGE_TOKEN, "<placeholder>" * num_tokens, 1) |
|
image_idx += 1 |
|
text = text.replace("<placeholder>", DEFAULT_IMAGE_TOKEN) |
|
|
|
assert len(grid_sizes) == image_idx, "Number of images does not match the number of image tokens in the text." |
|
|
|
text_inputs = self.tokenizer(text, **kwargs) |
|
return text_inputs |
|
|
|
def __call__( |
|
self, |
|
text: Optional[TextInput] = None, |
|
conversation: Optional[Conversation] = None, |
|
images: Optional[Union[BatchedImage, BatchedNamedImage]] = None, |
|
return_labels: bool = False, |
|
**kwargs: Unpack[Videollama3Qwen2ProcessorKwargs], |
|
) -> BatchFeature: |
|
if conversation is not None: |
|
if text is not None: |
|
raise ValueError("You cannot provide 'message' with 'text'.") |
|
return self._process_conversation(conversation, images, return_labels, **kwargs) |
|
return self._process_plain(text, images, return_labels, **kwargs) |
|
|
|
def batch_decode(self, *args, **kwargs): |
|
return self.tokenizer.batch_decode(*args, **kwargs) |
|
|
|
def decode(self, *args, **kwargs): |
|
return self.tokenizer.decode(*args, **kwargs) |
|
|
|
def apply_chat_template( |
|
self, |
|
conversation: Conversation, |
|
chat_template: Optional[str] = None, |
|
tokenize: bool = False, |
|
add_system_prompt: bool = False, |
|
add_generation_prompt: bool = False, |
|
image_token: Optional[str] = DEFAULT_IMAGE_TOKEN, |
|
**kwargs, |
|
) -> str: |
|
""" |
|
Similar to the `apply_chat_template` method on tokenizers, this method applies a Jinja template to input |
|
conversations to turn them into a single tokenizable string. |
|
|
|
Args: |
|
conversation (`List[Dict, str, str]`): |
|
The conversation to format. |
|
chat_template (`Optional[str]`, *optional*): |
|
The Jinja template to use for formatting the conversation. If not provided, the tokenizer's |
|
chat template is used. |
|
tokenize (`bool`, *optional*, defaults to `False`): |
|
Whether to tokenize the output or not. |
|
add_system_prompt (`bool`, *optional*, defaults to `False`): |
|
Whether to add the system prompt to the output or not. |
|
add_generation_prompt (`bool`, *optional*, defaults to `False`): |
|
Whether to add the generation prompt to the output or not. |
|
image_token (`Optional[str]`, *optional*, defaults to `<image>`): |
|
The token to use for indicating images in the conversation. |
|
**kwargs: |
|
Additional keyword arguments |
|
""" |
|
|
|
if chat_template is None: |
|
if self.chat_template is not None: |
|
chat_template = self.chat_template |
|
else: |
|
raise ValueError( |
|
"No chat template is set for this processor. Please either set the `chat_template` attribute, " |
|
"or provide a chat template as an argument. See " |
|
"https://huggingface.co/docs/transformers/main/en/chat_templating for more information." |
|
) |
|
return self.tokenizer.apply_chat_template( |
|
conversation, |
|
chat_template=chat_template, |
|
tokenize=tokenize, |
|
add_system_prompt=add_system_prompt, |
|
add_generation_prompt=add_generation_prompt, |
|
image_token=image_token, |
|
**kwargs |
|
) |
|
|
|
@property |
|
def model_input_names(self): |
|
tokenizer_input_names = self.tokenizer.model_input_names |
|
image_processor_input_names = self.image_processor.model_input_names |
|
return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) + ["modals"] |
|
|
|
|
|
def _merge_kwargs( |
|
self, |
|
ModelProcessorKwargs: ProcessingKwargs, |
|
tokenizer_init_kwargs: Optional[Dict] = None, |
|
**kwargs, |
|
) -> Dict[str, Dict]: |
|
""" |
|
Method to merge dictionaries of kwargs cleanly separated by modality within a Processor instance. |
|
The order of operations is as follows: |
|
1) kwargs passed as before have highest priority to preserve BC. |
|
```python |
|
high_priority_kwargs = {"crop_size" = {"height": 222, "width": 222}, "padding" = "max_length"} |
|
processor(..., **high_priority_kwargs) |
|
``` |
|
2) kwargs passed as modality-specific kwargs have second priority. This is the recommended API. |
|
```python |
|
processor(..., text_kwargs={"padding": "max_length"}, images_kwargs={"crop_size": {"height": 222, "width": 222}}}) |
|
``` |
|
3) kwargs passed during instantiation of a modality processor have fourth priority. |
|
```python |
|
tokenizer = tokenizer_class(..., {"padding": "max_length"}) |
|
image_processor = image_processor_class(...) |
|
processor(tokenizer, image_processor) # will pass max_length unless overriden by kwargs at call |
|
``` |
|
4) defaults kwargs specified at processor level have lowest priority. |
|
```python |
|
class MyProcessingKwargs(ProcessingKwargs, CommonKwargs, TextKwargs, ImagesKwargs, total=False): |
|
_defaults = { |
|
"text_kwargs": { |
|
"padding": "max_length", |
|
"max_length": 64, |
|
}, |
|
} |
|
``` |
|
Args: |
|
ModelProcessorKwargs (`ProcessingKwargs`): |
|
Typed dictionary of kwargs specifically required by the model passed. |
|
tokenizer_init_kwargs (`Dict`, *optional*): |
|
Dictionary of kwargs the tokenizer was instantiated with and need to take precedence over defaults. |
|
|
|
Returns: |
|
output_kwargs (`Dict`): |
|
Dictionary of per-modality kwargs to be passed to each modality-specific processor. |
|
|
|
""" |
|
|
|
output_kwargs = { |
|
"text_kwargs": {}, |
|
"images_kwargs": {}, |
|
"audio_kwargs": {}, |
|
"videos_kwargs": {}, |
|
"chat_template_kwargs": {}, |
|
"common_kwargs": {}, |
|
} |
|
|
|
default_kwargs = { |
|
"text_kwargs": {}, |
|
"images_kwargs": {}, |
|
"audio_kwargs": {}, |
|
"videos_kwargs": {}, |
|
"chat_template_kwargs": {}, |
|
"common_kwargs": {}, |
|
} |
|
|
|
used_keys = set() |
|
|
|
|
|
for modality in default_kwargs: |
|
default_kwargs[modality] = ModelProcessorKwargs._defaults.get(modality, {}).copy() |
|
|
|
for modality_key in ModelProcessorKwargs.__annotations__[modality].__annotations__.keys(): |
|
|
|
if modality_key in tokenizer_init_kwargs: |
|
value = ( |
|
getattr(self.tokenizer, modality_key) |
|
if hasattr(self.tokenizer, modality_key) |
|
else tokenizer_init_kwargs[modality_key] |
|
) |
|
default_kwargs[modality][modality_key] = value |
|
|
|
|
|
output_kwargs.update(default_kwargs) |
|
|
|
|
|
non_modality_kwargs = set(kwargs) - set(output_kwargs) |
|
for modality in output_kwargs: |
|
for modality_key in ModelProcessorKwargs.__annotations__[modality].__annotations__.keys(): |
|
|
|
if modality in kwargs: |
|
kwarg_value = kwargs[modality].pop(modality_key, "__empty__") |
|
|
|
if kwarg_value != "__empty__" and modality_key in non_modality_kwargs: |
|
raise ValueError( |
|
f"Keyword argument {modality_key} was passed two times:\n" |
|
f"in a dictionary for {modality} and as a **kwarg." |
|
) |
|
elif modality_key in kwargs: |
|
|
|
|
|
kwarg_value = kwargs.get(modality_key, "__empty__") |
|
else: |
|
kwarg_value = "__empty__" |
|
if kwarg_value != "__empty__": |
|
output_kwargs[modality][modality_key] = kwarg_value |
|
used_keys.add(modality_key) |
|
|
|
|
|
if any(key in default_kwargs for key in kwargs): |
|
|
|
for modality, subdict in kwargs.items(): |
|
if modality in default_kwargs: |
|
for subkey, subvalue in subdict.items(): |
|
if subkey not in used_keys: |
|
output_kwargs[modality][subkey] = subvalue |
|
used_keys.add(subkey) |
|
else: |
|
|
|
for key in kwargs: |
|
if key not in used_keys: |
|
output_kwargs["common_kwargs"][key] = kwargs[key] |
|
|
|
|
|
for modality in output_kwargs: |
|
output_kwargs[modality].update(output_kwargs["common_kwargs"]) |
|
return output_kwargs |
|
|