import os import json import argparse import numpy as np from tqdm import tqdm import torch import torch.distributed as dist from torch.utils.data import DataLoader from torchvision import transforms as T from data.dataset_for_clean_descrip import PoseHICODetDataset from data.convsersation import Conversation_For_Action_Pharse as Conversation import re from dataclasses import dataclass from transformers import Qwen3VLForConditionalGeneration from transformers import AutoTokenizer, AutoConfig, AutoProcessor def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) import os, json import torch @dataclass class DataCollatorForSupervisedDataset(object): def __init__(self, processor, data_path): self.processor = processor self.conv = Conversation( system='', data_path=data_path ) def __call__(self, data_dicts): """Collate examples for supervised fine-tuning.""" batch_prompts = [] batch_images = [] result_meta = [] for i, data_dict in enumerate(data_dicts): batch_images.append(data_dict['image']) batch_prompts.append(self.conv.get_prompt(data_dict['meta'])) result_meta.append(data_dict['meta']) messages = [] for prompt in zip(batch_prompts): messages.append([ {"role": "system", "content":[ {"type": "text", "text": self.conv.system},]}, {"role": "user", "content":[ {"type": "image"}, {"type": "text", "text": prompt},]}, ]) prompts = [self.processor.apply_chat_template(m, tokenize=False, add_generation_prompt=True) for m in messages] batch_tensors = self.processor( text=prompts, images=batch_images, return_tensors="pt", padding=True ) return batch_tensors, result_meta @torch.no_grad() def worker(model, processor, dataset, args, output_dir): rank = int(os.environ["LOCAL_RANK"]) world_size = int(os.environ["WORLD_SIZE"]) indices = list(range(rank, len(dataset), world_size)) print("==>" + " Worker {} Started, responsible for {} images".format(rank, len(indices))) sub_dataset = torch.utils.data.Subset(dataset, indices) batch_size = 16 data_loader = DataLoader(sub_dataset, batch_size=batch_size, shuffle=False, num_workers=0, collate_fn=DataCollatorForSupervisedDataset(processor, args.data_path)) labels = [] for batch_tensors, result_meta in tqdm(data_loader): input_ids = batch_tensors['input_ids'].cuda() batch_tensors = {k: v.cuda() for k, v in batch_tensors.items() if isinstance(v, torch.Tensor)} with torch.inference_mode(): output_dict = model.generate(do_sample=False, output_scores=True, return_dict_in_generate=True, max_new_tokens=1600, output_logits=True, **batch_tensors,) output_ids = output_dict['sequences'] for input_id, output_id, meta in zip(input_ids, output_ids, result_meta): input_token_len = input_id.shape[0] n_diff_input_output = (input_id != output_id[:input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] Sample: {n_diff_input_output} output_ids are not the same as the input_ids') #input_text = processor.tokenizer.batch_decode(output_id[:input_token_len].unsqueeze(0), skip_special_tokens=True)[0] output = processor.tokenizer.batch_decode(output_id[input_token_len:].unsqueeze(0), skip_special_tokens=True)[0] # print(output) # import pdb;pdb.set_trace() meta['action_description'] = output #import pdb;pdb.set_trace() labels.append(meta) local_rank = int(os.environ.get("LOCAL_RANK", "0")) output_path = os.path.join(args.output_dir, f'labels_{local_rank}.json') with open(output_path, "w", encoding="utf-8") as f: json.dump(labels, f, ensure_ascii=False, indent=2) def eval_model(args): torch.distributed.init_process_group(backend='nccl') rank = int(os.environ["LOCAL_RANK"]) world_size = int(os.environ["WORLD_SIZE"]) print('Init process group: world_size: {}, rank: {}'.format(world_size, rank)) torch.cuda.set_device(rank) disable_torch_init() model = Qwen3VLForConditionalGeneration.from_pretrained( args.model_path, torch_dtype=torch.bfloat16, trust_remote_code=True ) model = model.cuda() model.eval() processor = AutoProcessor.from_pretrained( args.model_path, trust_remote_code=True) processor.tokenizer.padding_side = "left" processor.tokenizer.pad_token = processor.tokenizer.eos_token dataset = PoseHICODetDataset( data_path=args.data_path, multimodal_cfg=dict(image_folder=os.path.join(args.data_path, 'Images/images/train2015'), data_augmentation=False, image_size=336,),) worker(model, processor, dataset, args, args.output_dir) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--model-path", type=str, default="facebook/opt-350m") parser.add_argument("--data-path", type=str, default="") parser.add_argument("--output-dir", type=str, default="") args = parser.parse_args() eval_model(args)