| | import cv2 |
| | import torch |
| | import torchvision |
| | from osdsynth.processor.wrappers.grounding_dino import get_grounding_dino_model |
| | from osdsynth.processor.wrappers.ram import get_tagging_model, run_tagging_model |
| | from osdsynth.processor.wrappers.sam import ( |
| | convert_detections_to_dict, |
| | convert_detections_to_list, |
| | crop_detections_with_xyxy, |
| | filter_detections, |
| | get_sam_predictor, |
| | get_sam_segmentation_from_xyxy, |
| | mask_subtract_contained, |
| | post_process_mask, |
| | sort_detections_by_area, |
| | ) |
| | from osdsynth.utils.logger import SkipImageException |
| | from osdsynth.visualizer.som import draw_som_on_image |
| | from PIL import Image |
| | import numpy as np |
| |
|
| | class SegmentImage: |
| | """Class to segment the image.""" |
| |
|
| | def __init__(self, cfg, logger, device, init_gdino=True, init_tagging=True, init_sam=True): |
| | self.cfg = cfg |
| | self.logger = logger |
| | self.device = device |
| |
|
| | if init_gdino: |
| | |
| | self.grounding_dino_model = get_grounding_dino_model(cfg, device) |
| | else: |
| | self.grounding_dino_model = None |
| |
|
| | if init_tagging: |
| | |
| | self.tagging_transform, self.tagging_model = get_tagging_model(cfg, device) |
| | else: |
| | self.tagging_transform = self.tagging_model = None |
| |
|
| | if init_sam: |
| | |
| | self.sam_predictor = get_sam_predictor(cfg.sam_variant, device) |
| | else: |
| | self.sam_predictor = None |
| |
|
| | pass |
| |
|
| | def process(self, image_bgr, two_class ,plot_som=True): |
| | """Segment the image.""" |
| |
|
| | image_rgb = cv2.cvtColor(image_bgr, cv2.COLOR_BGR2RGB) |
| | image_rgb_pil = Image.fromarray(image_rgb) |
| | |
| | |
| |
|
| | img_tagging = image_rgb_pil.resize((384, 384)) |
| | img_tagging = self.tagging_transform(img_tagging).unsqueeze(0).to(self.device) |
| |
|
| | |
| | if two_class is None: |
| | classes = run_tagging_model(self.cfg, img_tagging, self.tagging_model) |
| | else: |
| | classes = two_class |
| |
|
| | if len(classes) == 0: |
| | raise SkipImageException("No foreground objects detected by tagging model.") |
| |
|
| | |
| | detections = self.grounding_dino_model.predict_with_classes( |
| | image=image_bgr, |
| | classes=classes, |
| | box_threshold=self.cfg.box_threshold, |
| | text_threshold=self.cfg.text_threshold, |
| | ) |
| |
|
| |
|
| | if len(detections.class_id) < 1: |
| | raise SkipImageException("No object detected.") |
| | |
| |
|
| |
|
| | |
| | nms_idx = ( |
| | torchvision.ops.nms( |
| | torch.from_numpy(detections.xyxy), |
| | torch.from_numpy(detections.confidence), |
| | self.cfg.nms_threshold, |
| | ) |
| | .numpy() |
| | .tolist() |
| | ) |
| |
|
| | print(f"Before NMS: {len(detections.xyxy)} detections") |
| | detections.xyxy = detections.xyxy[nms_idx] |
| | detections.confidence = detections.confidence[nms_idx] |
| | detections.class_id = detections.class_id[nms_idx] |
| | print(f"After NMS: {len(detections.xyxy)} detections") |
| |
|
| | |
| | valid_idx = detections.class_id != -1 |
| | detections.xyxy = detections.xyxy[valid_idx] |
| | detections.confidence = detections.confidence[valid_idx] |
| | detections.class_id = detections.class_id[valid_idx] |
| |
|
| | |
| | detections.mask = get_sam_segmentation_from_xyxy( |
| | sam_predictor=self.sam_predictor, image=image_rgb, xyxy=detections.xyxy |
| | ) |
| |
|
| | |
| | detections_dict = convert_detections_to_dict(detections, classes) |
| | |
| | |
| | detections_dict = filter_detections(self.cfg, detections_dict, image_rgb) |
| |
|
| | if len(detections_dict["xyxy"]) < 1: |
| | raise SkipImageException("No object detected after filtering.") |
| |
|
| | |
| | detections_dict["subtracted_mask"], mask_contained = mask_subtract_contained( |
| | detections_dict["xyxy"], detections_dict["mask"], th1=0.05, th2=0.05 |
| | ) |
| |
|
| | |
| | detections_dict = sort_detections_by_area(detections_dict) |
| |
|
| | |
| | detections_dict = post_process_mask(detections_dict) |
| |
|
| | |
| | detections_list = convert_detections_to_list(detections_dict, classes) |
| | |
| | |
| | |
| |
|
| | detections_list = crop_detections_with_xyxy(self.cfg, image_rgb_pil, detections_list) |
| |
|
| | detections_list = segmentImage(detections_list, image_rgb_pil) |
| | |
| | detections_list = add_index_to_class(detections_list) |
| | |
| | if two_class is not None: |
| | if len(two_class)==2 and len(detections_list) != 2: |
| | raise SkipImageException("Not all objects detected.") |
| | |
| | if len(two_class)==1 and len(detections_list) != 1: |
| | raise SkipImageException("Not all objects detected.") |
| | |
| | if len(two_class)==3 and len(detections_list) != 3: |
| | raise SkipImageException("Not all objects detected.") |
| | |
| | if len(two_class)==2: |
| | detections_two_class = [detections_list[0]['class_name'][:-1], detections_list[1]['class_name'][:-1]] |
| | if two_class[0] not in detections_two_class or two_class[1] not in detections_two_class: |
| | raise SkipImageException("Not all objects detected.") |
| | |
| | if len(two_class)==3: |
| | detections_two_class = [detections_list[0]['class_name'][:-1], detections_list[1]['class_name'][:-1], detections_list[2]['class_name'][:-1]] |
| | if two_class[0] not in detections_two_class or two_class[1] not in detections_two_class or two_class[2] not in detections_two_class: |
| | raise SkipImageException("Not all objects detected.") |
| | |
| | |
| |
|
| | if plot_som: |
| | |
| | vis_som = draw_som_on_image( |
| | detections_dict, |
| | image_rgb, |
| | label_mode="1", |
| | alpha=0.4, |
| | anno_mode=["Mask", "Mark", "Box"], |
| | ) |
| | else: |
| | vis_som = None |
| | |
| | |
| | |
| | return vis_som, detections_list |
| |
|
| | |
| | def segmentImage(detections_list, image_rgb_pil): |
| | |
| | for i in range(len(detections_list)): |
| | image_pil = detections_list[i]['image_crop'] |
| | mask_pil = Image.fromarray(detections_list[i]['mask_crop']) |
| | |
| | image_rgba = image_pil.convert("RGBA") |
| | |
| | transparent_bg = Image.new("RGBA", image_rgba.size, (0, 0, 0, 0)) |
| |
|
| | |
| | segmented_image = Image.composite( |
| | image_rgba, |
| | transparent_bg, |
| | mask_pil |
| | ) |
| | |
| | detections_list[i]['image_segment'] = segmented_image |
| | |
| | return detections_list |
| | |
| | def skipbyconfidence(detections_list): |
| | skip_index = [] |
| | for i in range(len(detections_list)): |
| | if detections_list[i]['confidence'] < 0.3: |
| | skip_index.append(i) |
| | |
| | for i in skip_index[::-1]: |
| | del detections_list[i] |
| | |
| | return detections_list |
| | |
| | def add_bbox_and_taggingtext_to_image(image, detections_list): |
| | for i in range(len(detections_list)): |
| | bbox = detections_list[i]['xyxy'] |
| | label = detections_list[i]['class_name'] |
| | confidence = detections_list[i]['confidence'] |
| | |
| | cv2.rectangle(image, (int(bbox[0]), int(bbox[1])), (int(bbox[2]), int(bbox[3])), (0, 255, 0), 2) |
| | cv2.putText(image, f"{label} {confidence:.2f}", (int(bbox[0]), int((bbox[1]+bbox[3])/2)), cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 255, 0), 2) |
| | |
| | return image |
| |
|
| | def add_index_to_class(detections_list): |
| | |
| | class_index = {} |
| | for detection in detections_list: |
| | class_name = detection['class_name'] |
| | if class_name not in class_index: |
| | class_index[class_name] = 0 |
| | else: |
| | class_index[class_name] += 1 |
| |
|
| | detection['class_name'] = f"{class_name}{class_index[class_name]}" |
| | return detections_list |
| | |