[Inference.Core] MeshGraphormer Hand Refiner¶
Documentation¶
- Class name:
Inference_Core_MeshGraphormer-DepthMapPreprocessor
- Category:
ControlNet Preprocessors/Normal and Depth Estimators
- Output node:
False
This node is designed to preprocess images for depth map generation and hand pose estimation using the MeshGraphormer model. It handles the conversion of input images to a format suitable for the model, performs inference to generate depth maps and masks, and processes these outputs for further use in hand refinement tasks.
Input types¶
Required¶
image
- The input image or batch of images to be processed for depth map and mask generation. It plays a crucial role in the node's execution as it directly influences the quality and accuracy of the generated depth maps and masks.
- Comfy dtype:
IMAGE
- Python dtype:
torch.Tensor
Optional¶
mask_bbox_padding
- Determines the padding around detected bounding boxes for masks, influencing the size and coverage of the generated masks.
- Comfy dtype:
INT
- Python dtype:
int
resolution
- Specifies the resolution at which the depth map and mask detection should be performed, affecting the detail level of the output.
- Comfy dtype:
INT
- Python dtype:
int
mask_type
- Defines the strategy for mask generation, either based on depth values or tight bounding boxes, affecting the mask's shape and coverage.
- Comfy dtype:
COMBO[STRING]
- Python dtype:
str
mask_expand
- Controls the expansion of masks beyond their original boundaries, useful for ensuring complete coverage of hand areas in depth maps.
- Comfy dtype:
INT
- Python dtype:
int
rand_seed
- A seed value for random number generation, ensuring reproducibility of the mask generation process.
- Comfy dtype:
INT
- Python dtype:
int
Output types¶
IMAGE
- Comfy dtype:
IMAGE
- The processed images after depth map and mask generation, ready for further analysis or visualization.
- Python dtype:
torch.Tensor
- Comfy dtype:
INPAINTING_MASK
- Comfy dtype:
MASK
- The masks indicating areas to be inpainted or further processed, typically highlighting regions of interest such as hands.
- Python dtype:
torch.Tensor
- Comfy dtype:
Usage tips¶
- Infra type:
GPU
- Common nodes: unknown
Source code¶
class Mesh_Graphormer_Depth_Map_Preprocessor:
@classmethod
def INPUT_TYPES(s):
types = create_node_input_types(mask_bbox_padding=("INT", {"default": 30, "min": 0, "max": 100}))
types["optional"].update(
{
"mask_type": (["based_on_depth", "tight_bboxes", "original"], {"default": "based_on_depth"}),
"mask_expand": ("INT", {"default": 5, "min": -MAX_RESOLUTION, "max": MAX_RESOLUTION, "step": 1}),
"rand_seed": ("INT", {"default": 88, "min": 0, "max": 0xffffffffffffffff})
}
)
return types
RETURN_TYPES = ("IMAGE", "MASK")
RETURN_NAMES = ("IMAGE", "INPAINTING_MASK")
FUNCTION = "execute"
CATEGORY = "ControlNet Preprocessors/Normal and Depth Estimators"
def execute(self, image, mask_bbox_padding=30, mask_type="based_on_depth", mask_expand=5, resolution=512, rand_seed=88, **kwargs):
install_deps()
from controlnet_aux.mesh_graphormer import MeshGraphormerDetector
model = MeshGraphormerDetector.from_pretrained().to(model_management.get_torch_device())
depth_map_list = []
mask_list = []
for single_image in image:
np_image = np.asarray(single_image.cpu() * 255., dtype=np.uint8)
depth_map, mask, info = model(np_image, output_type="np", detect_resolution=resolution, mask_bbox_padding=mask_bbox_padding, seed=rand_seed)
if mask_type == "based_on_depth":
H, W = mask.shape[:2]
mask = cv2.resize(depth_map.copy(), (W, H))
mask[mask > 0] = 255
elif mask_type == "tight_bboxes":
mask = np.zeros_like(mask)
hand_bboxes = info["abs_boxes"]
for hand_bbox in hand_bboxes:
x_min, x_max, y_min, y_max = hand_bbox
mask[y_min:y_max+1, x_min:x_max+1, :] = 255 #HWC
mask = mask[:, :, :1]
depth_map_list.append(torch.from_numpy(depth_map.astype(np.float32) / 255.0))
mask_list.append(torch.from_numpy(mask.astype(np.float32) / 255.0))
depth_maps, masks = torch.stack(depth_map_list, dim=0), rearrange(torch.stack(mask_list, dim=0), "n h w 1 -> n 1 h w")
return depth_maps, expand_mask(masks, mask_expand, tapered_corners=True)