File size: 106,998 Bytes
cd15502
1
{"repo": "MultimediaTechLab/YOLO", "n_pairs": 46, "version": "v2_function_scoped", "contexts": {"tests/test_model/test_yolo.py::66": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py"], "used_names": ["YOLO"], "enclosing_function": "test_model_basic_status", "extracted_code": "# Source: yolo/model/yolo.py\nclass YOLO(nn.Module):\n    \"\"\"\n    A preliminary YOLO (You Only Look Once) model class still under development.\n\n    Parameters:\n        model_cfg: Configuration for the YOLO model. Expected to define the layers,\n                   parameters, and any other relevant configuration details.\n    \"\"\"\n\n    def __init__(self, model_cfg: ModelConfig, class_num: int = 80):\n        super(YOLO, self).__init__()\n        self.num_classes = class_num\n        self.layer_map = get_layer_map()  # Get the map Dict[str: Module]\n        self.model: List[YOLOLayer] = nn.ModuleList()\n        self.reg_max = getattr(model_cfg.anchor, \"reg_max\", 16)\n        self.build_model(model_cfg.model)\n\n    def build_model(self, model_arch: Dict[str, List[Dict[str, Dict[str, Dict]]]]):\n        self.layer_index = {}\n        output_dim, layer_idx = [3], 1\n        logger.info(f\":tractor: Building YOLO\")\n        for arch_name in model_arch:\n            if model_arch[arch_name]:\n                logger.info(f\"  :building_construction:  Building {arch_name}\")\n            for layer_idx, layer_spec in enumerate(model_arch[arch_name], start=layer_idx):\n                layer_type, layer_info = next(iter(layer_spec.items()))\n                layer_args = layer_info.get(\"args\", {})\n\n                # Get input source\n                source = self.get_source_idx(layer_info.get(\"source\", -1), layer_idx)\n\n                # Find in channels\n                if any(module in layer_type for module in [\"Conv\", \"ELAN\", \"ADown\", \"AConv\", \"CBLinear\"]):\n                    layer_args[\"in_channels\"] = output_dim[source]\n                if any(module in layer_type for module in [\"Detection\", \"Segmentation\", \"Classification\"]):\n                    if isinstance(source, list):\n                        layer_args[\"in_channels\"] = [output_dim[idx] for idx in source]\n                    else:\n                        layer_args[\"in_channel\"] = output_dim[source]\n                    layer_args[\"num_classes\"] = self.num_classes\n                    layer_args[\"reg_max\"] = self.reg_max\n\n                # create layers\n                layer = self.create_layer(layer_type, source, layer_info, **layer_args)\n                self.model.append(layer)\n\n                if layer.tags:\n                    if layer.tags in self.layer_index:\n                        raise ValueError(f\"Duplicate tag '{layer_info['tags']}' found.\")\n                    self.layer_index[layer.tags] = layer_idx\n\n                out_channels = self.get_out_channels(layer_type, layer_args, output_dim, source)\n                output_dim.append(out_channels)\n                setattr(layer, \"out_c\", out_channels)\n            layer_idx += 1\n\n    def forward(self, x, external: Optional[Dict] = None, shortcut: Optional[str] = None):\n        y = {0: x, **(external or {})}\n        output = dict()\n        for index, layer in enumerate(self.model, start=1):\n            if isinstance(layer.source, list):\n                model_input = [y[idx] for idx in layer.source]\n            else:\n                model_input = y[layer.source]\n\n            external_input = {source_name: y[source_name] for source_name in layer.external}\n\n            x = layer(model_input, **external_input)\n            y[-1] = x\n            if layer.usable:\n                y[index] = x\n            if layer.output:\n                output[layer.tags] = x\n                if layer.tags == shortcut:\n                    return output\n        return output\n\n    def get_out_channels(self, layer_type: str, layer_args: dict, output_dim: list, source: Union[int, list]):\n        if hasattr(layer_args, \"out_channels\"):\n            return layer_args[\"out_channels\"]\n        if layer_type == \"CBFuse\":\n            return output_dim[source[-1]]\n        if isinstance(source, int):\n            return output_dim[source]\n        if isinstance(source, list):\n            return sum(output_dim[idx] for idx in source)\n\n    def get_source_idx(self, source: Union[ListConfig, str, int], layer_idx: int):\n        if isinstance(source, ListConfig):\n            return [self.get_source_idx(index, layer_idx) for index in source]\n        if isinstance(source, str):\n            source = self.layer_index[source]\n        if source < -1:\n            source += layer_idx\n        if source > 0:  # Using Previous Layer's Output\n            self.model[source - 1].usable = True\n        return source\n\n    def create_layer(self, layer_type: str, source: Union[int, list], layer_info: Dict, **kwargs) -> YOLOLayer:\n        if layer_type in self.layer_map:\n            layer = self.layer_map[layer_type](**kwargs)\n            setattr(layer, \"layer_type\", layer_type)\n            setattr(layer, \"source\", source)\n            setattr(layer, \"in_c\", kwargs.get(\"in_channels\", None))\n            setattr(layer, \"output\", layer_info.get(\"output\", False))\n            setattr(layer, \"tags\", layer_info.get(\"tags\", None))\n            setattr(layer, \"external\", layer_info.get(\"external\", []))\n            setattr(layer, \"usable\", 0)\n            return layer\n        else:\n            raise ValueError(f\"Unsupported layer type: {layer_type}\")\n\n    def save_load_weights(self, weights: Union[Path, OrderedDict]):\n        \"\"\"\n        Update the model's weights with the provided weights.\n\n        args:\n            weights: A OrderedDict containing the new weights.\n        \"\"\"\n        if isinstance(weights, Path):\n            weights = torch.load(weights, map_location=torch.device(\"cpu\"), weights_only=False)\n        if \"state_dict\" in weights:\n            weights = {name.removeprefix(\"model.model.\"): key for name, key in weights[\"state_dict\"].items()}\n        model_state_dict = self.model.state_dict()\n\n        # TODO1: autoload old version weight\n        # TODO2: weight transform if num_class difference\n\n        error_dict = {\"Mismatch\": set(), \"Not Found\": set()}\n        for model_key, model_weight in model_state_dict.items():\n            if model_key not in weights:\n                error_dict[\"Not Found\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            if model_weight.shape != weights[model_key].shape:\n                error_dict[\"Mismatch\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            model_state_dict[model_key] = weights[model_key]\n\n        for error_name, error_set in error_dict.items():\n            error_dict = dict()\n            for layer_idx, *layer_name in error_set:\n                if layer_idx not in error_dict:\n                    error_dict[layer_idx] = [\".\".join(layer_name)]\n                else:\n                    error_dict[layer_idx].append(\".\".join(layer_name))\n            for layer_idx, layer_name in error_dict.items():\n                layer_name.sort()\n                logger.warning(f\":warning: Weight {error_name} for Layer {layer_idx}: {', '.join(layer_name)}\")\n\n        self.model.load_state_dict(model_state_dict)", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 6918}, "tests/test_tools/test_data_loader.py::48": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_validation_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::33": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_training_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_model/test_module.py::51": {"resolved_imports": ["yolo/model/module.py"], "used_names": ["CBLinear", "torch"], "enclosing_function": "test_cblinear", "extracted_code": "# Source: yolo/model/module.py\nclass CBLinear(nn.Module):\n    \"\"\"Convolutional block that outputs multiple feature maps split along the channel dimension.\"\"\"\n\n    def __init__(self, in_channels: int, out_channels: List[int], kernel_size: int = 1, **kwargs):\n        super(CBLinear, self).__init__()\n        kwargs.setdefault(\"padding\", auto_pad(kernel_size, **kwargs))\n        self.conv = nn.Conv2d(in_channels, sum(out_channels), kernel_size, **kwargs)\n        self.out_channels = list(out_channels)\n\n    def forward(self, x: Tensor) -> Tuple[Tensor]:\n        x = self.conv(x)\n        return x.split(self.out_channels, dim=1)", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 626}, "tests/test_utils/test_bounding_box_utils.py::37": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["allclose", "calculate_iou", "tensor"], "enclosing_function": "test_calculate_iou_2d", "extracted_code": "# Source: yolo/utils/bounding_box_utils.py\ndef calculate_iou(bbox1, bbox2, metrics=\"iou\") -> Tensor:\n    metrics = metrics.lower()\n    EPS = 1e-7\n    dtype = bbox1.dtype\n    bbox1 = bbox1.to(torch.float32)\n    bbox2 = bbox2.to(torch.float32)\n\n    # Expand dimensions if necessary\n    if bbox1.ndim == 2 and bbox2.ndim == 2:\n        bbox1 = bbox1.unsqueeze(1)  # (Ax4) -> (Ax1x4)\n        bbox2 = bbox2.unsqueeze(0)  # (Bx4) -> (1xBx4)\n    elif bbox1.ndim == 3 and bbox2.ndim == 3:\n        bbox1 = bbox1.unsqueeze(2)  # (BZxAx4) -> (BZxAx1x4)\n        bbox2 = bbox2.unsqueeze(1)  # (BZxBx4) -> (BZx1xBx4)\n\n    # Calculate intersection coordinates\n    xmin_inter = torch.max(bbox1[..., 0], bbox2[..., 0])\n    ymin_inter = torch.max(bbox1[..., 1], bbox2[..., 1])\n    xmax_inter = torch.min(bbox1[..., 2], bbox2[..., 2])\n    ymax_inter = torch.min(bbox1[..., 3], bbox2[..., 3])\n\n    # Calculate intersection area\n    intersection_area = torch.clamp(xmax_inter - xmin_inter, min=0) * torch.clamp(ymax_inter - ymin_inter, min=0)\n\n    # Calculate area of each bbox\n    area_bbox1 = (bbox1[..., 2] - bbox1[..., 0]) * (bbox1[..., 3] - bbox1[..., 1])\n    area_bbox2 = (bbox2[..., 2] - bbox2[..., 0]) * (bbox2[..., 3] - bbox2[..., 1])\n\n    # Calculate union area\n    union_area = area_bbox1 + area_bbox2 - intersection_area\n\n    # Calculate IoU\n    iou = intersection_area / (union_area + EPS)\n    if metrics == \"iou\":\n        return iou.to(dtype)\n\n    # Calculate centroid distance\n    cx1 = (bbox1[..., 2] + bbox1[..., 0]) / 2\n    cy1 = (bbox1[..., 3] + bbox1[..., 1]) / 2\n    cx2 = (bbox2[..., 2] + bbox2[..., 0]) / 2\n    cy2 = (bbox2[..., 3] + bbox2[..., 1]) / 2\n    cent_dis = (cx1 - cx2) ** 2 + (cy1 - cy2) ** 2\n\n    # Calculate diagonal length of the smallest enclosing box\n    c_x = torch.max(bbox1[..., 2], bbox2[..., 2]) - torch.min(bbox1[..., 0], bbox2[..., 0])\n    c_y = torch.max(bbox1[..., 3], bbox2[..., 3]) - torch.min(bbox1[..., 1], bbox2[..., 1])\n    diag_dis = c_x**2 + c_y**2 + EPS\n\n    diou = iou - (cent_dis / diag_dis)\n    if metrics == \"diou\":\n        return diou.to(dtype)\n\n    # Compute aspect ratio penalty term\n    arctan = torch.atan((bbox1[..., 2] - bbox1[..., 0]) / (bbox1[..., 3] - bbox1[..., 1] + EPS)) - torch.atan(\n        (bbox2[..., 2] - bbox2[..., 0]) / (bbox2[..., 3] - bbox2[..., 1] + EPS)\n    )\n    v = (4 / (math.pi**2)) * (arctan**2)\n    with torch.no_grad():\n        alpha = v / (v - iou + 1 + EPS)\n    # Compute CIoU\n    ciou = diou - alpha * v\n    return ciou.to(dtype)", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 2502}, "tests/test_tools/test_data_loader.py::27": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["Config", "Path", "create_dataloader"], "enclosing_function": "test_create_dataloader_cache", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/data_loader.py\ndef create_dataloader(data_cfg: DataConfig, dataset_cfg: DatasetConfig, task: str = \"train\"):\n    if task == \"inference\":\n        return StreamDataLoader(data_cfg)\n\n    if getattr(dataset_cfg, \"auto_download\", False):\n        prepare_dataset(dataset_cfg, task)\n    dataset = YoloDataset(data_cfg, dataset_cfg, task)\n\n    return DataLoader(\n        dataset,\n        batch_size=data_cfg.batch_size,\n        num_workers=data_cfg.cpu_num,\n        pin_memory=data_cfg.pin_memory,\n        collate_fn=collate_fn,\n    )", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 971}, "tests/test_model/test_module.py::43": {"resolved_imports": ["yolo/model/module.py"], "used_names": ["ADown", "torch"], "enclosing_function": "test_adown", "extracted_code": "# Source: yolo/model/module.py\nclass ADown(nn.Module):\n    \"\"\"Downsampling module combining average and max pooling with convolution for feature reduction.\"\"\"\n\n    def __init__(self, in_channels: int, out_channels: int):\n        super().__init__()\n        half_in_channels = in_channels // 2\n        half_out_channels = out_channels // 2\n        mid_layer = {\"kernel_size\": 3, \"stride\": 2}\n        self.avg_pool = Pool(\"avg\", kernel_size=2, stride=1)\n        self.conv1 = Conv(half_in_channels, half_out_channels, **mid_layer)\n        self.max_pool = Pool(\"max\", **mid_layer)\n        self.conv2 = Conv(half_in_channels, half_out_channels, kernel_size=1)\n\n    def forward(self, x: Tensor) -> Tensor:\n        x = self.avg_pool(x)\n        x1, x2 = x.chunk(2, dim=1)\n        x1 = self.conv1(x1)\n        x2 = self.max_pool(x2)\n        x2 = self.conv2(x2)\n        return torch.cat((x1, x2), dim=1)", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 891}, "tests/test_utils/test_bounding_box_utils.py::145": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["Anc2Box", "AnchorConfig", "Config", "create_model", "torch"], "enclosing_function": "test_anc2box_autoanchor", "extracted_code": "# Source: yolo/__init__.py\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n    YOLORichProgressBar,\n\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n    \"AugmentationComposer\",\n    \"ImageLogger\",\n\n\n# Source: yolo/config/config.py\nclass AnchorConfig:\n    strides: List[int]\n    reg_max: Optional[int]\n    anchor_num: Optional[int]\n    anchor: List[List[int]]\n\n\n# Source: yolo/utils/bounding_box_utils.py\nclass Anc2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        self.head_num = len(anchor_cfg.anchor)\n        self.anchor_grids = self.generate_anchors(image_size)\n        self.anchor_scale = tensor(anchor_cfg.anchor, device=device).view(self.head_num, 1, -1, 1, 1, 2)\n        self.anchor_num = self.anchor_scale.size(2)\n        self.class_num = model.num_classes\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        dummy_input = torch.zeros(1, 3, H, W).to(self.device)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head.shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def generate_anchors(self, image_size: List[int]):\n        anchor_grids = []\n        for stride in self.strides:\n            W, H = image_size[0] // stride, image_size[1] // stride\n            anchor_h, anchor_w = torch.meshgrid([torch.arange(H), torch.arange(W)], indexing=\"ij\")\n            anchor_grid = torch.stack((anchor_w, anchor_h), 2).view((1, 1, H, W, 2)).float().to(self.device)\n            anchor_grids.append(anchor_grid)\n        return anchor_grids\n\n    def update(self, image_size):\n        self.anchor_grids = self.generate_anchors(image_size)\n\n    def __call__(self, predicts: List[Tensor]):\n        preds_box, preds_cls, preds_cnf = [], [], []\n        for layer_idx, predict in enumerate(predicts):\n            predict = rearrange(predict, \"B (L C) h w -> B L h w C\", L=self.anchor_num)\n            pred_box, pred_cnf, pred_cls = predict.split((4, 1, self.class_num), dim=-1)\n            pred_box = pred_box.sigmoid()\n            pred_box[..., 0:2] = (pred_box[..., 0:2] * 2.0 - 0.5 + self.anchor_grids[layer_idx]) * self.strides[\n                layer_idx\n            ]\n            pred_box[..., 2:4] = (pred_box[..., 2:4] * 2) ** 2 * self.anchor_scale[layer_idx]\n            preds_box.append(rearrange(pred_box, \"B L h w A -> B (L h w) A\"))\n            preds_cls.append(rearrange(pred_cls, \"B L h w C -> B (L h w) C\"))\n            preds_cnf.append(rearrange(pred_cnf, \"B L h w C -> B (L h w) C\"))\n\n        preds_box = torch.concat(preds_box, dim=1)\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_cnf = torch.concat(preds_cnf, dim=1)\n\n        preds_box = transform_bbox(preds_box, \"xycwh -> xyxy\")\n        return preds_cls, None, preds_box, preds_cnf.sigmoid()", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4808}, "tests/test_utils/test_bounding_box_utils.py::45": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["allclose", "calculate_iou", "tensor"], "enclosing_function": "test_calculate_iou_3d", "extracted_code": "# Source: yolo/utils/bounding_box_utils.py\ndef calculate_iou(bbox1, bbox2, metrics=\"iou\") -> Tensor:\n    metrics = metrics.lower()\n    EPS = 1e-7\n    dtype = bbox1.dtype\n    bbox1 = bbox1.to(torch.float32)\n    bbox2 = bbox2.to(torch.float32)\n\n    # Expand dimensions if necessary\n    if bbox1.ndim == 2 and bbox2.ndim == 2:\n        bbox1 = bbox1.unsqueeze(1)  # (Ax4) -> (Ax1x4)\n        bbox2 = bbox2.unsqueeze(0)  # (Bx4) -> (1xBx4)\n    elif bbox1.ndim == 3 and bbox2.ndim == 3:\n        bbox1 = bbox1.unsqueeze(2)  # (BZxAx4) -> (BZxAx1x4)\n        bbox2 = bbox2.unsqueeze(1)  # (BZxBx4) -> (BZx1xBx4)\n\n    # Calculate intersection coordinates\n    xmin_inter = torch.max(bbox1[..., 0], bbox2[..., 0])\n    ymin_inter = torch.max(bbox1[..., 1], bbox2[..., 1])\n    xmax_inter = torch.min(bbox1[..., 2], bbox2[..., 2])\n    ymax_inter = torch.min(bbox1[..., 3], bbox2[..., 3])\n\n    # Calculate intersection area\n    intersection_area = torch.clamp(xmax_inter - xmin_inter, min=0) * torch.clamp(ymax_inter - ymin_inter, min=0)\n\n    # Calculate area of each bbox\n    area_bbox1 = (bbox1[..., 2] - bbox1[..., 0]) * (bbox1[..., 3] - bbox1[..., 1])\n    area_bbox2 = (bbox2[..., 2] - bbox2[..., 0]) * (bbox2[..., 3] - bbox2[..., 1])\n\n    # Calculate union area\n    union_area = area_bbox1 + area_bbox2 - intersection_area\n\n    # Calculate IoU\n    iou = intersection_area / (union_area + EPS)\n    if metrics == \"iou\":\n        return iou.to(dtype)\n\n    # Calculate centroid distance\n    cx1 = (bbox1[..., 2] + bbox1[..., 0]) / 2\n    cy1 = (bbox1[..., 3] + bbox1[..., 1]) / 2\n    cx2 = (bbox2[..., 2] + bbox2[..., 0]) / 2\n    cy2 = (bbox2[..., 3] + bbox2[..., 1]) / 2\n    cent_dis = (cx1 - cx2) ** 2 + (cy1 - cy2) ** 2\n\n    # Calculate diagonal length of the smallest enclosing box\n    c_x = torch.max(bbox1[..., 2], bbox2[..., 2]) - torch.min(bbox1[..., 0], bbox2[..., 0])\n    c_y = torch.max(bbox1[..., 3], bbox2[..., 3]) - torch.min(bbox1[..., 1], bbox2[..., 1])\n    diag_dis = c_x**2 + c_y**2 + EPS\n\n    diou = iou - (cent_dis / diag_dis)\n    if metrics == \"diou\":\n        return diou.to(dtype)\n\n    # Compute aspect ratio penalty term\n    arctan = torch.atan((bbox1[..., 2] - bbox1[..., 0]) / (bbox1[..., 3] - bbox1[..., 1] + EPS)) - torch.atan(\n        (bbox2[..., 2] - bbox2[..., 0]) / (bbox2[..., 3] - bbox2[..., 1] + EPS)\n    )\n    v = (4 / (math.pi**2)) * (arctan**2)\n    with torch.no_grad():\n        alpha = v / (v - iou + 1 + EPS)\n    # Compute CIoU\n    ciou = diou - alpha * v\n    return ciou.to(dtype)", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 2502}, "tests/test_tools/test_dataset_preparation.py::25": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/dataset_preparation.py"], "used_names": ["Config", "Path", "os", "prepare_dataset", "shutil"], "enclosing_function": "test_prepare_dataset", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/dataset_preparation.py\ndef prepare_dataset(dataset_cfg: DatasetConfig, task: str):\n    \"\"\"\n    Prepares dataset by downloading and unzipping if necessary.\n    \"\"\"\n    # TODO: do EDA of dataset\n    data_dir = Path(dataset_cfg.path)\n    for data_type, settings in dataset_cfg.auto_download.items():\n        base_url = settings[\"base_url\"]\n        for dataset_type, dataset_args in settings.items():\n            if dataset_type != \"annotations\" and dataset_cfg.get(task, task) != dataset_type:\n                continue\n            file_name = f\"{dataset_args.get('file_name', dataset_type)}.zip\"\n            url = f\"{base_url}{file_name}\"\n            local_zip_path = data_dir / file_name\n            extract_to = data_dir / data_type if data_type != \"annotations\" else data_dir\n            final_place = extract_to / dataset_type\n\n            final_place.mkdir(parents=True, exist_ok=True)\n            if check_files(final_place, dataset_args.get(\"file_num\")):\n                logger.info(f\":white_check_mark: Dataset {dataset_type: <12} already verified.\")\n                continue\n\n            if not local_zip_path.exists():\n                download_file(url, local_zip_path)\n            unzip_file(local_zip_path, extract_to)\n\n            if not check_files(final_place, dataset_args.get(\"file_num\")):\n                logger.error(f\"Error verifying the {dataset_type} dataset after extraction.\")", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 1842}, "tests/test_model/test_module.py::29": {"resolved_imports": ["yolo/model/module.py"], "used_names": ["Pool", "torch"], "enclosing_function": "test_pool_max", "extracted_code": "# Source: yolo/model/module.py\nclass Pool(nn.Module):\n    \"\"\"A generic pooling block supporting 'max' and 'avg' pooling methods.\"\"\"\n\n    def __init__(self, method: str = \"max\", kernel_size: _size_2_t = 2, **kwargs):\n        super().__init__()\n        kwargs.setdefault(\"padding\", auto_pad(kernel_size, **kwargs))\n        pool_classes = {\"max\": nn.MaxPool2d, \"avg\": nn.AvgPool2d}\n        self.pool = pool_classes[method.lower()](kernel_size=kernel_size, **kwargs)\n\n    def forward(self, x: Tensor) -> Tensor:\n        return self.pool(x)", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 535}, "tests/test_tools/test_data_loader.py::45": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_validation_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_solver.py::78": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py", "yolo/tools/data_loader.py", "yolo/tools/solver.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["TrainModel", "Trainer", "YOLO"], "enclosing_function": "test_model_trainer_initialization", "extracted_code": "# Source: yolo/model/yolo.py\nclass YOLO(nn.Module):\n    \"\"\"\n    A preliminary YOLO (You Only Look Once) model class still under development.\n\n    Parameters:\n        model_cfg: Configuration for the YOLO model. Expected to define the layers,\n                   parameters, and any other relevant configuration details.\n    \"\"\"\n\n    def __init__(self, model_cfg: ModelConfig, class_num: int = 80):\n        super(YOLO, self).__init__()\n        self.num_classes = class_num\n        self.layer_map = get_layer_map()  # Get the map Dict[str: Module]\n        self.model: List[YOLOLayer] = nn.ModuleList()\n        self.reg_max = getattr(model_cfg.anchor, \"reg_max\", 16)\n        self.build_model(model_cfg.model)\n\n    def build_model(self, model_arch: Dict[str, List[Dict[str, Dict[str, Dict]]]]):\n        self.layer_index = {}\n        output_dim, layer_idx = [3], 1\n        logger.info(f\":tractor: Building YOLO\")\n        for arch_name in model_arch:\n            if model_arch[arch_name]:\n                logger.info(f\"  :building_construction:  Building {arch_name}\")\n            for layer_idx, layer_spec in enumerate(model_arch[arch_name], start=layer_idx):\n                layer_type, layer_info = next(iter(layer_spec.items()))\n                layer_args = layer_info.get(\"args\", {})\n\n                # Get input source\n                source = self.get_source_idx(layer_info.get(\"source\", -1), layer_idx)\n\n                # Find in channels\n                if any(module in layer_type for module in [\"Conv\", \"ELAN\", \"ADown\", \"AConv\", \"CBLinear\"]):\n                    layer_args[\"in_channels\"] = output_dim[source]\n                if any(module in layer_type for module in [\"Detection\", \"Segmentation\", \"Classification\"]):\n                    if isinstance(source, list):\n                        layer_args[\"in_channels\"] = [output_dim[idx] for idx in source]\n                    else:\n                        layer_args[\"in_channel\"] = output_dim[source]\n                    layer_args[\"num_classes\"] = self.num_classes\n                    layer_args[\"reg_max\"] = self.reg_max\n\n                # create layers\n                layer = self.create_layer(layer_type, source, layer_info, **layer_args)\n                self.model.append(layer)\n\n                if layer.tags:\n                    if layer.tags in self.layer_index:\n                        raise ValueError(f\"Duplicate tag '{layer_info['tags']}' found.\")\n                    self.layer_index[layer.tags] = layer_idx\n\n                out_channels = self.get_out_channels(layer_type, layer_args, output_dim, source)\n                output_dim.append(out_channels)\n                setattr(layer, \"out_c\", out_channels)\n            layer_idx += 1\n\n    def forward(self, x, external: Optional[Dict] = None, shortcut: Optional[str] = None):\n        y = {0: x, **(external or {})}\n        output = dict()\n        for index, layer in enumerate(self.model, start=1):\n            if isinstance(layer.source, list):\n                model_input = [y[idx] for idx in layer.source]\n            else:\n                model_input = y[layer.source]\n\n            external_input = {source_name: y[source_name] for source_name in layer.external}\n\n            x = layer(model_input, **external_input)\n            y[-1] = x\n            if layer.usable:\n                y[index] = x\n            if layer.output:\n                output[layer.tags] = x\n                if layer.tags == shortcut:\n                    return output\n        return output\n\n    def get_out_channels(self, layer_type: str, layer_args: dict, output_dim: list, source: Union[int, list]):\n        if hasattr(layer_args, \"out_channels\"):\n            return layer_args[\"out_channels\"]\n        if layer_type == \"CBFuse\":\n            return output_dim[source[-1]]\n        if isinstance(source, int):\n            return output_dim[source]\n        if isinstance(source, list):\n            return sum(output_dim[idx] for idx in source)\n\n    def get_source_idx(self, source: Union[ListConfig, str, int], layer_idx: int):\n        if isinstance(source, ListConfig):\n            return [self.get_source_idx(index, layer_idx) for index in source]\n        if isinstance(source, str):\n            source = self.layer_index[source]\n        if source < -1:\n            source += layer_idx\n        if source > 0:  # Using Previous Layer's Output\n            self.model[source - 1].usable = True\n        return source\n\n    def create_layer(self, layer_type: str, source: Union[int, list], layer_info: Dict, **kwargs) -> YOLOLayer:\n        if layer_type in self.layer_map:\n            layer = self.layer_map[layer_type](**kwargs)\n            setattr(layer, \"layer_type\", layer_type)\n            setattr(layer, \"source\", source)\n            setattr(layer, \"in_c\", kwargs.get(\"in_channels\", None))\n            setattr(layer, \"output\", layer_info.get(\"output\", False))\n            setattr(layer, \"tags\", layer_info.get(\"tags\", None))\n            setattr(layer, \"external\", layer_info.get(\"external\", []))\n            setattr(layer, \"usable\", 0)\n            return layer\n        else:\n            raise ValueError(f\"Unsupported layer type: {layer_type}\")\n\n    def save_load_weights(self, weights: Union[Path, OrderedDict]):\n        \"\"\"\n        Update the model's weights with the provided weights.\n\n        args:\n            weights: A OrderedDict containing the new weights.\n        \"\"\"\n        if isinstance(weights, Path):\n            weights = torch.load(weights, map_location=torch.device(\"cpu\"), weights_only=False)\n        if \"state_dict\" in weights:\n            weights = {name.removeprefix(\"model.model.\"): key for name, key in weights[\"state_dict\"].items()}\n        model_state_dict = self.model.state_dict()\n\n        # TODO1: autoload old version weight\n        # TODO2: weight transform if num_class difference\n\n        error_dict = {\"Mismatch\": set(), \"Not Found\": set()}\n        for model_key, model_weight in model_state_dict.items():\n            if model_key not in weights:\n                error_dict[\"Not Found\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            if model_weight.shape != weights[model_key].shape:\n                error_dict[\"Mismatch\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            model_state_dict[model_key] = weights[model_key]\n\n        for error_name, error_set in error_dict.items():\n            error_dict = dict()\n            for layer_idx, *layer_name in error_set:\n                if layer_idx not in error_dict:\n                    error_dict[layer_idx] = [\".\".join(layer_name)]\n                else:\n                    error_dict[layer_idx].append(\".\".join(layer_name))\n            for layer_idx, layer_name in error_dict.items():\n                layer_name.sort()\n                logger.warning(f\":warning: Weight {error_name} for Layer {layer_idx}: {', '.join(layer_name)}\")\n\n        self.model.load_state_dict(model_state_dict)\n\n\n# Source: yolo/tools/solver.py\nclass TrainModel(ValidateModel):\n    def __init__(self, cfg: Config):\n        super().__init__(cfg)\n        self.cfg = cfg\n        self.train_loader = create_dataloader(self.cfg.task.data, self.cfg.dataset, self.cfg.task.task)\n\n    def setup(self, stage):\n        super().setup(stage)\n        self.loss_fn = create_loss_function(self.cfg, self.vec2box)\n\n    def train_dataloader(self):\n        return self.train_loader\n\n    def on_train_epoch_start(self):\n        self.trainer.optimizers[0].next_epoch(\n            ceil(len(self.train_loader) / self.trainer.world_size), self.current_epoch\n        )\n        self.vec2box.update(self.cfg.image_size)\n\n    def training_step(self, batch, batch_idx):\n        lr_dict = self.trainer.optimizers[0].next_batch()\n        batch_size, images, targets, *_ = batch\n        predicts = self(images)\n        aux_predicts = self.vec2box(predicts[\"AUX\"])\n        main_predicts = self.vec2box(predicts[\"Main\"])\n        loss, loss_item = self.loss_fn(aux_predicts, main_predicts, targets)\n        self.log_dict(\n            loss_item,\n            prog_bar=True,\n            on_epoch=True,\n            batch_size=batch_size,\n            rank_zero_only=True,\n        )\n        self.log_dict(lr_dict, prog_bar=False, logger=True, on_epoch=False, rank_zero_only=True)\n        return loss * batch_size\n\n    def configure_optimizers(self):\n        optimizer = create_optimizer(self.model, self.cfg.task.optimizer)\n        scheduler = create_scheduler(optimizer, self.cfg.task.scheduler)\n        return [optimizer], [scheduler]", "n_imports_parsed": 11, "n_files_resolved": 5, "n_chars_extracted": 8503}, "tests/test_utils/test_bounding_box_utils.py::123": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["Config", "Vec2Box", "compose", "create_model", "initialize", "torch"], "enclosing_function": "test_vec2box_autoanchor", "extracted_code": "# Source: yolo/__init__.py\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n    YOLORichProgressBar,\n\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n    \"AugmentationComposer\",\n    \"ImageLogger\",\n\n\n# Source: yolo/utils/bounding_box_utils.py\nclass Vec2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(device), scaler.to(device)\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        # TODO: need accelerate dummy test\n        dummy_input = torch.zeros(1, 3, H, W)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head[2].shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def update(self, image_size):\n        \"\"\"\n        image_size: W, H\n        \"\"\"\n        if self.image_size == image_size:\n            return\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(self.device), scaler.to(self.device)\n\n    def __call__(self, predicts):\n        preds_cls, preds_anc, preds_box = [], [], []\n        for layer_output in predicts:\n            pred_cls, pred_anc, pred_box = layer_output\n            preds_cls.append(rearrange(pred_cls, \"B C h w -> B (h w) C\"))\n            preds_anc.append(rearrange(pred_anc, \"B A R h w -> B (h w) R A\"))\n            preds_box.append(rearrange(pred_box, \"B X h w -> B (h w) X\"))\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_anc = torch.concat(preds_anc, dim=1)\n        preds_box = torch.concat(preds_box, dim=1)\n\n        pred_LTRB = preds_box * self.scaler.view(1, -1, 1)\n        lt, rb = pred_LTRB.chunk(2, dim=-1)\n        preds_box = torch.cat([self.anchor_grid - lt, self.anchor_grid + rb], dim=-1)\n        return preds_cls, preds_anc, preds_box", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4002}, "tests/test_utils/test_bounding_box_utils.py::140": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["Anc2Box", "AnchorConfig", "Config", "create_model", "torch"], "enclosing_function": "test_anc2box_autoanchor", "extracted_code": "# Source: yolo/__init__.py\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n    YOLORichProgressBar,\n\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n    \"AugmentationComposer\",\n    \"ImageLogger\",\n\n\n# Source: yolo/config/config.py\nclass AnchorConfig:\n    strides: List[int]\n    reg_max: Optional[int]\n    anchor_num: Optional[int]\n    anchor: List[List[int]]\n\n\n# Source: yolo/utils/bounding_box_utils.py\nclass Anc2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        self.head_num = len(anchor_cfg.anchor)\n        self.anchor_grids = self.generate_anchors(image_size)\n        self.anchor_scale = tensor(anchor_cfg.anchor, device=device).view(self.head_num, 1, -1, 1, 1, 2)\n        self.anchor_num = self.anchor_scale.size(2)\n        self.class_num = model.num_classes\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        dummy_input = torch.zeros(1, 3, H, W).to(self.device)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head.shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def generate_anchors(self, image_size: List[int]):\n        anchor_grids = []\n        for stride in self.strides:\n            W, H = image_size[0] // stride, image_size[1] // stride\n            anchor_h, anchor_w = torch.meshgrid([torch.arange(H), torch.arange(W)], indexing=\"ij\")\n            anchor_grid = torch.stack((anchor_w, anchor_h), 2).view((1, 1, H, W, 2)).float().to(self.device)\n            anchor_grids.append(anchor_grid)\n        return anchor_grids\n\n    def update(self, image_size):\n        self.anchor_grids = self.generate_anchors(image_size)\n\n    def __call__(self, predicts: List[Tensor]):\n        preds_box, preds_cls, preds_cnf = [], [], []\n        for layer_idx, predict in enumerate(predicts):\n            predict = rearrange(predict, \"B (L C) h w -> B L h w C\", L=self.anchor_num)\n            pred_box, pred_cnf, pred_cls = predict.split((4, 1, self.class_num), dim=-1)\n            pred_box = pred_box.sigmoid()\n            pred_box[..., 0:2] = (pred_box[..., 0:2] * 2.0 - 0.5 + self.anchor_grids[layer_idx]) * self.strides[\n                layer_idx\n            ]\n            pred_box[..., 2:4] = (pred_box[..., 2:4] * 2) ** 2 * self.anchor_scale[layer_idx]\n            preds_box.append(rearrange(pred_box, \"B L h w A -> B (L h w) A\"))\n            preds_cls.append(rearrange(pred_cls, \"B L h w C -> B (L h w) C\"))\n            preds_cnf.append(rearrange(pred_cnf, \"B L h w C -> B (L h w) C\"))\n\n        preds_box = torch.concat(preds_box, dim=1)\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_cnf = torch.concat(preds_cnf, dim=1)\n\n        preds_box = transform_bbox(preds_box, \"xycwh -> xyxy\")\n        return preds_cls, None, preds_box, preds_cnf.sigmoid()", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4808}, "tests/test_utils/test_module_utils.py::41": {"resolved_imports": ["yolo/utils/module_utils.py"], "used_names": ["create_activation_function", "pytest"], "enclosing_function": "test_get_activation_invalid", "extracted_code": "# Source: yolo/utils/module_utils.py\ndef create_activation_function(activation: str) -> nn.Module:\n    \"\"\"\n    Retrieves an activation function from the PyTorch nn module based on its name, case-insensitively.\n    \"\"\"\n    if not activation or activation.lower() in [\"false\", \"none\"]:\n        return nn.Identity()\n\n    activation_map = {\n        name.lower(): obj\n        for name, obj in nn.modules.activation.__dict__.items()\n        if isinstance(obj, type) and issubclass(obj, nn.Module)\n    }\n    if activation.lower() in activation_map:\n        return activation_map[activation.lower()](inplace=True)\n    else:\n        raise ValueError(f\"Activation function '{activation}' is not found in torch.nn\")", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 704}, "tests/test_utils/test_module_utils.py::28": {"resolved_imports": ["yolo/utils/module_utils.py"], "used_names": ["auto_pad", "pytest"], "enclosing_function": "test_auto_pad", "extracted_code": "# Source: yolo/utils/module_utils.py\ndef auto_pad(kernel_size: _size_2_t, dilation: _size_2_t = 1, **kwargs) -> Tuple[int, int]:\n    \"\"\"\n    Auto Padding for the convolution blocks\n    \"\"\"\n    if isinstance(kernel_size, int):\n        kernel_size = (kernel_size, kernel_size)\n    if isinstance(dilation, int):\n        dilation = (dilation, dilation)\n\n    pad_h = ((kernel_size[0] - 1) * dilation[0]) // 2\n    pad_w = ((kernel_size[1] - 1) * dilation[1]) // 2\n    return (pad_h, pad_w)", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 483}, "tests/test_model/test_module.py::50": {"resolved_imports": ["yolo/model/module.py"], "used_names": ["CBLinear", "torch"], "enclosing_function": "test_cblinear", "extracted_code": "# Source: yolo/model/module.py\nclass CBLinear(nn.Module):\n    \"\"\"Convolutional block that outputs multiple feature maps split along the channel dimension.\"\"\"\n\n    def __init__(self, in_channels: int, out_channels: List[int], kernel_size: int = 1, **kwargs):\n        super(CBLinear, self).__init__()\n        kwargs.setdefault(\"padding\", auto_pad(kernel_size, **kwargs))\n        self.conv = nn.Conv2d(in_channels, sum(out_channels), kernel_size, **kwargs)\n        self.out_channels = list(out_channels)\n\n    def forward(self, x: Tensor) -> Tuple[Tensor]:\n        x = self.conv(x)\n        return x.split(self.out_channels, dim=1)", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 626}, "tests/test_utils/test_bounding_box_utils.py::101": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["pytest", "transform_bbox"], "enclosing_function": "test_transform_bbox_invalid_format", "extracted_code": "# Source: yolo/utils/bounding_box_utils.py\ndef transform_bbox(bbox: Tensor, indicator=\"xywh -> xyxy\"):\n    data_type = bbox.dtype\n    in_type, out_type = indicator.replace(\" \", \"\").split(\"->\")\n\n    if in_type not in [\"xyxy\", \"xywh\", \"xycwh\"] or out_type not in [\"xyxy\", \"xywh\", \"xycwh\"]:\n        raise ValueError(\"Invalid input or output format\")\n\n    if in_type == \"xywh\":\n        x_min = bbox[..., 0]\n        y_min = bbox[..., 1]\n        x_max = bbox[..., 0] + bbox[..., 2]\n        y_max = bbox[..., 1] + bbox[..., 3]\n    elif in_type == \"xyxy\":\n        x_min = bbox[..., 0]\n        y_min = bbox[..., 1]\n        x_max = bbox[..., 2]\n        y_max = bbox[..., 3]\n    elif in_type == \"xycwh\":\n        x_min = bbox[..., 0] - bbox[..., 2] / 2\n        y_min = bbox[..., 1] - bbox[..., 3] / 2\n        x_max = bbox[..., 0] + bbox[..., 2] / 2\n        y_max = bbox[..., 1] + bbox[..., 3] / 2\n\n    if out_type == \"xywh\":\n        bbox = torch.stack([x_min, y_min, x_max - x_min, y_max - y_min], dim=-1)\n    elif out_type == \"xyxy\":\n        bbox = torch.stack([x_min, y_min, x_max, y_max], dim=-1)\n    elif out_type == \"xycwh\":\n        bbox = torch.stack([(x_min + x_max) / 2, (y_min + y_max) / 2, x_max - x_min, y_max - y_min], dim=-1)\n\n    return bbox.to(dtype=data_type)", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 1263}, "tests/test_utils/test_bounding_box_utils.py::126": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["Config", "Vec2Box", "compose", "create_model", "initialize", "torch"], "enclosing_function": "test_vec2box_autoanchor", "extracted_code": "# Source: yolo/__init__.py\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n    YOLORichProgressBar,\n\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n    \"AugmentationComposer\",\n    \"ImageLogger\",\n\n\n# Source: yolo/utils/bounding_box_utils.py\nclass Vec2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(device), scaler.to(device)\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        # TODO: need accelerate dummy test\n        dummy_input = torch.zeros(1, 3, H, W)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head[2].shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def update(self, image_size):\n        \"\"\"\n        image_size: W, H\n        \"\"\"\n        if self.image_size == image_size:\n            return\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(self.device), scaler.to(self.device)\n\n    def __call__(self, predicts):\n        preds_cls, preds_anc, preds_box = [], [], []\n        for layer_output in predicts:\n            pred_cls, pred_anc, pred_box = layer_output\n            preds_cls.append(rearrange(pred_cls, \"B C h w -> B (h w) C\"))\n            preds_anc.append(rearrange(pred_anc, \"B A R h w -> B (h w) R A\"))\n            preds_box.append(rearrange(pred_box, \"B X h w -> B (h w) X\"))\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_anc = torch.concat(preds_anc, dim=1)\n        preds_box = torch.concat(preds_box, dim=1)\n\n        pred_LTRB = preds_box * self.scaler.view(1, -1, 1)\n        lt, rb = pred_LTRB.chunk(2, dim=-1)\n        preds_box = torch.cat([self.anchor_grid - lt, self.anchor_grid + rb], dim=-1)\n        return preds_cls, preds_anc, preds_box", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4002}, "tests/test_tools/test_data_loader.py::62": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["StreamDataLoader"], "enclosing_function": "test_file_stream_data_loader_frame", "extracted_code": "# Source: yolo/tools/data_loader.py\nclass StreamDataLoader:\n    def __init__(self, data_cfg: DataConfig):\n        self.source = data_cfg.source\n        self.running = True\n        self.is_stream = isinstance(self.source, int) or str(self.source).lower().startswith(\"rtmp://\")\n\n        self.transform = AugmentationComposer([], data_cfg.image_size)\n        self.stop_event = Event()\n\n        if self.is_stream:\n            import cv2\n\n            self.cap = cv2.VideoCapture(self.source)\n        else:\n            self.source = Path(self.source)\n            self.queue = Queue()\n            self.thread = Thread(target=self.load_source)\n            self.thread.start()\n\n    def load_source(self):\n        if self.source.is_dir():  # image folder\n            self.load_image_folder(self.source)\n        elif any(self.source.suffix.lower().endswith(ext) for ext in [\".mp4\", \".avi\", \".mkv\"]):  # Video file\n            self.load_video_file(self.source)\n        else:  # Single image\n            self.process_image(self.source)\n\n    def load_image_folder(self, folder):\n        folder_path = Path(folder)\n        for file_path in folder_path.rglob(\"*\"):\n            if self.stop_event.is_set():\n                break\n            if file_path.suffix.lower() in [\".jpg\", \".jpeg\", \".png\", \".bmp\"]:\n                self.process_image(file_path)\n\n    def process_image(self, image_path):\n        image = Image.open(image_path).convert(\"RGB\")\n        if image is None:\n            raise ValueError(f\"Error loading image: {image_path}\")\n        self.process_frame(image)\n\n    def load_video_file(self, video_path):\n        import cv2\n\n        cap = cv2.VideoCapture(str(video_path))\n        while self.running:\n            ret, frame = cap.read()\n            if not ret:\n                break\n            self.process_frame(frame)\n        cap.release()\n\n    def process_frame(self, frame):\n        if isinstance(frame, np.ndarray):\n            # TODO: we don't need cv2\n            import cv2\n\n            frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n            frame = Image.fromarray(frame)\n        origin_frame = frame\n        frame, _, rev_tensor = self.transform(frame, torch.zeros(0, 5))\n        frame = frame[None]\n        rev_tensor = rev_tensor[None]\n        if not self.is_stream:\n            self.queue.put((frame, rev_tensor, origin_frame))\n        else:\n            self.current_frame = (frame, rev_tensor, origin_frame)\n\n    def __iter__(self) -> Generator[Tensor, None, None]:\n        return self\n\n    def __next__(self) -> Tensor:\n        if self.is_stream:\n            ret, frame = self.cap.read()\n            if not ret:\n                self.stop()\n                raise StopIteration\n            self.process_frame(frame)\n            return self.current_frame\n        else:\n            try:\n                frame = self.queue.get(timeout=1)\n                return frame\n            except Empty:\n                raise StopIteration\n\n    def stop(self):\n        self.running = False\n        if self.is_stream:\n            self.cap.release()\n        else:\n            self.thread.join(timeout=1)\n\n    def __len__(self):\n        return self.queue.qsize() if not self.is_stream else 0", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 3190}, "tests/test_utils/test_bounding_box_utils.py::113": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["generate_anchors"], "enclosing_function": "test_generate_anchors", "extracted_code": "# Source: yolo/utils/bounding_box_utils.py\ndef generate_anchors(image_size: List[int], strides: List[int]):\n    \"\"\"\n    Find the anchor maps for each w, h.\n\n    Args:\n        image_size List: the image size of augmented image size\n        strides List[8, 16, 32, ...]: the stride size for each predicted layer\n\n    Returns:\n        all_anchors [HW x 2]:\n        all_scalers [HW]: The index of the best targets for each anchors\n    \"\"\"\n    W, H = image_size\n    anchors = []\n    scaler = []\n    for stride in strides:\n        anchor_num = W // stride * H // stride\n        scaler.append(torch.full((anchor_num,), stride))\n        shift = stride // 2\n        h = torch.arange(0, H, stride) + shift\n        w = torch.arange(0, W, stride) + shift\n        if torch.__version__ >= \"2.3.0\":\n            anchor_h, anchor_w = torch.meshgrid(h, w, indexing=\"ij\")\n        else:\n            anchor_h, anchor_w = torch.meshgrid(h, w)\n        anchor = torch.stack([anchor_w.flatten(), anchor_h.flatten()], dim=-1)\n        anchors.append(anchor)\n    all_anchors = torch.cat(anchors, dim=0)\n    all_scalers = torch.cat(scaler, dim=0)\n    return all_anchors, all_scalers\n\nclass Anc2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        self.head_num = len(anchor_cfg.anchor)\n        self.anchor_grids = self.generate_anchors(image_size)\n        self.anchor_scale = tensor(anchor_cfg.anchor, device=device).view(self.head_num, 1, -1, 1, 1, 2)\n        self.anchor_num = self.anchor_scale.size(2)\n        self.class_num = model.num_classes\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        dummy_input = torch.zeros(1, 3, H, W).to(self.device)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head.shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def generate_anchors(self, image_size: List[int]):\n        anchor_grids = []\n        for stride in self.strides:\n            W, H = image_size[0] // stride, image_size[1] // stride\n            anchor_h, anchor_w = torch.meshgrid([torch.arange(H), torch.arange(W)], indexing=\"ij\")\n            anchor_grid = torch.stack((anchor_w, anchor_h), 2).view((1, 1, H, W, 2)).float().to(self.device)\n            anchor_grids.append(anchor_grid)\n        return anchor_grids\n\n    def update(self, image_size):\n        self.anchor_grids = self.generate_anchors(image_size)\n\n    def __call__(self, predicts: List[Tensor]):\n        preds_box, preds_cls, preds_cnf = [], [], []\n        for layer_idx, predict in enumerate(predicts):\n            predict = rearrange(predict, \"B (L C) h w -> B L h w C\", L=self.anchor_num)\n            pred_box, pred_cnf, pred_cls = predict.split((4, 1, self.class_num), dim=-1)\n            pred_box = pred_box.sigmoid()\n            pred_box[..., 0:2] = (pred_box[..., 0:2] * 2.0 - 0.5 + self.anchor_grids[layer_idx]) * self.strides[\n                layer_idx\n            ]\n            pred_box[..., 2:4] = (pred_box[..., 2:4] * 2) ** 2 * self.anchor_scale[layer_idx]\n            preds_box.append(rearrange(pred_box, \"B L h w A -> B (L h w) A\"))\n            preds_cls.append(rearrange(pred_cls, \"B L h w C -> B (L h w) C\"))\n            preds_cnf.append(rearrange(pred_cnf, \"B L h w C -> B (L h w) C\"))\n\n        preds_box = torch.concat(preds_box, dim=1)\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_cnf = torch.concat(preds_cnf, dim=1)\n\n        preds_box = transform_bbox(preds_box, \"xycwh -> xyxy\")\n        return preds_cls, None, preds_box, preds_cnf.sigmoid()", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4067}, "tests/test_tools/test_loss_functions.py::54": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py", "yolo/tools/loss_functions.py", "yolo/utils/bounding_box_utils.py"], "used_names": [], "enclosing_function": "test_yolo_loss", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 4, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::46": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_validation_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::64": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["StreamDataLoader"], "enclosing_function": "test_file_stream_data_loader_frame", "extracted_code": "# Source: yolo/tools/data_loader.py\nclass StreamDataLoader:\n    def __init__(self, data_cfg: DataConfig):\n        self.source = data_cfg.source\n        self.running = True\n        self.is_stream = isinstance(self.source, int) or str(self.source).lower().startswith(\"rtmp://\")\n\n        self.transform = AugmentationComposer([], data_cfg.image_size)\n        self.stop_event = Event()\n\n        if self.is_stream:\n            import cv2\n\n            self.cap = cv2.VideoCapture(self.source)\n        else:\n            self.source = Path(self.source)\n            self.queue = Queue()\n            self.thread = Thread(target=self.load_source)\n            self.thread.start()\n\n    def load_source(self):\n        if self.source.is_dir():  # image folder\n            self.load_image_folder(self.source)\n        elif any(self.source.suffix.lower().endswith(ext) for ext in [\".mp4\", \".avi\", \".mkv\"]):  # Video file\n            self.load_video_file(self.source)\n        else:  # Single image\n            self.process_image(self.source)\n\n    def load_image_folder(self, folder):\n        folder_path = Path(folder)\n        for file_path in folder_path.rglob(\"*\"):\n            if self.stop_event.is_set():\n                break\n            if file_path.suffix.lower() in [\".jpg\", \".jpeg\", \".png\", \".bmp\"]:\n                self.process_image(file_path)\n\n    def process_image(self, image_path):\n        image = Image.open(image_path).convert(\"RGB\")\n        if image is None:\n            raise ValueError(f\"Error loading image: {image_path}\")\n        self.process_frame(image)\n\n    def load_video_file(self, video_path):\n        import cv2\n\n        cap = cv2.VideoCapture(str(video_path))\n        while self.running:\n            ret, frame = cap.read()\n            if not ret:\n                break\n            self.process_frame(frame)\n        cap.release()\n\n    def process_frame(self, frame):\n        if isinstance(frame, np.ndarray):\n            # TODO: we don't need cv2\n            import cv2\n\n            frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n            frame = Image.fromarray(frame)\n        origin_frame = frame\n        frame, _, rev_tensor = self.transform(frame, torch.zeros(0, 5))\n        frame = frame[None]\n        rev_tensor = rev_tensor[None]\n        if not self.is_stream:\n            self.queue.put((frame, rev_tensor, origin_frame))\n        else:\n            self.current_frame = (frame, rev_tensor, origin_frame)\n\n    def __iter__(self) -> Generator[Tensor, None, None]:\n        return self\n\n    def __next__(self) -> Tensor:\n        if self.is_stream:\n            ret, frame = self.cap.read()\n            if not ret:\n                self.stop()\n                raise StopIteration\n            self.process_frame(frame)\n            return self.current_frame\n        else:\n            try:\n                frame = self.queue.get(timeout=1)\n                return frame\n            except Empty:\n                raise StopIteration\n\n    def stop(self):\n        self.running = False\n        if self.is_stream:\n            self.cap.release()\n        else:\n            self.thread.join(timeout=1)\n\n    def __len__(self):\n        return self.queue.qsize() if not self.is_stream else 0", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 3190}, "tests/test_tools/test_data_augmentation.py::28": {"resolved_imports": ["yolo/tools/data_augmentation.py"], "used_names": ["HorizontalFlip", "Image", "torch"], "enclosing_function": "test_horizontal_flip", "extracted_code": "# Source: yolo/tools/data_augmentation.py\nclass HorizontalFlip:\n    \"\"\"Randomly horizontally flips the image along with the bounding boxes.\"\"\"\n\n    def __init__(self, prob=0.5):\n        self.prob = prob\n\n    def __call__(self, image, boxes):\n        if torch.rand(1) < self.prob:\n            image = TF.hflip(image)\n            boxes[:, [1, 3]] = 1 - boxes[:, [3, 1]]\n        return image, boxes", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 395}, "tests/test_model/test_yolo.py::46": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py"], "used_names": ["Config", "OmegaConf", "YOLO", "compose", "initialize"], "enclosing_function": "test_build_model_v7", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/model/yolo.py\nclass YOLO(nn.Module):\n    \"\"\"\n    A preliminary YOLO (You Only Look Once) model class still under development.\n\n    Parameters:\n        model_cfg: Configuration for the YOLO model. Expected to define the layers,\n                   parameters, and any other relevant configuration details.\n    \"\"\"\n\n    def __init__(self, model_cfg: ModelConfig, class_num: int = 80):\n        super(YOLO, self).__init__()\n        self.num_classes = class_num\n        self.layer_map = get_layer_map()  # Get the map Dict[str: Module]\n        self.model: List[YOLOLayer] = nn.ModuleList()\n        self.reg_max = getattr(model_cfg.anchor, \"reg_max\", 16)\n        self.build_model(model_cfg.model)\n\n    def build_model(self, model_arch: Dict[str, List[Dict[str, Dict[str, Dict]]]]):\n        self.layer_index = {}\n        output_dim, layer_idx = [3], 1\n        logger.info(f\":tractor: Building YOLO\")\n        for arch_name in model_arch:\n            if model_arch[arch_name]:\n                logger.info(f\"  :building_construction:  Building {arch_name}\")\n            for layer_idx, layer_spec in enumerate(model_arch[arch_name], start=layer_idx):\n                layer_type, layer_info = next(iter(layer_spec.items()))\n                layer_args = layer_info.get(\"args\", {})\n\n                # Get input source\n                source = self.get_source_idx(layer_info.get(\"source\", -1), layer_idx)\n\n                # Find in channels\n                if any(module in layer_type for module in [\"Conv\", \"ELAN\", \"ADown\", \"AConv\", \"CBLinear\"]):\n                    layer_args[\"in_channels\"] = output_dim[source]\n                if any(module in layer_type for module in [\"Detection\", \"Segmentation\", \"Classification\"]):\n                    if isinstance(source, list):\n                        layer_args[\"in_channels\"] = [output_dim[idx] for idx in source]\n                    else:\n                        layer_args[\"in_channel\"] = output_dim[source]\n                    layer_args[\"num_classes\"] = self.num_classes\n                    layer_args[\"reg_max\"] = self.reg_max\n\n                # create layers\n                layer = self.create_layer(layer_type, source, layer_info, **layer_args)\n                self.model.append(layer)\n\n                if layer.tags:\n                    if layer.tags in self.layer_index:\n                        raise ValueError(f\"Duplicate tag '{layer_info['tags']}' found.\")\n                    self.layer_index[layer.tags] = layer_idx\n\n                out_channels = self.get_out_channels(layer_type, layer_args, output_dim, source)\n                output_dim.append(out_channels)\n                setattr(layer, \"out_c\", out_channels)\n            layer_idx += 1\n\n    def forward(self, x, external: Optional[Dict] = None, shortcut: Optional[str] = None):\n        y = {0: x, **(external or {})}\n        output = dict()\n        for index, layer in enumerate(self.model, start=1):\n            if isinstance(layer.source, list):\n                model_input = [y[idx] for idx in layer.source]\n            else:\n                model_input = y[layer.source]\n\n            external_input = {source_name: y[source_name] for source_name in layer.external}\n\n            x = layer(model_input, **external_input)\n            y[-1] = x\n            if layer.usable:\n                y[index] = x\n            if layer.output:\n                output[layer.tags] = x\n                if layer.tags == shortcut:\n                    return output\n        return output\n\n    def get_out_channels(self, layer_type: str, layer_args: dict, output_dim: list, source: Union[int, list]):\n        if hasattr(layer_args, \"out_channels\"):\n            return layer_args[\"out_channels\"]\n        if layer_type == \"CBFuse\":\n            return output_dim[source[-1]]\n        if isinstance(source, int):\n            return output_dim[source]\n        if isinstance(source, list):\n            return sum(output_dim[idx] for idx in source)\n\n    def get_source_idx(self, source: Union[ListConfig, str, int], layer_idx: int):\n        if isinstance(source, ListConfig):\n            return [self.get_source_idx(index, layer_idx) for index in source]\n        if isinstance(source, str):\n            source = self.layer_index[source]\n        if source < -1:\n            source += layer_idx\n        if source > 0:  # Using Previous Layer's Output\n            self.model[source - 1].usable = True\n        return source\n\n    def create_layer(self, layer_type: str, source: Union[int, list], layer_info: Dict, **kwargs) -> YOLOLayer:\n        if layer_type in self.layer_map:\n            layer = self.layer_map[layer_type](**kwargs)\n            setattr(layer, \"layer_type\", layer_type)\n            setattr(layer, \"source\", source)\n            setattr(layer, \"in_c\", kwargs.get(\"in_channels\", None))\n            setattr(layer, \"output\", layer_info.get(\"output\", False))\n            setattr(layer, \"tags\", layer_info.get(\"tags\", None))\n            setattr(layer, \"external\", layer_info.get(\"external\", []))\n            setattr(layer, \"usable\", 0)\n            return layer\n        else:\n            raise ValueError(f\"Unsupported layer type: {layer_type}\")\n\n    def save_load_weights(self, weights: Union[Path, OrderedDict]):\n        \"\"\"\n        Update the model's weights with the provided weights.\n\n        args:\n            weights: A OrderedDict containing the new weights.\n        \"\"\"\n        if isinstance(weights, Path):\n            weights = torch.load(weights, map_location=torch.device(\"cpu\"), weights_only=False)\n        if \"state_dict\" in weights:\n            weights = {name.removeprefix(\"model.model.\"): key for name, key in weights[\"state_dict\"].items()}\n        model_state_dict = self.model.state_dict()\n\n        # TODO1: autoload old version weight\n        # TODO2: weight transform if num_class difference\n\n        error_dict = {\"Mismatch\": set(), \"Not Found\": set()}\n        for model_key, model_weight in model_state_dict.items():\n            if model_key not in weights:\n                error_dict[\"Not Found\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            if model_weight.shape != weights[model_key].shape:\n                error_dict[\"Mismatch\"].add(tuple(model_key.split(\".\")[:-2]))\n                continue\n            model_state_dict[model_key] = weights[model_key]\n\n        for error_name, error_set in error_dict.items():\n            error_dict = dict()\n            for layer_idx, *layer_name in error_set:\n                if layer_idx not in error_dict:\n                    error_dict[layer_idx] = [\".\".join(layer_name)]\n                else:\n                    error_dict[layer_idx].append(\".\".join(layer_name))\n            for layer_idx, layer_name in error_dict.items():\n                layer_name.sort()\n                logger.warning(f\":warning: Weight {error_name} for Layer {layer_idx}: {', '.join(layer_name)}\")\n\n        self.model.load_state_dict(model_state_dict)", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 7342}, "tests/test_tools/test_dataset_preparation.py::22": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/dataset_preparation.py"], "used_names": ["Config", "Path", "os", "prepare_dataset", "shutil"], "enclosing_function": "test_prepare_dataset", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/dataset_preparation.py\ndef prepare_dataset(dataset_cfg: DatasetConfig, task: str):\n    \"\"\"\n    Prepares dataset by downloading and unzipping if necessary.\n    \"\"\"\n    # TODO: do EDA of dataset\n    data_dir = Path(dataset_cfg.path)\n    for data_type, settings in dataset_cfg.auto_download.items():\n        base_url = settings[\"base_url\"]\n        for dataset_type, dataset_args in settings.items():\n            if dataset_type != \"annotations\" and dataset_cfg.get(task, task) != dataset_type:\n                continue\n            file_name = f\"{dataset_args.get('file_name', dataset_type)}.zip\"\n            url = f\"{base_url}{file_name}\"\n            local_zip_path = data_dir / file_name\n            extract_to = data_dir / data_type if data_type != \"annotations\" else data_dir\n            final_place = extract_to / dataset_type\n\n            final_place.mkdir(parents=True, exist_ok=True)\n            if check_files(final_place, dataset_args.get(\"file_num\")):\n                logger.info(f\":white_check_mark: Dataset {dataset_type: <12} already verified.\")\n                continue\n\n            if not local_zip_path.exists():\n                download_file(url, local_zip_path)\n            unzip_file(local_zip_path, extract_to)\n\n            if not check_files(final_place, dataset_args.get(\"file_num\")):\n                logger.error(f\"Error verifying the {dataset_type} dataset after extraction.\")", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 1842}, "tests/test_tools/test_data_augmentation.py::70": {"resolved_imports": ["yolo/tools/data_augmentation.py"], "used_names": ["Image", "Mosaic", "torch"], "enclosing_function": "test_mosaic", "extracted_code": "# Source: yolo/tools/data_augmentation.py\nclass Mosaic:\n    \"\"\"Applies the Mosaic augmentation to a batch of images and their corresponding boxes.\"\"\"\n\n    def __init__(self, prob=0.5):\n        self.prob = prob\n        self.parent = None\n\n    def set_parent(self, parent):\n        self.parent = parent\n\n    def __call__(self, image, boxes):\n        if torch.rand(1) >= self.prob:\n            return image, boxes\n\n        assert self.parent is not None, \"Parent is not set. Mosaic cannot retrieve image size.\"\n\n        img_sz = self.parent.base_size  # Assuming `image_size` is defined in parent\n        more_data = self.parent.get_more_data(3)  # get 3 more images randomly\n\n        data = [(image, boxes)] + more_data\n        mosaic_image = Image.new(\"RGB\", (2 * img_sz, 2 * img_sz), (114, 114, 114))\n        vectors = np.array([(-1, -1), (0, -1), (-1, 0), (0, 0)])\n        center = np.array([img_sz, img_sz])\n        all_labels = []\n\n        for (image, boxes), vector in zip(data, vectors):\n            this_w, this_h = image.size\n            coord = tuple(center + vector * np.array([this_w, this_h]))\n\n            mosaic_image.paste(image, coord)\n            xmin, ymin, xmax, ymax = boxes[:, 1], boxes[:, 2], boxes[:, 3], boxes[:, 4]\n            xmin = (xmin * this_w + coord[0]) / (2 * img_sz)\n            xmax = (xmax * this_w + coord[0]) / (2 * img_sz)\n            ymin = (ymin * this_h + coord[1]) / (2 * img_sz)\n            ymax = (ymax * this_h + coord[1]) / (2 * img_sz)\n\n            adjusted_boxes = torch.stack([boxes[:, 0], xmin, ymin, xmax, ymax], dim=1)\n            all_labels.append(adjusted_boxes)\n\n        all_labels = torch.cat(all_labels, dim=0)\n        mosaic_image = mosaic_image.resize((img_sz, img_sz))\n        return mosaic_image, all_labels", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 1768}, "tests/test_utils/test_module_utils.py::55": {"resolved_imports": ["yolo/utils/module_utils.py"], "used_names": ["divide_into_chunks", "pytest", "re"], "enclosing_function": "test_divide_into_chunks_non_divisible_length", "extracted_code": "# Source: yolo/utils/module_utils.py\ndef divide_into_chunks(input_list, chunk_num):\n    \"\"\"\n    Args: input_list: [0, 1, 2, 3, 4, 5], chunk: 2\n    Return: [[0, 1, 2], [3, 4, 5]]\n    \"\"\"\n    list_size = len(input_list)\n\n    if list_size % chunk_num != 0:\n        raise ValueError(\n            f\"The length of the input list ({list_size}) must be exactly divisible by the number of chunks ({chunk_num}).\"\n        )\n\n    chunk_size = list_size // chunk_num\n    return [input_list[i : i + chunk_size] for i in range(0, list_size, chunk_size)]", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 538}, "tests/test_model/test_yolo.py::77": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py"], "used_names": ["torch"], "enclosing_function": "test_yolo_forward_output_shape", "extracted_code": "", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::72": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["StreamDataLoader"], "enclosing_function": "test_directory_stream_data_loader_frame", "extracted_code": "# Source: yolo/tools/data_loader.py\nclass StreamDataLoader:\n    def __init__(self, data_cfg: DataConfig):\n        self.source = data_cfg.source\n        self.running = True\n        self.is_stream = isinstance(self.source, int) or str(self.source).lower().startswith(\"rtmp://\")\n\n        self.transform = AugmentationComposer([], data_cfg.image_size)\n        self.stop_event = Event()\n\n        if self.is_stream:\n            import cv2\n\n            self.cap = cv2.VideoCapture(self.source)\n        else:\n            self.source = Path(self.source)\n            self.queue = Queue()\n            self.thread = Thread(target=self.load_source)\n            self.thread.start()\n\n    def load_source(self):\n        if self.source.is_dir():  # image folder\n            self.load_image_folder(self.source)\n        elif any(self.source.suffix.lower().endswith(ext) for ext in [\".mp4\", \".avi\", \".mkv\"]):  # Video file\n            self.load_video_file(self.source)\n        else:  # Single image\n            self.process_image(self.source)\n\n    def load_image_folder(self, folder):\n        folder_path = Path(folder)\n        for file_path in folder_path.rglob(\"*\"):\n            if self.stop_event.is_set():\n                break\n            if file_path.suffix.lower() in [\".jpg\", \".jpeg\", \".png\", \".bmp\"]:\n                self.process_image(file_path)\n\n    def process_image(self, image_path):\n        image = Image.open(image_path).convert(\"RGB\")\n        if image is None:\n            raise ValueError(f\"Error loading image: {image_path}\")\n        self.process_frame(image)\n\n    def load_video_file(self, video_path):\n        import cv2\n\n        cap = cv2.VideoCapture(str(video_path))\n        while self.running:\n            ret, frame = cap.read()\n            if not ret:\n                break\n            self.process_frame(frame)\n        cap.release()\n\n    def process_frame(self, frame):\n        if isinstance(frame, np.ndarray):\n            # TODO: we don't need cv2\n            import cv2\n\n            frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n            frame = Image.fromarray(frame)\n        origin_frame = frame\n        frame, _, rev_tensor = self.transform(frame, torch.zeros(0, 5))\n        frame = frame[None]\n        rev_tensor = rev_tensor[None]\n        if not self.is_stream:\n            self.queue.put((frame, rev_tensor, origin_frame))\n        else:\n            self.current_frame = (frame, rev_tensor, origin_frame)\n\n    def __iter__(self) -> Generator[Tensor, None, None]:\n        return self\n\n    def __next__(self) -> Tensor:\n        if self.is_stream:\n            ret, frame = self.cap.read()\n            if not ret:\n                self.stop()\n                raise StopIteration\n            self.process_frame(frame)\n            return self.current_frame\n        else:\n            try:\n                frame = self.queue.get(timeout=1)\n                return frame\n            except Empty:\n                raise StopIteration\n\n    def stop(self):\n        self.running = False\n        if self.is_stream:\n            self.cap.release()\n        else:\n            self.thread.join(timeout=1)\n\n    def __len__(self):\n        return self.queue.qsize() if not self.is_stream else 0", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 3190}, "tests/test_tools/test_data_loader.py::47": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_validation_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_utils/test_module_utils.py::49": {"resolved_imports": ["yolo/utils/module_utils.py"], "used_names": ["divide_into_chunks"], "enclosing_function": "test_divide_into_chunks", "extracted_code": "# Source: yolo/utils/module_utils.py\ndef divide_into_chunks(input_list, chunk_num):\n    \"\"\"\n    Args: input_list: [0, 1, 2, 3, 4, 5], chunk: 2\n    Return: [[0, 1, 2], [3, 4, 5]]\n    \"\"\"\n    list_size = len(input_list)\n\n    if list_size % chunk_num != 0:\n        raise ValueError(\n            f\"The length of the input list ({list_size}) must be exactly divisible by the number of chunks ({chunk_num}).\"\n        )\n\n    chunk_size = list_size // chunk_num\n    return [input_list[i : i + chunk_size] for i in range(0, list_size, chunk_size)]", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 538}, "tests/test_utils/test_bounding_box_utils.py::114": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["generate_anchors"], "enclosing_function": "test_generate_anchors", "extracted_code": "# Source: yolo/utils/bounding_box_utils.py\ndef generate_anchors(image_size: List[int], strides: List[int]):\n    \"\"\"\n    Find the anchor maps for each w, h.\n\n    Args:\n        image_size List: the image size of augmented image size\n        strides List[8, 16, 32, ...]: the stride size for each predicted layer\n\n    Returns:\n        all_anchors [HW x 2]:\n        all_scalers [HW]: The index of the best targets for each anchors\n    \"\"\"\n    W, H = image_size\n    anchors = []\n    scaler = []\n    for stride in strides:\n        anchor_num = W // stride * H // stride\n        scaler.append(torch.full((anchor_num,), stride))\n        shift = stride // 2\n        h = torch.arange(0, H, stride) + shift\n        w = torch.arange(0, W, stride) + shift\n        if torch.__version__ >= \"2.3.0\":\n            anchor_h, anchor_w = torch.meshgrid(h, w, indexing=\"ij\")\n        else:\n            anchor_h, anchor_w = torch.meshgrid(h, w)\n        anchor = torch.stack([anchor_w.flatten(), anchor_h.flatten()], dim=-1)\n        anchors.append(anchor)\n    all_anchors = torch.cat(anchors, dim=0)\n    all_scalers = torch.cat(scaler, dim=0)\n    return all_anchors, all_scalers\n\nclass Anc2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        self.head_num = len(anchor_cfg.anchor)\n        self.anchor_grids = self.generate_anchors(image_size)\n        self.anchor_scale = tensor(anchor_cfg.anchor, device=device).view(self.head_num, 1, -1, 1, 1, 2)\n        self.anchor_num = self.anchor_scale.size(2)\n        self.class_num = model.num_classes\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        dummy_input = torch.zeros(1, 3, H, W).to(self.device)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head.shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def generate_anchors(self, image_size: List[int]):\n        anchor_grids = []\n        for stride in self.strides:\n            W, H = image_size[0] // stride, image_size[1] // stride\n            anchor_h, anchor_w = torch.meshgrid([torch.arange(H), torch.arange(W)], indexing=\"ij\")\n            anchor_grid = torch.stack((anchor_w, anchor_h), 2).view((1, 1, H, W, 2)).float().to(self.device)\n            anchor_grids.append(anchor_grid)\n        return anchor_grids\n\n    def update(self, image_size):\n        self.anchor_grids = self.generate_anchors(image_size)\n\n    def __call__(self, predicts: List[Tensor]):\n        preds_box, preds_cls, preds_cnf = [], [], []\n        for layer_idx, predict in enumerate(predicts):\n            predict = rearrange(predict, \"B (L C) h w -> B L h w C\", L=self.anchor_num)\n            pred_box, pred_cnf, pred_cls = predict.split((4, 1, self.class_num), dim=-1)\n            pred_box = pred_box.sigmoid()\n            pred_box[..., 0:2] = (pred_box[..., 0:2] * 2.0 - 0.5 + self.anchor_grids[layer_idx]) * self.strides[\n                layer_idx\n            ]\n            pred_box[..., 2:4] = (pred_box[..., 2:4] * 2) ** 2 * self.anchor_scale[layer_idx]\n            preds_box.append(rearrange(pred_box, \"B L h w A -> B (L h w) A\"))\n            preds_cls.append(rearrange(pred_cls, \"B L h w C -> B (L h w) C\"))\n            preds_cnf.append(rearrange(pred_cnf, \"B L h w C -> B (L h w) C\"))\n\n        preds_box = torch.concat(preds_box, dim=1)\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_cnf = torch.concat(preds_cnf, dim=1)\n\n        preds_box = transform_bbox(preds_box, \"xycwh -> xyxy\")\n        return preds_cls, None, preds_box, preds_cnf.sigmoid()", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4067}, "tests/test_tools/test_loss_functions.py::56": {"resolved_imports": ["yolo/config/config.py", "yolo/model/yolo.py", "yolo/tools/loss_functions.py", "yolo/utils/bounding_box_utils.py"], "used_names": [], "enclosing_function": "test_yolo_loss", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 4, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::63": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["StreamDataLoader"], "enclosing_function": "test_file_stream_data_loader_frame", "extracted_code": "# Source: yolo/tools/data_loader.py\nclass StreamDataLoader:\n    def __init__(self, data_cfg: DataConfig):\n        self.source = data_cfg.source\n        self.running = True\n        self.is_stream = isinstance(self.source, int) or str(self.source).lower().startswith(\"rtmp://\")\n\n        self.transform = AugmentationComposer([], data_cfg.image_size)\n        self.stop_event = Event()\n\n        if self.is_stream:\n            import cv2\n\n            self.cap = cv2.VideoCapture(self.source)\n        else:\n            self.source = Path(self.source)\n            self.queue = Queue()\n            self.thread = Thread(target=self.load_source)\n            self.thread.start()\n\n    def load_source(self):\n        if self.source.is_dir():  # image folder\n            self.load_image_folder(self.source)\n        elif any(self.source.suffix.lower().endswith(ext) for ext in [\".mp4\", \".avi\", \".mkv\"]):  # Video file\n            self.load_video_file(self.source)\n        else:  # Single image\n            self.process_image(self.source)\n\n    def load_image_folder(self, folder):\n        folder_path = Path(folder)\n        for file_path in folder_path.rglob(\"*\"):\n            if self.stop_event.is_set():\n                break\n            if file_path.suffix.lower() in [\".jpg\", \".jpeg\", \".png\", \".bmp\"]:\n                self.process_image(file_path)\n\n    def process_image(self, image_path):\n        image = Image.open(image_path).convert(\"RGB\")\n        if image is None:\n            raise ValueError(f\"Error loading image: {image_path}\")\n        self.process_frame(image)\n\n    def load_video_file(self, video_path):\n        import cv2\n\n        cap = cv2.VideoCapture(str(video_path))\n        while self.running:\n            ret, frame = cap.read()\n            if not ret:\n                break\n            self.process_frame(frame)\n        cap.release()\n\n    def process_frame(self, frame):\n        if isinstance(frame, np.ndarray):\n            # TODO: we don't need cv2\n            import cv2\n\n            frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n            frame = Image.fromarray(frame)\n        origin_frame = frame\n        frame, _, rev_tensor = self.transform(frame, torch.zeros(0, 5))\n        frame = frame[None]\n        rev_tensor = rev_tensor[None]\n        if not self.is_stream:\n            self.queue.put((frame, rev_tensor, origin_frame))\n        else:\n            self.current_frame = (frame, rev_tensor, origin_frame)\n\n    def __iter__(self) -> Generator[Tensor, None, None]:\n        return self\n\n    def __next__(self) -> Tensor:\n        if self.is_stream:\n            ret, frame = self.cap.read()\n            if not ret:\n                self.stop()\n                raise StopIteration\n            self.process_frame(frame)\n            return self.current_frame\n        else:\n            try:\n                frame = self.queue.get(timeout=1)\n                return frame\n            except Empty:\n                raise StopIteration\n\n    def stop(self):\n        self.running = False\n        if self.is_stream:\n            self.cap.release()\n        else:\n            self.thread.join(timeout=1)\n\n    def __len__(self):\n        return self.queue.qsize() if not self.is_stream else 0", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 3190}, "tests/test_tools/test_data_loader.py::34": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_training_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_utils/test_bounding_box_utils.py::127": {"resolved_imports": ["yolo/__init__.py", "yolo/config/config.py", "yolo/utils/bounding_box_utils.py"], "used_names": ["Config", "Vec2Box", "compose", "create_model", "initialize", "torch"], "enclosing_function": "test_vec2box_autoanchor", "extracted_code": "# Source: yolo/__init__.py\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n\nfrom yolo.config.config import Config, NMSConfig\nfrom yolo.model.yolo import create_model\nfrom yolo.tools.data_loader import AugmentationComposer, create_dataloader\nfrom yolo.tools.drawer import draw_bboxes\nfrom yolo.tools.solver import TrainModel\nfrom yolo.utils.bounding_box_utils import Anc2Box, Vec2Box, bbox_nms, create_converter\nfrom yolo.utils.deploy_utils import FastModelLoader\nfrom yolo.utils.logging_utils import (\n    ImageLogger,\n    YOLORichModelSummary,\n    YOLORichProgressBar,\n\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n\nall = [\n    \"create_model\",\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n\n    \"Config\",\n    \"YOLORichProgressBar\",\n    \"NMSConfig\",\n    \"YOLORichModelSummary\",\n    \"validate_log_directory\",\n    \"draw_bboxes\",\n    \"Vec2Box\",\n    \"Anc2Box\",\n    \"bbox_nms\",\n    \"create_converter\",\n    \"AugmentationComposer\",\n    \"ImageLogger\",\n\n\n# Source: yolo/utils/bounding_box_utils.py\nclass Vec2Box:\n    def __init__(self, model: YOLO, anchor_cfg: AnchorConfig, image_size, device):\n        self.device = device\n\n        if hasattr(anchor_cfg, \"strides\"):\n            logger.info(f\":japanese_not_free_of_charge_button: Found stride of model {anchor_cfg.strides}\")\n            self.strides = anchor_cfg.strides\n        else:\n            logger.info(\":teddy_bear: Found no stride of model, performed a dummy test for auto-anchor size\")\n            self.strides = self.create_auto_anchor(model, image_size)\n\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(device), scaler.to(device)\n\n    def create_auto_anchor(self, model: YOLO, image_size):\n        W, H = image_size\n        # TODO: need accelerate dummy test\n        dummy_input = torch.zeros(1, 3, H, W)\n        dummy_output = model(dummy_input)\n        strides = []\n        for predict_head in dummy_output[\"Main\"]:\n            _, _, *anchor_num = predict_head[2].shape\n            strides.append(W // anchor_num[1])\n        return strides\n\n    def update(self, image_size):\n        \"\"\"\n        image_size: W, H\n        \"\"\"\n        if self.image_size == image_size:\n            return\n        anchor_grid, scaler = generate_anchors(image_size, self.strides)\n        self.image_size = image_size\n        self.anchor_grid, self.scaler = anchor_grid.to(self.device), scaler.to(self.device)\n\n    def __call__(self, predicts):\n        preds_cls, preds_anc, preds_box = [], [], []\n        for layer_output in predicts:\n            pred_cls, pred_anc, pred_box = layer_output\n            preds_cls.append(rearrange(pred_cls, \"B C h w -> B (h w) C\"))\n            preds_anc.append(rearrange(pred_anc, \"B A R h w -> B (h w) R A\"))\n            preds_box.append(rearrange(pred_box, \"B X h w -> B (h w) X\"))\n        preds_cls = torch.concat(preds_cls, dim=1)\n        preds_anc = torch.concat(preds_anc, dim=1)\n        preds_box = torch.concat(preds_box, dim=1)\n\n        pred_LTRB = preds_box * self.scaler.view(1, -1, 1)\n        lt, rb = pred_LTRB.chunk(2, dim=-1)\n        preds_box = torch.cat([self.anchor_grid - lt, self.anchor_grid + rb], dim=-1)\n        return preds_cls, preds_anc, preds_box", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 4002}, "tests/test_tools/test_data_loader.py::24": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["Config", "Path", "create_dataloader"], "enclosing_function": "test_create_dataloader_cache", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/data_loader.py\ndef create_dataloader(data_cfg: DataConfig, dataset_cfg: DatasetConfig, task: str = \"train\"):\n    if task == \"inference\":\n        return StreamDataLoader(data_cfg)\n\n    if getattr(dataset_cfg, \"auto_download\", False):\n        prepare_dataset(dataset_cfg, task)\n    dataset = YoloDataset(data_cfg, dataset_cfg, task)\n\n    return DataLoader(\n        dataset,\n        batch_size=data_cfg.batch_size,\n        num_workers=data_cfg.cpu_num,\n        pin_memory=data_cfg.pin_memory,\n        collate_fn=collate_fn,\n    )", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 971}, "tests/test_tools/test_data_loader.py::40": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_training_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::35": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["DataLoader", "Path"], "enclosing_function": "test_training_data_loader_correctness", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_tools/test_data_loader.py::25": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["Config", "Path", "create_dataloader"], "enclosing_function": "test_create_dataloader_cache", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/data_loader.py\ndef create_dataloader(data_cfg: DataConfig, dataset_cfg: DatasetConfig, task: str = \"train\"):\n    if task == \"inference\":\n        return StreamDataLoader(data_cfg)\n\n    if getattr(dataset_cfg, \"auto_download\", False):\n        prepare_dataset(dataset_cfg, task)\n    dataset = YoloDataset(data_cfg, dataset_cfg, task)\n\n    return DataLoader(\n        dataset,\n        batch_size=data_cfg.batch_size,\n        num_workers=data_cfg.cpu_num,\n        pin_memory=data_cfg.pin_memory,\n        collate_fn=collate_fn,\n    )", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 971}, "tests/test_tools/test_data_augmentation.py::71": {"resolved_imports": ["yolo/tools/data_augmentation.py"], "used_names": ["Image", "Mosaic", "torch"], "enclosing_function": "test_mosaic", "extracted_code": "# Source: yolo/tools/data_augmentation.py\nclass Mosaic:\n    \"\"\"Applies the Mosaic augmentation to a batch of images and their corresponding boxes.\"\"\"\n\n    def __init__(self, prob=0.5):\n        self.prob = prob\n        self.parent = None\n\n    def set_parent(self, parent):\n        self.parent = parent\n\n    def __call__(self, image, boxes):\n        if torch.rand(1) >= self.prob:\n            return image, boxes\n\n        assert self.parent is not None, \"Parent is not set. Mosaic cannot retrieve image size.\"\n\n        img_sz = self.parent.base_size  # Assuming `image_size` is defined in parent\n        more_data = self.parent.get_more_data(3)  # get 3 more images randomly\n\n        data = [(image, boxes)] + more_data\n        mosaic_image = Image.new(\"RGB\", (2 * img_sz, 2 * img_sz), (114, 114, 114))\n        vectors = np.array([(-1, -1), (0, -1), (-1, 0), (0, 0)])\n        center = np.array([img_sz, img_sz])\n        all_labels = []\n\n        for (image, boxes), vector in zip(data, vectors):\n            this_w, this_h = image.size\n            coord = tuple(center + vector * np.array([this_w, this_h]))\n\n            mosaic_image.paste(image, coord)\n            xmin, ymin, xmax, ymax = boxes[:, 1], boxes[:, 2], boxes[:, 3], boxes[:, 4]\n            xmin = (xmin * this_w + coord[0]) / (2 * img_sz)\n            xmax = (xmax * this_w + coord[0]) / (2 * img_sz)\n            ymin = (ymin * this_h + coord[1]) / (2 * img_sz)\n            ymax = (ymax * this_h + coord[1]) / (2 * img_sz)\n\n            adjusted_boxes = torch.stack([boxes[:, 0], xmin, ymin, xmax, ymax], dim=1)\n            all_labels.append(adjusted_boxes)\n\n        all_labels = torch.cat(all_labels, dim=0)\n        mosaic_image = mosaic_image.resize((img_sz, img_sz))\n        return mosaic_image, all_labels", "n_imports_parsed": 6, "n_files_resolved": 1, "n_chars_extracted": 1768}, "tests/test_tools/test_data_loader.py::26": {"resolved_imports": ["yolo/config/config.py", "yolo/tools/data_loader.py"], "used_names": ["Config", "Path", "create_dataloader"], "enclosing_function": "test_create_dataloader_cache", "extracted_code": "# Source: yolo/config/config.py\nclass Config:\n    task: Union[TrainConfig, InferenceConfig, ValidationConfig]\n    dataset: DatasetConfig\n    model: ModelConfig\n    name: str\n\n    accelerator: Optional[str]\n    device: Union[str, int, List[int]]\n    cpu_num: int\n\n    image_size: List[int]\n\n    out_path: str\n    exist_ok: bool\n\n    lucky_number: 10\n    use_wandb: bool\n    use_tensorboard: bool\n\n    weight: Optional[str]\n\n\n# Source: yolo/tools/data_loader.py\ndef create_dataloader(data_cfg: DataConfig, dataset_cfg: DatasetConfig, task: str = \"train\"):\n    if task == \"inference\":\n        return StreamDataLoader(data_cfg)\n\n    if getattr(dataset_cfg, \"auto_download\", False):\n        prepare_dataset(dataset_cfg, task)\n    dataset = YoloDataset(data_cfg, dataset_cfg, task)\n\n    return DataLoader(\n        dataset,\n        batch_size=data_cfg.batch_size,\n        num_workers=data_cfg.cpu_num,\n        pin_memory=data_cfg.pin_memory,\n        collate_fn=collate_fn,\n    )", "n_imports_parsed": 5, "n_files_resolved": 2, "n_chars_extracted": 971}, "tests/test_model/test_module.py::22": {"resolved_imports": ["yolo/model/module.py"], "used_names": ["Conv", "torch"], "enclosing_function": "test_conv", "extracted_code": "# Source: yolo/model/module.py\nclass Conv(nn.Module):\n    \"\"\"A basic convolutional block that includes convolution, batch normalization, and activation.\"\"\"\n\n    def __init__(\n        self,\n        in_channels: int,\n        out_channels: int,\n        kernel_size: _size_2_t,\n        *,\n        activation: Optional[str] = \"SiLU\",\n        **kwargs,\n    ):\n        super().__init__()\n        kwargs.setdefault(\"padding\", auto_pad(kernel_size, **kwargs))\n        self.conv = nn.Conv2d(in_channels, out_channels, kernel_size, bias=False, **kwargs)\n        self.bn = nn.BatchNorm2d(out_channels, eps=1e-3, momentum=3e-2)\n        self.act = create_activation_function(activation)\n\n    def forward(self, x: Tensor) -> Tensor:\n        return self.act(self.bn(self.conv(x)))", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 764}}}