| import argparse |
| import glob |
| import os |
| import re |
| import sys |
|
|
| import numpy as np |
| import torch |
| import SimpleITK as sitk |
|
|
| |
| os.environ.setdefault("MONAI_SKIP_SUBMODULES", "1") |
| _repo_root = os.path.abspath(os.path.dirname(__file__)) |
| if "" in sys.path: |
| sys.path.remove("") |
| if _repo_root in sys.path: |
| sys.path.remove(_repo_root) |
| import monai |
| sys.path.insert(0, _repo_root) |
|
|
| from monai.inferers import SlidingWindowInferer |
| from monai.utils import set_determinism |
|
|
| from light_training.dataloading.dataset import MedicalDataset, get_train_val_test_loader_from_train |
| from light_training.evaluation.metric import dice |
| from light_training.prediction import Predictor |
| from light_training.trainer import Trainer |
|
|
| set_determinism(123) |
|
|
|
|
| def _parse_csv_ints(s: str, n: int): |
| parts = [p.strip() for p in str(s).split(",") if p.strip()] |
| if len(parts) != n: |
| raise ValueError(f"expect {n} integers like '128,128,128', got: {s}") |
| return [int(x) for x in parts] |
|
|
|
|
| def _parse_csv_floats(s: str, n: int): |
| parts = [p.strip() for p in str(s).split(",") if p.strip()] |
| if len(parts) != n: |
| raise ValueError(f"expect {n} floats like '1,1,1', got: {s}") |
| return [float(x) for x in parts] |
|
|
|
|
| def _find_ckpt_from_logdir(logdir: str, prefer: str = "best") -> str: |
| model_dir = os.path.join(logdir, "model") |
| if not os.path.isdir(model_dir): |
| raise FileNotFoundError(f"model dir not found: {model_dir}") |
|
|
| best = sorted(glob.glob(os.path.join(model_dir, "best_model_*.pt"))) |
| final = sorted(glob.glob(os.path.join(model_dir, "final_model_*.pt"))) |
| tmp = sorted(glob.glob(os.path.join(model_dir, "tmp_model_ep*.pt"))) |
| any_pt = sorted(glob.glob(os.path.join(model_dir, "*.pt"))) |
|
|
| def pick_by_score(paths): |
| |
| scored = [] |
| for p in paths: |
| m = re.search(r"_(\d+\\.?\\d*)\\.pt$", os.path.basename(p)) |
| if m is None: |
| continue |
| try: |
| scored.append((float(m.group(1)), p)) |
| except ValueError: |
| continue |
| if scored: |
| scored.sort(key=lambda x: x[0], reverse=True) |
| return scored[0][1] |
| return None |
|
|
| if prefer == "best": |
| picked = pick_by_score(best) or (best[-1] if best else None) |
| if picked: |
| return picked |
| if prefer in {"best", "final"}: |
| picked = pick_by_score(final) or (final[-1] if final else None) |
| if picked: |
| return picked |
| if prefer in {"best", "final", "latest"}: |
| if tmp: |
| tmp.sort(key=lambda p: os.path.getmtime(p), reverse=True) |
| return tmp[0] |
| if any_pt: |
| any_pt.sort(key=lambda p: os.path.getmtime(p), reverse=True) |
| return any_pt[0] |
|
|
| raise FileNotFoundError(f"no checkpoint found under: {model_dir}") |
|
|
|
|
| class BraTSTrainer(Trainer): |
| def __init__( |
| self, |
| ckpt_path: str, |
| save_path: str, |
| patch_size, |
| sw_batch_size: int = 2, |
| overlap: float = 0.5, |
| mirror_axes=(0, 1, 2), |
| raw_spacing=(1.0, 1.0, 1.0), |
| device="cuda:0", |
| print_dice: bool = False, |
| ): |
| super().__init__( |
| env_type="pytorch", |
| max_epochs=1, |
| batch_size=1, |
| device=device, |
| val_every=1, |
| num_gpus=1, |
| logdir="", |
| master_port=17751, |
| training_script=__file__, |
| ) |
|
|
| self.patch_size = patch_size |
| self.augmentation = False |
| self.print_dice = print_dice |
| self.save_path = save_path |
| self.raw_spacing = raw_spacing |
|
|
| from model_segmamba.segmamba import SegMamba |
|
|
| self.model = SegMamba( |
| in_chans=4, |
| out_chans=4, |
| depths=[2, 2, 2, 2], |
| feat_size=[48, 96, 192, 384], |
| ) |
| self.load_state_dict(ckpt_path, strict=True) |
| self.model.eval() |
|
|
| window_infer = SlidingWindowInferer( |
| roi_size=patch_size, |
| sw_batch_size=sw_batch_size, |
| overlap=overlap, |
| progress=True, |
| mode="gaussian", |
| ) |
|
|
| self.predictor = Predictor( |
| window_infer=window_infer, |
| mirror_axes=list(mirror_axes) if mirror_axes is not None else None, |
| ) |
|
|
| os.makedirs(self.save_path, exist_ok=True) |
| |
| def convert_labels(self, labels): |
| |
| result = [(labels == 1) | (labels == 3), (labels == 1) | (labels == 3) | (labels == 2), labels == 3] |
| |
| return torch.cat(result, dim=1).float() |
|
|
| def get_input(self, batch): |
| image = batch["data"] |
| label = batch["seg"] |
| properties = batch["properties"] |
| label = self.convert_labels(label) |
|
|
| return image, label, properties |
| |
| def validation_step(self, batch): |
| image, label, properties = self.get_input(batch) |
| |
| |
| |
| |
| |
| |
|
|
| logits = self.predictor.maybe_mirror_and_predict(image, self.model, device=self.device) |
| pred_lbl = logits.argmax(dim=1) |
| pred_3c = self.convert_labels(pred_lbl[:, None])[0].cpu().numpy().astype(np.uint8) |
|
|
| if self.print_dice: |
| gt_3c = label[0].cpu().numpy() |
| dices = [dice(pred_3c[i], gt_3c[i]) for i in range(3)] |
| print(dices) |
|
|
| case_name = properties.get("name", "") |
| if isinstance(case_name, (list, tuple)) and len(case_name) > 0: |
| case_name = case_name[0] |
|
|
| out_path = os.path.join(self.save_path, f"{case_name}.nii.gz") |
| pred_itk = sitk.GetImageFromArray(pred_3c, isVector=False) |
| pred_itk.SetSpacing((float(self.raw_spacing[0]), float(self.raw_spacing[1]), float(self.raw_spacing[2]), 1.0)) |
| sitk.WriteImage(pred_itk, out_path) |
| print(f"saved: {out_path}") |
| |
| return 0 |
|
|
| def convert_labels_dim0(self, labels): |
| |
| result = [(labels == 1) | (labels == 3), (labels == 1) | (labels == 3) | (labels == 2), labels == 3] |
| |
| return torch.cat(result, dim=0).float() |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser(description="SegMamba inference/prediction for BraTS2023.") |
| parser.add_argument("--data_dir", type=str, default="./data/fullres/train", help="Preprocessed data directory (contains *.npz).") |
| parser.add_argument("--split", type=str, default="test", choices=["train", "val", "test", "all"]) |
| parser.add_argument("--train_rate", type=float, default=0.7) |
| parser.add_argument("--val_rate", type=float, default=0.1) |
| parser.add_argument("--test_rate", type=float, default=0.2) |
| parser.add_argument("--seed", type=int, default=42) |
|
|
| parser.add_argument("--ckpt", type=str, default="", help="Checkpoint path (*.pt). If empty, will search under --logdir/model.") |
| parser.add_argument("--logdir", type=str, default="./logs/segmamba", help="Training logdir to locate checkpoints when --ckpt is empty.") |
| parser.add_argument("--ckpt_prefer", type=str, default="best", choices=["best", "final", "latest"]) |
|
|
| parser.add_argument("--save_dir", type=str, default="./prediction_results/segmamba", help="Directory to save prediction nii.gz.") |
| parser.add_argument("--device", type=str, default="cuda:0") |
| parser.add_argument("--patch_size", type=str, default="128,128,128") |
| parser.add_argument("--sw_batch_size", type=int, default=2) |
| parser.add_argument("--overlap", type=float, default=0.5) |
| parser.add_argument("--raw_spacing", type=str, default="1,1,1", help="Spacing used when saving NIfTI, e.g. '1,1,1'.") |
| parser.add_argument("--no_mirror", action="store_true", help="Disable mirror TTA.") |
| parser.add_argument("--print_dice", action="store_true", help="Print dice against preprocessed seg (if available).") |
| args = parser.parse_args() |
|
|
| patch_size = _parse_csv_ints(args.patch_size, 3) |
| raw_spacing = _parse_csv_floats(args.raw_spacing, 3) |
|
|
| ckpt_path = args.ckpt.strip() |
| if ckpt_path == "": |
| ckpt_path = _find_ckpt_from_logdir(args.logdir, prefer=args.ckpt_prefer) |
| if not os.path.isfile(ckpt_path): |
| raise FileNotFoundError(f"checkpoint not found: {ckpt_path}") |
| print(f"Using checkpoint: {ckpt_path}") |
|
|
| trainer = BraTSTrainer( |
| ckpt_path=ckpt_path, |
| save_path=args.save_dir, |
| patch_size=patch_size, |
| sw_batch_size=args.sw_batch_size, |
| overlap=args.overlap, |
| mirror_axes=None if args.no_mirror else (0, 1, 2), |
| raw_spacing=raw_spacing, |
| device=args.device, |
| print_dice=args.print_dice, |
| ) |
|
|
| if args.split == "all": |
| all_paths = sorted(glob.glob(os.path.join(args.data_dir, "*.npz"))) |
| ds = MedicalDataset(all_paths, test=False) |
| else: |
| train_ds, val_ds, test_ds = get_train_val_test_loader_from_train( |
| args.data_dir, |
| train_rate=args.train_rate, |
| val_rate=args.val_rate, |
| test_rate=args.test_rate, |
| seed=args.seed, |
| ) |
| ds = {"train": train_ds, "val": val_ds, "test": test_ds}[args.split] |
|
|
| trainer.validation_single_gpu(ds) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|