コード例 #1
0
    def __init__(
        self,
        summary_writer=None,
        interval=1,
        log_dir="./runs",
        tag_name="val_dice",
        compute_metric=True,
        images=True,
        image_interval=1,
        max_channels=1,
        max_frames=64,
        add_scalar=True,
        merge_scalar=False,
        fold_size=0,
    ):
        self.writer = SummaryWriter(
            log_dir=log_dir) if summary_writer is None else summary_writer
        self.interval = interval
        self.tag_name = tag_name
        self.compute_metric = compute_metric
        self.images = images
        self.image_interval = image_interval
        self.max_channels = max_channels
        self.max_frames = max_frames
        self.add_scalar = add_scalar
        self.merge_scalar = merge_scalar
        self.fold_size = fold_size
        self.logger = logging.getLogger(__name__)

        if torch.distributed.is_initialized():
            self.tag_name = "{}-r{}".format(self.tag_name,
                                            torch.distributed.get_rank())

        self.plot_data = {}
        self.metric_data = {}
コード例 #2
0
"""
"""## Download dataset

Downloads and extracts the dataset.  
The dataset comes from http://medicaldecathlon.com/.
"""

md5 = "410d4a301da4e5b2f6f86ec3ddba524e"

root_dir = "//home//imoreira//Data"
#root_dir = "C:\\Users\\isasi\\Downloads"
data_dir = os.path.join(root_dir, "Kidneys_Data")
out_dir = os.path.join(root_dir, "Kidneys_Best_Model")
tensorboard_dir = "//home//imoreira//Data//Tensorboard_Kidneys"

writer = SummaryWriter(log_dir=tensorboard_dir)
"""## Set MSD Spleen dataset path"""

train_images = sorted(glob.glob(os.path.join(data_dir, "imagesTr",
                                             "*.nii.gz")))
train_labels = sorted(glob.glob(os.path.join(data_dir, "labelsTr",
                                             "*.nii.gz")))
data_dicts = [{
    "image": image_name,
    "label": label_name
} for image_name, label_name in zip(train_images, train_labels)]
#n = len(data_dicts)
#train_files, val_files = data_dicts[:-3], data_dicts[-3:]
#train_files, val_files = data_dicts[:int(n*0.8)], data_dicts[int(n*0.2):]

val_files, train_files, test_files = data_dicts[0:8], data_dicts[
コード例 #3
0
class DeepgrowStatsHandler:
    def __init__(
        self,
        summary_writer=None,
        interval=1,
        log_dir="./runs",
        tag_name="val_dice",
        compute_metric=True,
        images=True,
        image_interval=1,
        max_channels=1,
        max_frames=64,
        add_scalar=True,
        merge_scalar=False,
        fold_size=0,
    ):
        self.writer = SummaryWriter(
            log_dir=log_dir) if summary_writer is None else summary_writer
        self.interval = interval
        self.tag_name = tag_name
        self.compute_metric = compute_metric
        self.images = images
        self.image_interval = image_interval
        self.max_channels = max_channels
        self.max_frames = max_frames
        self.add_scalar = add_scalar
        self.merge_scalar = merge_scalar
        self.fold_size = fold_size
        self.logger = logging.getLogger(__name__)

        if torch.distributed.is_initialized():
            self.tag_name = "{}-r{}".format(self.tag_name,
                                            torch.distributed.get_rank())

        self.plot_data = {}
        self.metric_data = {}

    def attach(self, engine: Engine) -> None:
        engine.add_event_handler(
            Events.ITERATION_COMPLETED(every=self.interval), self, "iteration")
        engine.add_event_handler(Events.EPOCH_COMPLETED(every=1), self,
                                 "epoch")

    def write_images(self, epoch):
        if not self.plot_data or not len(self.plot_data):
            return

        all_imgs = []
        for region in sorted(self.plot_data.keys()):
            metric = self.metric_data.get(region)
            region_data = self.plot_data[region]
            if len(region_data[0].shape) == 3:
                ti = Image.new("RGB", region_data[0].shape[1:])
                d = ImageDraw.Draw(ti)
                t = "region: {}".format(region)
                if self.compute_metric:
                    t = t + "\ndice: {:.4f}".format(metric.mean())
                    t = t + "\nstdev: {:.4f}".format(metric.stdev())
                d.multiline_text((10, 10), t, fill=(255, 255, 0))
                ti = rescale_array(
                    np.rollaxis(np.array(ti), 2, 0)[0][np.newaxis])
                all_imgs.append(ti)
            all_imgs.extend(region_data)

        if len(all_imgs[0].shape) == 3:
            img_tensor = make_grid(tensor=torch.from_numpy(np.array(all_imgs)),
                                   nrow=4,
                                   normalize=True,
                                   pad_value=2)
            self.writer.add_image(tag=f"Deepgrow Regions ({self.tag_name})",
                                  img_tensor=img_tensor,
                                  global_step=epoch)

        if len(all_imgs[0].shape) == 4:
            for region in sorted(self.plot_data.keys()):
                tags = [
                    f"region_{region}_image", f"region_{region}_label",
                    f"region_{region}_output"
                ]
                if torch.distributed.is_initialized():
                    rank = "r{}-".format(torch.distributed.get_rank())
                    tags = [rank + tags[0], rank + tags[1], rank + tags[2]]
                for i in range(3):
                    img = self.plot_data[region][i]
                    img = np.moveaxis(img, -3, -1)
                    plot_2d_or_3d_image(img[np.newaxis], epoch, self.writer, 0,
                                        self.max_channels, self.max_frames,
                                        tags[i])

        self.logger.info(
            "Saved {} Regions {} into Tensorboard at epoch: {}".format(
                len(self.plot_data), sorted([*self.plot_data]), epoch))
        self.writer.flush()

    def write_region_metrics(self, epoch):
        metric_sum = 0
        means = {}
        for region in self.metric_data:
            metric = self.metric_data[region].mean()
            self.logger.info(
                "Epoch[{}] Metrics -- Region: {:0>2d}, {}: {:.4f}".format(
                    epoch, region, self.tag_name, metric))

            if self.merge_scalar:
                means["{:0>2d}".format(region)] = metric
            else:
                self.writer.add_scalar(
                    "{}_{:0>2d}".format(self.tag_name, region), metric, epoch)
            metric_sum += metric

        if self.merge_scalar:
            means["avg"] = metric_sum / len(self.metric_data)
            self.writer.add_scalars("{}_region".format(self.tag_name), means,
                                    epoch)
        elif len(self.metric_data) > 1:
            metric_avg = metric_sum / len(self.metric_data)
            self.writer.add_scalar("{}_regions_avg".format(self.tag_name),
                                   metric_avg, epoch)
        self.writer.flush()

    def __call__(self, engine: Engine, action) -> None:
        total_steps = engine.state.iteration
        if total_steps < engine.state.epoch_length:
            total_steps = engine.state.epoch_length * (engine.state.epoch -
                                                       1) + total_steps

        if action == "epoch" and not self.fold_size:
            epoch = engine.state.epoch
        elif self.fold_size and total_steps % self.fold_size == 0:
            epoch = int(total_steps / self.fold_size)
        else:
            epoch = None

        if epoch:
            if self.images and epoch % self.image_interval == 0:
                self.write_images(epoch)
            if self.add_scalar:
                self.write_region_metrics(epoch)

        if action == "epoch" or epoch:
            self.plot_data = {}
            self.metric_data = {}
            return

        device = engine.state.device
        batch_data = engine.state.batch
        output_data = engine.state.output

        for bidx in range(len(batch_data.get("region", []))):
            region = batch_data.get("region")[bidx]
            region = region.item() if torch.is_tensor(region) else region

            if self.images and self.plot_data.get(region) is None:
                self.plot_data[region] = [
                    rescale_array(
                        batch_data["image"][bidx][0].detach().cpu().numpy()[
                            np.newaxis], 0, 1),
                    rescale_array(
                        batch_data["label"][bidx].detach().cpu().numpy(), 0,
                        1),
                    rescale_array(
                        output_data["pred"][bidx].detach().cpu().numpy(), 0,
                        1),
                ]

            if self.compute_metric:
                if self.metric_data.get(region) is None:
                    self.metric_data[region] = RegionDice()
                self.metric_data[region].update(
                    y_pred=output_data["pred"][bidx].to(device),
                    y=batch_data["label"][bidx].to(device),
                    batched=False)
コード例 #4
0
)
from monai.utils import first, set_determinism
from numpy import math

print_config()
print("INFERENCE")

md5 = "410d4a301da4e5b2f6f86ec3ddba524e"

root_dir = "//home//imoreira//Data"
#root_dir = "C:\\Users\\isasi\\Downloads"
data_dir = os.path.join(root_dir, "Kidneys_Data")
out_dir = os.path.join(root_dir, "Kidneys_Best_Model")
tensorboard_dir = "//home//imoreira//Data//Tensorboard_Kidneys"

writer = SummaryWriter(log_dir=tensorboard_dir)
"""## Set dataset path"""

train_images = sorted(glob.glob(os.path.join(data_dir, "imagesTr",
                                             "*.nii.gz")))
train_labels = sorted(glob.glob(os.path.join(data_dir, "labelsTr",
                                             "*.nii.gz")))
data_dicts = [{
    "image": image_name,
    "label": label_name
} for image_name, label_name in zip(train_images, train_labels)]
#
# data_dicts = [
#     {"image": image_name}
#     for image_name in train_images
# ]