Пример #1
0
config = EasyDict()

config.cols_to_use = [0,1,2,3]
# may want to change: feat_size, layer_dims, etc. 
config.model_kwargs = {"num_features":1, "num_classes": 3}
config.data_path = "/fast_scratch/WatChMaL/data/pointnet/pointnet_trainval.h5"
config.indices_file = "/fast_scratch/WatChMaL/data/pointnet/pointnet_trainval_idxs.npz"

#make sure to change this
config.dump_path = "/home/dgreen/training_outputs/pointnet2/no_time/adam/"

config.num_data_workers = 0 # Sometime crashes if we do multiprocessing
config.device = 'cuda:6'

config.optimizer = "Adam"
config.optimizer_kwargs = {"lr":1e-3, "betas": (0.9, 0.999)}

config.use_scheduler = False
config.scheduler_kwargs = {"mode":"min", "min_lr":1e-6, "patience":1, "verbose":True}
config.scheduler_step = 190

config.batch_size = 32
config.epochs = 20

config.report_interval = 200
config.num_val_batches  = 256
config.valid_interval   = 1000

config.validate_batch_size = 32
config.validate_dump_interval = 256
Пример #2
0
## Data paths
config.data_path = "/app/test_data/split_h5/IWCDmPMT_4pi_fulltank_test_graphnet_trainval.h5"
config.indices_file = "/app/test_data/split_h5/IWCDmPMT_4pi_fulltank_test_graphnet_trainval_idxs.npz"
config.edge_index_pickle = "/app/GraphNets/metadata/edges_dict.pkl"

## Log location
config.dump_path = "/app/GraphNets/dump/gcn"

## Computer Parameters
config.num_data_workers = 0  # Sometime crashes if we do multiprocessing
config.device = 'gpu'
config.gpu_list = [0]

# Optimizer Parameters
config.optimizer = "Adam"
config.optimizer_kwargs = {"lr": 0.01, "weight_decay": 5e-4}

## Training parameters
config.batch_size = 32
config.epochs = 1

## Logging parameters for training
config.report_interval = 10  # 100
config.num_val_batches = 32
config.valid_interval = 100  # 10000

## Validating parameters
config.validate_batch_size = 32
config.validate_dump_interval = 256
Пример #3
0
}

config.data_path = "/data/WatChMaL/data/IWCDmPMT_4pi_fulltank_test/split_h5/IWCDmPMT_4pi_fulltank_test_graphnet_trainval.h5"
config.indices_file = "/data/WatChMaL/data/IWCDmPMT_4pi_fulltank_test/split_h5/IWCDmPMT_4pi_fulltank_test_graphnet_trainval_idxs.npz"
config.edge_index_pickle = "/data/WatChMaL/graphnets/visualization/mpmt_edges_dict.pkl"

config.dump_path = "/data/WatChMaL/graphnets/GraphNets/dump/" + config.model_name

config.num_data_workers = 0  # Sometime crashes if we do multiprocessing
config.device = 'gpu'
config.gpu_list = [6]

config.optimizer = "SGD"
config.optimizer_kwargs = {
    "lr": 0.01,
    "weight_decay": 1e-3,
    "momentum": 0.9,
    "nesterov": True
}

config.scheduler_kwargs = {
    "mode": "min",
    "min_lr": 1e-6,
    "patience": 1,
    "verbose": True
}
config.scheduler_step = 190

config.batch_size = 64
config.epochs = 15

config.report_interval = 50
Пример #4
0
## Data paths
config.data_path = "/fast_scratch/NeutronGNN/iwcd_mpmt_shorttank_neutrongnn_trainval.h5"
config.indices_file = "/fast_scratch/NeutronGNN/iwcd_mpmt_shorttank_neutrongnn_trainval_idxs.npz"

## Log location
config.dump_path = "/fast_scratch/NeutronGNN/dump/" + config.model_name

## Computer parameters
config.num_data_workers = 0 # Sometime crashes if we do multiprocessing
config.device = 'gpu'
config.gpu_list = [1]

## Optimizer parameters
config.optimizer = "Adam"
config.optimizer_kwargs = {"lr":1e-3}
#config.optimizer_kwargs = {"lr":0.01, "weight_decay":5e-4}

## Scheduler parameters
config.scheduler_kwargs = {"mode":"min", "min_lr":1e-6, "patience":1, "verbose":True}
config.scheduler_step = 190

## Training parameters
config.batch_size = 64
config.epochs = 5

## Logging parameters for training
config.report_interval = 50
config.num_val_batches  = 16
config.valid_interval   = 200