def test_remote_cloud(self): global base_port # FIX! we should increment this from 1 to N? for i in range(1, 10): # timeout wants to be larger for large numbers of hosts * node_count # don't want to reload jar/flatfile, so use build_cloud ### base_port += portsPerNode * node_count print "base_port:", base_port timeoutSecs = max(60, 8 * totalNodes) print "totalNodes:", totalNodes, "timeoutSecs:", timeoutSecs h2o.check_port_group(base_port) # FIX! would have to recopy the flatfile if we change base_port? hmm.. won't work ### h2o.write_flatfile(node_count=node_count, base_port=base_port) # FIX! ..just use_flatfile=False for now on these subsequent ones. rely on multicast # Could make the biggest cloud first and count down? supposed to not matter if flatfile has too many? start = time.time() h2o.build_cloud(node_count, base_port=base_port, hosts=h2o_hosts.hosts, use_flatfile=False, timeoutSecs=timeoutSecs, retryDelaySecs=0.5) print "Cloud of", len( h2o.nodes), "built in", time.time() - start, "seconds" check_cloud_and_setup_next()
def setUpClass(cls): # do the first one to build up hosts # so we don't repeatedly copy the jar # have to make sure base_port is the same on both! print "base_port:", base_port h2o.check_port_group(base_port) h2o.write_flatfile(node_count=node_count, base_port=base_port) start = time.time() h2o_hosts.build_cloud_with_hosts(node_count, base_port=base_port, use_flatfile=True, java_heap_GB=1) print "jar/flatfile copied and Cloud of", len(h2o.nodes), "built in", time.time()-start, "seconds" # have to remember total # of nodes for the next class. it will stay the same # when we tear down the cloud, we zero the nodes list global totalNodes totalNodes = len(h2o.nodes) check_cloud_and_setup_next()
def setUpClass(cls): # do the first one to build up hosts # so we don't repeatedly copy the jar # have to make sure base_port is the same on both! print "base_port:", base_port h2o.check_port_group(base_port) start = time.time() # writes it's own flatfile h2o_hosts.build_cloud_with_hosts(node_count, base_port=base_port, use_flatfile=True, java_heap_GB=1) print "jar/flatfile copied and Cloud of", len(h2o.nodes), "built in", time.time()-start, "seconds" # have to remember total # of nodes for the next class. it will stay the same # when we tear down the cloud, we zero the nodes list global totalNodes totalNodes = len(h2o.nodes) check_cloud_and_setup_next()
def test_remote_cloud(self): global base_port # FIX! we should increment this from 1 to N? for i in range(1,10): # timeout wants to be larger for large numbers of hosts * node_count # don't want to reload jar/flatfile, so use build_cloud base_port += portsPerNode * node_count print "base_port:", base_port timeoutSecs = max(60, 8 * totalNodes) print "totalNodes:", totalNodes, "timeoutSecs:", timeoutSecs h2o.check_port_group(base_port) # FIX! have to recopy the flatfile if we change base_port? hmm.. won't work h2o.write_flatfile(node_count=node_count, base_port=base_port) # FIX! ..just use_flatfile=False for now on these subsequent ones. rely on multicast start = time.time() h2o.build_cloud(node_count, base_port=base_port, hosts=h2o_hosts.hosts, use_flatfile=False, timeoutSecs=timeoutSecs, retryDelaySecs=0.5) print "Cloud of", len(h2o.nodes), "built in", time.time()-start, "seconds" check_cloud_and_setup_next()