'centerOfRotation': '281401.0288298179 -4639090.577582279 4354217.4974804', 'VerticalExaggeration': '10', 'speed': '0.1', } }, grdTerrain=os.path.join(parentDir, 'michigan_lld.grd')) # Set start and end dates for all loads from sources that contain data # beyond the temporal bounds of the campaign # startdate = datetime.datetime(2016, 7, 24) # Fixed start enddate = datetime.datetime(2016, 8, 24) # Fixed end # default location of thredds and dods data: cl.tdsBase = 'http://odss.mbari.org/thredds/' cl.dodsBase = cl.tdsBase + 'dodsC/' ##################################################################### # LRAUV ##################################################################### def find_urls(base, search_str): INV_NS = "http://www.unidata.ucar.edu/namespaces/thredds/InvCatalog/v1.0" url = os.path.join(base, 'catalog.xml') print "Crawling: %s" % url skips = Crawl.SKIPS + [ ".*Courier*", ".*Express*", ".*Normal*, '.*Priority*", ".*.cfg$" ] u = urlparse.urlsplit(url) name, ext = os.path.splitext(u.path)
'VerticalExaggeration': '10', 'speed': '0.1', } }, grdTerrain = os.path.join(parentDir, 'michigan_lld.grd') ) # Set start and end dates for all loads from sources that contain data # beyond the temporal bounds of the campaign # startdate = datetime.datetime(2016, 7, 24) # Fixed start enddate = datetime.datetime(2016, 8, 24) # Fixed end # default location of thredds and dods data: cl.tdsBase = 'http://odss.mbari.org/thredds/' cl.dodsBase = cl.tdsBase + 'dodsC/' ##################################################################### # LRAUV ##################################################################### def find_urls(base, search_str): INV_NS = "http://www.unidata.ucar.edu/namespaces/thredds/InvCatalog/v1.0" url = os.path.join(base, 'catalog.xml') print "Crawling: %s" % url skips = Crawl.SKIPS + [".*Courier*", ".*Express*", ".*Normal*, '.*Priority*", ".*.cfg$" ] u = urlparse.urlsplit(url) name, ext = os.path.splitext(u.path) if ext == ".html": u = urlparse.urlsplit(url.replace(".html", ".xml")) url = u.geturl()