Skip to content

emthompson-usgs/shakemap-scenarios

 
 

Repository files navigation

Build Status codecov

shakemap-scenarios

Overview

Running a scenario is very different than running real events for a few reasons:

  • There isn't any data to be concerned about
  • The date/time doesn't really matter unless required for a specific exercise
  • Scenarios are organized in COMCAT by scenario catalogs, and each catalog needs to have an associated "catalog page" on the earthquake website that describes how the scenarios are created

In the future, all of this will be run within Python in ShakeMap 4.0. However, this code was created to address some issues that could not be handled in the current version of ShakeMap, such as the use of multiple GMPEs, and the inclusion of new GMPEs that are not available in ShakeMap 3.5. However, the code in this repository only handles the generation of the ground motion grids. The generation of products (e.g., maps, shapefiles, etc.) and transferring of the products to COMCAT is still handled with ShakeMap 3.5, but there are some helper functions in here to facilitate the interaction of the different codes.

Installation and Dependencies

  • ShakeMap 3.5: http://usgs.github.io/shakemap/manual_index.html
  • The python dependencies are the same as shakelib. See the setup_env.sh script.
  • The install.sh script installs this package and all dependencies. It is regularly tested on OSX and Ubuntu.

Workflow

Paths

You need to configure some paths in a conf file. Create a file in the user home directory named scenarios.conf with the following contents:

[system]
    # Base directory of shakemap 3.5 installation
    shakehome = /home/user/shake

    # PDL paths
    pdlbin = /home/user/ProductClient/ProductClient.jar
    key = /home/user/ProductClient/key
    pdlconf = /home/user/ProductClient/scenarioconfig.ini

    # Catalog name, i.e., "eventsource"
    catalog = test

[data]
    # Path to Vs30 grid
    vs30file = /home/user/data/Global_30_sec/global_vs30_ca_waor_ut_jp.grd


[modeling]
    # GMPE; found in openquake.hazardlib.gsim or specified in gmpe_sets.conf
    gmpe=active_crustal_nshmp2014

    # IPE; either VirtualIPE or found in shakelib.grind.ipe (does not yet exist)
    ipe=VirtualIPE  # unused

    # GMICE; found in shakelib.grind.gmice
    gmice = WGRW12  # unused

    # Cross correlation function
    ccf = LB13    # unused

[interp]
    # IMT list
    imt_list = PGA    # unused

    # Intensity measure component (IMC); currently either "Larger" or "RotD50"
    component = Larger

Note that the ipe and some other entries is currently a placeholder and cannot be changed.

Rupture Set

Generally, we start with scenario sources defined in JSON files. The subdirectory "rupture_sets" includes a few examples. Note that there is more than one acceptable format for the rupture sets because different sources of ruptures use different representations. Here is an example rupture set JSON file with only a single simple rupture (extracted from US_MT_2016.json):

{
    "name": "2016 Montana Scenarios",
    "info": "Derived by expert opinion with input from Michael Stickney, Kathy Haller.",
    "format":"nshmp",
    "events":
    [
        {
            "id":"greatfalls_se",
            "desc":"Great Falls",
            "mag":6.01,
            "width":12.3,
            "dip":90,
            "rake":0,
            "ztor":1.0,
            "lats":[47.542, 47.499],
            "lons":[-111.435, -111.406]
        }
    ]
}

If you create a rupture set for scenarios, I would like to archive them in this repository. So please send it in as a pull request or email the file to me. For more complex ruptures (e.g., multiple disconnected traces) or if you simply prefer to specify the lat/lon/depth of each quadrilateral corner, there is an example of how to to this in rupture_sets/example.json.

Input Directories

The mkinputdir command line program will generate an input directory for each of the events in a rupture set. The arguments are explained with the help message:

$ mkinputdir -h
usage: mkinputdir [-h] -f FILE [-r REFERENCE] [-d {-1,0,1,2}]
                  [-i [INDEX [INDEX ...]]] [-s SHAKEHOME]

Create ShakeMap input directory. This is designed primarily to be used with an
input file that provides rupture information.

optional arguments:
  -h, --help            show this help message and exit
  -f FILE, --file FILE  File with rupture information.
  -r REFERENCE, --reference REFERENCE
                        Reference for rupture source.
  -d {-1,0,1,2}, --dirind {-1,0,1,2}
                        Directivity; -1 for no directivity (default); 0 and 2
                        are the two opposing unilateral directions, 1 is for
                        bilateral.
  -i [INDEX [INDEX ...]], --index [INDEX [INDEX ...]]
                        List of rupture indices to run. Useful if you do not
                        want to run all ruptures in the file.

A few things to note:

  • The -f argument is the only required one, but the python argparse lists it as optional.
  • This is where directivity is selected if needed.
  • The resulting input directories will use the current time/date as the time/date for the scenario by default. This can be edited in the resulting event.xml file if required by the scenario exercise.

Grid Calculations

The runscenarios command line program will calculate the ground motion grids for all of events in the current ShakeMap data directory, which is basically just a wrapper around the mkscenariogrids program that runs on an individual event directory. I usually use runscenarios because I don't keep old event directories around in the data directory, and when I re-run things, I usually do the full set all at once. Some key things to note:

  • This is where you set what GMPE to use. Currently this only supports the NSHMP GMPE sets/weights, but it is easy to add new ones, or use a single GMPE.
  • runscenarios has an additional argument for the number of processors to use. If you are running a large number of events, it will run them in parallel, but the the code is not written to parallelize the calculations within a single scenario.

Run ShakeMap 3.5

The input directories now have all the required files, as well as the *estimates.grd and *sd.grd files. So ShakeMap 3.5 is run to generate the various products that are transferred to COMCAT. This runs ShakeMap 3.5 up through genex but does not run transfer.

Note that I have not written a command line argument for doing this, but there is a helper method run_one_old_shakemap in this repository. Here is a python script that uses this to run ShakeMap 3.5 on all of the events in the current data directory:

import os
from configobj import ConfigObj
from scenarios.utils import run_one_old_shakemap
config = ConfigObj(os.path.join(os.path.expanduser('~'), 'scenarios.conf'))
datadir = os.path.join(config['system']['shakehome'], 'data')
id_str = next(os.walk(datadir))[1]
n = len(id_str)
logs = [None]*n
for i in range(0, n):
    logs[i] = run_one_old_shakemap(id_str[i])

Note: just running ShakeMap 3.5 programs without this helper function probably will not work. Also, if anything goes wrong, the list of logs can be helpful for troubleshooting.

Transfer

IMPORTANT: Be sure to double/triple check that the catalog code is correct in [SHAKEHOME]/config/transfer.conf, which is specified in two places. For catalog code BSSC2014, this is what the pertinent section of the conf file should look like:

pdl_java : /usr/bin/java
pdl_client : /home/<user>/ProductClient/ProductClient.jar
pdl_source : us
pdl_type : shakemap
pdl_scenariotype : shakemap-scenario
pdl_code : bssc2014<EVENT>
pdl_eventsource : bssc2014
pdl_eventsourcecode : <EVENT>
pdl_privatekey : /home/<user>/ProductClient/comcat_atlaskey
pdl_config: /home/<user>/ProductClient/scenarioconfig.ini

If you mess this up, it creates havoc in COMCAT because of the complex nature of association, the lack of our ability to manually un-associate, and that we cannot really delete anything. It is extremely difficult to fix.

Unlike for older scenarios, we now have to also send an origin. Here is a script that sends origins for all of the events in the data directory:

import os
from scenarios.utils import send_origin
from configobj import ConfigObj
from impactutils.io.cmd import get_command_output
config = ConfigObj(os.path.join(os.path.expanduser('~'), 'scenarios.conf'))
datadir = os.path.join(config['system']['shakehome'], 'data')
id_str = next(os.walk(datadir))[1]
n = len(id_str)
logs = [None]*n
for i in range(0, n):
    logs[i] = send_origin(id_str[i])

To send the associated ShakeMap, we simply construct a call to the ShakeMap 3.5 transfer program:

import os
from configobj import ConfigObj
from impactutils.io.cmd import get_command_output
config = ConfigObj(os.path.join(os.path.expanduser('~'), 'scenarios.conf'))
datadir = os.path.join(config['system']['shakehome'], 'data')
shakebin = os.path.join(config['system']['shakehome'], 'bin')
id_str = next(os.walk(datadir))[1]
n = len(id_str)
logs = [None]*n
for i in range(n):
    calltransfer = shakebin + '/transfer -event ' + id_str[i] + ' -pdl -scenario'
    logs[i] = get_command_output(calltransfer)

Sometimes it is difficult to know if the transfer has succeeded because of caching. The logs are meant to help. It can also be helpful to know the caching rules:

Event age Cache time
7 days or less 1 min
30 days or less 15 min
More than 30 days 1 day

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 98.4%
  • Shell 1.6%