def main(cfg): # Start kraken print(pyfiglet.figlet_format("kraken")) logging.info("Starting kraken") # Parse and read the config if os.path.isfile(cfg): with open(cfg, 'r') as f: config = yaml.full_load(f) global kubeconfig_path, wait_duration kubeconfig_path = config["kraken"].get("kubeconfig_path", "") chaos_scenarios = config["kraken"].get("chaos_scenarios", []) litmus_version = config['kraken'].get("litmus_version", 'v1.9.1') litmus_uninstall = config['kraken'].get("litmus_uninstall", False) wait_duration = config["tunings"].get("wait_duration", 60) iterations = config["tunings"].get("iterations", 1) daemon_mode = config["tunings"].get("daemon_mode", False) deploy_performance_dashboards = config["performance_monitoring"].get( "deploy_dashboards", False) dashboard_repo = config["performance_monitoring"].get( "repo", "https://github.com/cloud-bulldozer/performance-dashboards.git" ) # noqa # Initialize clients if not os.path.isfile(kubeconfig_path): kubeconfig_path = None logging.info("Initializing client to talk to the Kubernetes cluster") kubecli.initialize_clients(kubeconfig_path) # find node kraken might be running on kubecli.find_kraken_node() # Cluster info logging.info("Fetching cluster info") cluster_version = runcommand.invoke("kubectl get clusterversion") cluster_info = runcommand.invoke( "kubectl cluster-info | awk 'NR==1' | sed -r " "'s/\x1B\[([0-9]{1,3}(;[0-9]{1,2})?)?[mGK]//g'") # noqa logging.info("\n%s%s" % (cluster_version, cluster_info)) # Deploy performance dashboards if deploy_performance_dashboards: performance_dashboards.setup(dashboard_repo) # Initialize the start iteration to 0 iteration = 0 # Set the number of iterations to loop to infinity if daemon mode is # enabled or else set it to the provided iterations count in the config if daemon_mode: logging.info( "Daemon mode enabled, kraken will cause chaos forever\n") logging.info("Ignoring the iterations set") iterations = float('inf') else: logging.info( "Daemon mode not enabled, will run through %s iterations\n" % str(iterations)) iterations = int(iterations) failed_post_scenarios = [] litmus_namespaces = [] litmus_installed = False # Loop to run the chaos starts here while (int(iteration) < iterations): # Inject chaos scenarios specified in the config logging.info("Executing scenarios for iteration " + str(iteration)) if chaos_scenarios: for scenario in chaos_scenarios: scenario_type = list(scenario.keys())[0] scenarios_list = scenario[scenario_type] if scenarios_list: # Inject pod chaos scenarios specified in the config if scenario_type == "pod_scenarios": failed_post_scenarios = pod_scenarios( scenarios_list, config, failed_post_scenarios) # Inject node chaos scenarios specified in the config elif scenario_type == "node_scenarios": node_scenarios(scenarios_list, config) # Inject time skew chaos scenarios specified in the config elif scenario_type == "time_scenarios": time_scenarios(scenarios_list, config) elif scenario_type == "litmus_scenarios": if not litmus_installed: common_litmus.install_litmus(litmus_version) common_litmus.deploy_all_experiments( litmus_version) litmus_installed = True litmus_namespaces = litmus_scenarios( scenarios_list, config, litmus_namespaces, litmus_uninstall) iteration += 1 logging.info("") if litmus_uninstall and litmus_installed: for namespace in litmus_namespaces: common_litmus.delete_chaos(namespace) common_litmus.delete_experiments() common_litmus.uninstall_litmus(litmus_version) if failed_post_scenarios: logging.error( "Post scenarios are still failing at the end of all iterations" ) sys.exit(1) else: logging.error("Cannot find a config at %s, please check" % (cfg)) sys.exit(1)
def main(cfg): # Start kraken print(pyfiglet.figlet_format("kraken")) logging.info("Starting kraken") # Parse and read the config if os.path.isfile(cfg): with open(cfg, "r") as f: config = yaml.full_load(f) global kubeconfig_path, wait_duration distribution = config["kraken"].get("distribution", "openshift") kubeconfig_path = config["kraken"].get("kubeconfig_path", "") chaos_scenarios = config["kraken"].get("chaos_scenarios", []) publish_running_status = config["kraken"].get("publish_kraken_status", False) port = config["kraken"].get("port", "8081") run_signal = config["kraken"].get("signal_state", "RUN") litmus_version = config["kraken"].get("litmus_version", "v1.9.1") litmus_uninstall = config["kraken"].get("litmus_uninstall", False) wait_duration = config["tunings"].get("wait_duration", 60) iterations = config["tunings"].get("iterations", 1) daemon_mode = config["tunings"].get("daemon_mode", False) deploy_performance_dashboards = config["performance_monitoring"].get( "deploy_dashboards", False) dashboard_repo = config["performance_monitoring"].get( "repo", "https://github.com/cloud-bulldozer/performance-dashboards.git" ) # noqa capture_metrics = config["performance_monitoring"].get( "capture_metrics", False) kube_burner_url = config["performance_monitoring"].get( "kube_burner_binary_url", "https://github.com/cloud-bulldozer/kube-burner/releases/download/v0.9.1/kube-burner-0.9.1-Linux-x86_64.tar.gz", # noqa ) config_path = config["performance_monitoring"].get( "config_path", "config/kube_burner.yaml") metrics_profile = config["performance_monitoring"].get( "metrics_profile_path", "config/metrics-aggregated.yaml") prometheus_url = config["performance_monitoring"].get( "prometheus_url", "") prometheus_bearer_token = config["performance_monitoring"].get( "prometheus_bearer_token", "") run_uuid = config["performance_monitoring"].get("uuid", "") enable_alerts = config["performance_monitoring"].get( "enable_alerts", False) alert_profile = config["performance_monitoring"].get( "alert_profile", "") # Initialize clients if not os.path.isfile(kubeconfig_path): logging.error( "Cannot read the kubeconfig file at %s, please check" % kubeconfig_path) sys.exit(1) logging.info("Initializing client to talk to the Kubernetes cluster") os.environ["KUBECONFIG"] = str(kubeconfig_path) kubecli.initialize_clients(kubeconfig_path) # find node kraken might be running on kubecli.find_kraken_node() # Set up kraken url to track signal if not 0 <= int(port) <= 65535: logging.info("Using port 8081 as %s isn't a valid port number" % (port)) port = 8081 address = ("0.0.0.0", port) # If publish_running_status is False this should keep us going in our loop below if publish_running_status: server_address = address[0] port = address[1] logging.info("Publishing kraken status at http://%s:%s" % (server_address, port)) server.start_server(address) publish_kraken_status(run_signal) # Cluster info logging.info("Fetching cluster info") cluster_version = runcommand.invoke("kubectl get clusterversion", 60) cluster_info = runcommand.invoke( "kubectl cluster-info | awk 'NR==1' | sed -r " "'s/\x1B\[([0-9]{1,3}(;[0-9]{1,2})?)?[mGK]//g'", 60) # noqa logging.info("\n%s%s" % (cluster_version, cluster_info)) # Deploy performance dashboards if deploy_performance_dashboards: performance_dashboards.setup(dashboard_repo) # Generate uuid for the run if run_uuid: logging.info("Using the uuid defined by the user for the run: %s" % run_uuid) else: run_uuid = str(uuid.uuid4()) logging.info("Generated a uuid for the run: %s" % run_uuid) # Initialize the start iteration to 0 iteration = 0 # Set the number of iterations to loop to infinity if daemon mode is # enabled or else set it to the provided iterations count in the config if daemon_mode: logging.info( "Daemon mode enabled, kraken will cause chaos forever\n") logging.info("Ignoring the iterations set") iterations = float("inf") else: logging.info( "Daemon mode not enabled, will run through %s iterations\n" % str(iterations)) iterations = int(iterations) failed_post_scenarios = [] litmus_installed = False # Capture the start time start_time = int(time.time()) # Loop to run the chaos starts here while int(iteration) < iterations and run_signal != "STOP": # Inject chaos scenarios specified in the config logging.info("Executing scenarios for iteration " + str(iteration)) if chaos_scenarios: for scenario in chaos_scenarios: if publish_running_status: run_signal = server.get_status(address) if run_signal == "PAUSE": while publish_running_status and run_signal == "PAUSE": logging.info( "Pausing Kraken run, waiting for %s seconds and will re-poll signal" % str(wait_duration)) time.sleep(wait_duration) run_signal = server.get_status(address) if run_signal == "STOP": logging.info("Received STOP signal; ending Kraken run") break scenario_type = list(scenario.keys())[0] scenarios_list = scenario[scenario_type] if scenarios_list: # Inject pod chaos scenarios specified in the config if scenario_type == "pod_scenarios": logging.info("Running pod scenarios") failed_post_scenarios = pod_scenarios.run( kubeconfig_path, scenarios_list, config, failed_post_scenarios, wait_duration) elif scenario_type == "container_scenarios": logging.info("Running container scenarios") failed_post_scenarios = pod_scenarios.container_run( kubeconfig_path, scenarios_list, config, failed_post_scenarios, wait_duration) # Inject node chaos scenarios specified in the config elif scenario_type == "node_scenarios": logging.info("Running node scenarios") nodeaction.run(scenarios_list, config, wait_duration) # Inject time skew chaos scenarios specified in the config elif scenario_type == "time_scenarios": logging.info("Running time skew scenarios") time_actions.run(scenarios_list, config, wait_duration) # Inject litmus based chaos scenarios elif scenario_type == "litmus_scenarios": logging.info("Running litmus scenarios") litmus_namespace = "litmus" if not litmus_installed: # Will always uninstall first common_litmus.delete_chaos(litmus_namespace) common_litmus.delete_chaos_experiments( litmus_namespace) common_litmus.uninstall_litmus( litmus_version, litmus_namespace) common_litmus.install_litmus( litmus_version, litmus_namespace) common_litmus.deploy_all_experiments( litmus_version, litmus_namespace) litmus_installed = True common_litmus.run( scenarios_list, config, litmus_uninstall, wait_duration, litmus_namespace, ) # Inject cluster shutdown scenarios elif scenario_type == "cluster_shut_down_scenarios": shut_down.run(scenarios_list, config, wait_duration) # Inject namespace chaos scenarios elif scenario_type == "namespace_scenarios": logging.info("Running namespace scenarios") namespace_actions.run(scenarios_list, config, wait_duration, failed_post_scenarios, kubeconfig_path) # Inject zone failures elif scenario_type == "zone_outages": logging.info("Inject zone outages") zone_outages.run(scenarios_list, config, wait_duration) # Application outages elif scenario_type == "application_outages": logging.info("Injecting application outage") application_outage.run(scenarios_list, config, wait_duration) # PVC scenarios elif scenario_type == "pvc_scenarios": logging.info("Running PVC scenario") pvc_scenario.run(scenarios_list, config) iteration += 1 logging.info("") # Capture the end time end_time = int(time.time()) # Capture metrics for the run if capture_metrics: logging.info("Capturing metrics") kube_burner.setup(kube_burner_url) kube_burner.scrape_metrics( distribution, run_uuid, prometheus_url, prometheus_bearer_token, start_time, end_time, config_path, metrics_profile, ) # Check for the alerts specified if enable_alerts: logging.info("Alerts checking is enabled") kube_burner.setup(kube_burner_url) if alert_profile: kube_burner.alerts( distribution, prometheus_url, prometheus_bearer_token, start_time, end_time, alert_profile, ) else: logging.error("Alert profile is not defined") sys.exit(1) if litmus_uninstall and litmus_installed: common_litmus.delete_chaos(litmus_namespace) common_litmus.delete_chaos_experiments(litmus_namespace) common_litmus.uninstall_litmus(litmus_version, litmus_namespace) if failed_post_scenarios: logging.error( "Post scenarios are still failing at the end of all iterations" ) sys.exit(1) run_dir = os.getcwd() + "/kraken.report" logging.info( "Successfully finished running Kraken. UUID for the run: %s. Report generated at %s. Exiting" % (run_uuid, run_dir)) else: logging.error("Cannot find a config at %s, please check" % (cfg)) sys.exit(1)
def main(cfg): # Start kraken print(pyfiglet.figlet_format("kraken")) logging.info("Starting kraken") # Parse and read the config if os.path.isfile(cfg): with open(cfg, "r") as f: config = yaml.full_load(f) global kubeconfig_path, wait_duration distribution = config["kraken"].get("distribution", "openshift") kubeconfig_path = config["kraken"].get("kubeconfig_path", "") chaos_scenarios = config["kraken"].get("chaos_scenarios", []) litmus_version = config["kraken"].get("litmus_version", "v1.9.1") litmus_uninstall = config["kraken"].get("litmus_uninstall", False) wait_duration = config["tunings"].get("wait_duration", 60) iterations = config["tunings"].get("iterations", 1) daemon_mode = config["tunings"].get("daemon_mode", False) deploy_performance_dashboards = config["performance_monitoring"].get("deploy_dashboards", False) dashboard_repo = config["performance_monitoring"].get( "repo", "https://github.com/cloud-bulldozer/performance-dashboards.git" ) # noqa capture_metrics = config["performance_monitoring"].get("capture_metrics", False) kube_burner_url = config["performance_monitoring"].get( "kube_burner_binary_url", "https://github.com/cloud-bulldozer/kube-burner/releases/download/v0.9.1/kube-burner-0.9.1-Linux-x86_64.tar.gz", # noqa ) config_path = config["performance_monitoring"].get("config_path", "config/kube_burner.yaml") metrics_profile = config["performance_monitoring"].get("metrics_profile_path", "config/metrics-aggregated.yaml") prometheus_url = config["performance_monitoring"].get("prometheus_url", "") prometheus_bearer_token = config["performance_monitoring"].get("prometheus_bearer_token", "") run_uuid = config["performance_monitoring"].get("uuid", "") enable_alerts = config["performance_monitoring"].get("enable_alerts", False) alert_profile = config["performance_monitoring"].get("alert_profile", "") # Initialize clients if not os.path.isfile(kubeconfig_path): kubeconfig_path = None logging.info("Initializing client to talk to the Kubernetes cluster") kubecli.initialize_clients(kubeconfig_path) # find node kraken might be running on kubecli.find_kraken_node() # Cluster info logging.info("Fetching cluster info") cluster_version = runcommand.invoke("kubectl get clusterversion") cluster_info = runcommand.invoke( "kubectl cluster-info | awk 'NR==1' | sed -r " "'s/\x1B\[([0-9]{1,3}(;[0-9]{1,2})?)?[mGK]//g'" ) # noqa logging.info("\n%s%s" % (cluster_version, cluster_info)) # Deploy performance dashboards if deploy_performance_dashboards: performance_dashboards.setup(dashboard_repo) # Generate uuid for the run if run_uuid: logging.info("Using the uuid defined by the user for the run: %s" % run_uuid) else: run_uuid = str(uuid.uuid4()) logging.info("Generated a uuid for the run: %s" % run_uuid) # Initialize the start iteration to 0 iteration = 0 # Set the number of iterations to loop to infinity if daemon mode is # enabled or else set it to the provided iterations count in the config if daemon_mode: logging.info("Daemon mode enabled, kraken will cause chaos forever\n") logging.info("Ignoring the iterations set") iterations = float("inf") else: logging.info("Daemon mode not enabled, will run through %s iterations\n" % str(iterations)) iterations = int(iterations) failed_post_scenarios = [] litmus_namespaces = [] litmus_installed = False # Capture the start time start_time = int(time.time()) # Loop to run the chaos starts here while int(iteration) < iterations: # Inject chaos scenarios specified in the config logging.info("Executing scenarios for iteration " + str(iteration)) if chaos_scenarios: for scenario in chaos_scenarios: scenario_type = list(scenario.keys())[0] scenarios_list = scenario[scenario_type] if scenarios_list: # Inject pod chaos scenarios specified in the config if scenario_type == "pod_scenarios": failed_post_scenarios = pod_scenarios.run( kubeconfig_path, scenarios_list, config, failed_post_scenarios, wait_duration ) # Inject node chaos scenarios specified in the config elif scenario_type == "node_scenarios": nodeaction.run(scenarios_list, config, wait_duration) # Inject time skew chaos scenarios specified in the config elif scenario_type == "time_scenarios": time_actions.run(scenarios_list, config, wait_duration) elif scenario_type == "litmus_scenarios": if not litmus_installed: common_litmus.install_litmus(litmus_version) common_litmus.deploy_all_experiments(litmus_version) litmus_installed = True litmus_namespaces = common_litmus.run( scenarios_list, config, litmus_namespaces, litmus_uninstall, wait_duration ) elif scenario_type == "cluster_shut_down_scenarios": shut_down.run(scenarios_list, config, wait_duration) iteration += 1 logging.info("") # Capture the end time end_time = int(time.time()) # Capture metrics for the run if capture_metrics: logging.info("Capturing metrics") kube_burner.setup(kube_burner_url) kube_burner.scrape_metrics( distribution, run_uuid, prometheus_url, prometheus_bearer_token, start_time, end_time, config_path, metrics_profile, ) # Check for the alerts specified if enable_alerts: logging.info("Alerts checking is enabled") kube_burner.setup(kube_burner_url) if alert_profile: kube_burner.alerts( distribution, prometheus_url, prometheus_bearer_token, start_time, end_time, alert_profile, ) else: logging.error("Alert profile is not defined") sys.exit(1) if litmus_uninstall and litmus_installed: for namespace in litmus_namespaces: common_litmus.delete_chaos(namespace) common_litmus.delete_experiments() common_litmus.uninstall_litmus(litmus_version) if failed_post_scenarios: logging.error("Post scenarios are still failing at the end of all iterations") sys.exit(1) else: logging.error("Cannot find a config at %s, please check" % (cfg)) sys.exit(1)