SetFetchDepth(), PreparePaths(), WriteInfo(), DownloadUrlFile(control), WgetDownload(wpull_args, accept_on_exit_code=AcceptAny(), env={ 'ITEM_IDENT': ItemInterpolation('%(ident)s'), 'LOG_KEY': ItemInterpolation('%(log_key)s'), 'REDIS_URL': REDIS_URL, 'PATH': os.environ['PATH'] }), RelabelIfAborted(control), WriteInfo(), MoveFiles(), LimitConcurrent( 2, RsyncUpload(target=RSYNC_URL, target_source_path=ItemInterpolation("%(data_dir)s"), files=ItemValue("all_target_files"), extra_args=['--partial', '--partial-dir', '.rsync-tmp'])), StopHeartbeat(), MarkItemAsDone(control, EXPIRE_TIME)) def stop_control(): #control.flag_logging_thread_for_termination() control.unregister_pipeline(pipeline_id) pipeline.on_cleanup += stop_control pipeline.running_status = "Running" def status_running(): pipeline.running_status = "Running"
), RelabelIfAborted(control), WriteInfo(), MoveFiles(), LimitConcurrent(2, RsyncUpload( target = RSYNC_URL, target_source_path = ItemInterpolation("%(data_dir)s"), files=ItemValue("all_target_files"), extra_args = [ '--partial', '--partial-dir', '.rsync-tmp' ] ) ), StopHeartbeat(), MarkItemAsDone(control, EXPIRE_TIME) ) def stop_control(): control.unregister_pipeline(pipeline_id) pipeline.on_cleanup += stop_control pipeline.running_status = "Running" def status_running(): pipeline.running_status = "Running" pipeline.on_stop_canceled += status_running