def init(logFilenameArg, defaultModuleSpecified, **params): # # Do not add in CWD to sys.path # cwd = os.getcwd() # if cwd not in sys.path: # sys.path = [cwd] + sys.path #initialize logging if logFilenameArg is not None: try: set_log_file(open(logFilenameArg, 'w')) except: print "ERROR: log file open failed, logging is disabled. Please verify the log file path is correct." init_spark(**params) #persist=persistState, resume=resumeState, persistIntentions=persistIntentionState global runAgent runAgent = new_agent(runAgent, step_init_fn, PRINT_TRACE) #if we have loaded from persist data, we still override switch to the #default module specified on the command line. this has the nice behavior #that if there is no persisted data, SPARK loads normally. we also #switch to the default module if we did not load from persisted data #print "ready to load default module", runAgent.loaded_from_persist_data, defaultModuleSpecified if not runAgent.loaded_from_persist_data or defaultModuleSpecified: return load_module(runAgent, defaultModule)
def jd_start_debugger(defaultModule, initParams=None): print "" print "Welcome to the Java debugger for SPARK v%s"%VERSION print "" if initParams is None: initParams = {} initParams['persist']=False initParams['resume']=False else: if initParams.get('unbuffered') is not None: from spark.main import nobuffer sys.stdout=nobuffer(sys.stdout) print "Output is unbuffered" del initParams['unbuffered'] elif initParams.get('python') is not None: import code code.interact() del initParams['python'] elif initParams.get('coverage'): import coverage coverage.start() del initParams['coverage'] p = {} keys = initParams.keys() for key in keys: p[key] = initParams.get(key) initParams = p try: init_spark(**initParams) jd_new_agent() print "test agent initialized" load_module(testagent, defaultModule) return None except LocatedError, err: return TermErrorWrapper(err)
def jd_start_debugger(defaultModule, initParams=None): print "" print "Welcome to the Java debugger for SPARK v%s" % VERSION print "" if initParams is None: initParams = {} initParams['persist'] = False initParams['resume'] = False else: if initParams.get('unbuffered') is not None: from spark.main import nobuffer sys.stdout = nobuffer(sys.stdout) print "Output is unbuffered" del initParams['unbuffered'] elif initParams.get('python') is not None: import code code.interact() del initParams['python'] elif initParams.get('coverage'): import coverage coverage.start() del initParams['coverage'] p = {} keys = initParams.keys() for key in keys: p[key] = initParams.get(key) initParams = p try: init_spark(**initParams) jd_new_agent() print "test agent initialized" load_module(testagent, defaultModule) return None except LocatedError, err: return TermErrorWrapper(err)
def main(argv): from spark.internal.init import init_spark params = {} init_spark(**params) #no persist resume if len(argv) > 1: if argv[1] == "-test": new_agent() modules = get_files() generate_indexes(modules) doc_module(default_module) generate_all_symbols_indexes(modules) return new_agent() modules = get_files() print "====================", modules generate_indexes(modules) print "====================" for mod in modules: print "====================" doc_module(mod) #generate all symbols must run last generate_all_symbols_indexes(modules)