def get_task(self, name): """See docstring in class Loader. """ try: num = self.tasks_order[name] # Here we expose an undocumented behavior, so that cmsMake can # import a task even without the whole contest; this is not to # be relied upon in general except AttributeError: num = 1 task_path = os.path.join(self.path, name) # We first look for the yaml file inside the task folder, # and eventually fallback to a yaml file in its parent folder. try: conf = yaml.safe_load( io.open(os.path.join(task_path, "task.yaml"), "rt", encoding="utf-8")) except IOError: conf = yaml.safe_load( io.open(os.path.join(self.path, name + ".yaml"), "rt", encoding="utf-8")) logger.info("Loading parameters for task %s." % name) # Here we update the time of the last import touch(os.path.join(task_path, ".itime")) # If this file is not deleted, then the import failed touch(os.path.join(task_path, ".import_error")) args = {} args["num"] = num load(conf, args, ["name", "nome_breve"]) load(conf, args, ["title", "nome"]) assert name == args["name"] if args["name"] == args["title"]: logger.warning("Short name equals long name (title). " "Please check.") primary_language = load(conf, None, "primary_language") if primary_language is None: primary_language = 'it' paths = [os.path.join(task_path, "statement", "statement.pdf"), os.path.join(task_path, "testo", "testo.pdf")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Statement for task %s (lang: %s)" % (name, primary_language)) break else: logger.critical("Couldn't find any task statement, aborting...") sys.exit(1) args["statements"] = [Statement(primary_language, digest)] args["primary_statements"] = '["%s"]' % (primary_language) args["attachments"] = [] # FIXME Use auxiliary args["submission_format"] = [ SubmissionFormatElement("%s.%%l" % name)] # Use the new token settings format if detected. if "token_mode" in conf: load(conf, args, "token_mode") load(conf, args, "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_interval", conv=make_timedelta) load(conf, args, "token_gen_max") # Otherwise fall back on the old one. else: logger.warning( "%s.yaml uses a deprecated format for token settings which " "will soon stop being supported, you're advised to update it.", name) # Determine the mode. if conf.get("token_initial", None) is None: args["token_mode"] = "disabled" elif conf.get("token_gen_number", 0) > 0 and \ conf.get("token_gen_time", 0) == 0: args["token_mode"] = "infinite" else: args["token_mode"] = "finite" # Set the old default values. args["token_gen_initial"] = 0 args["token_gen_number"] = 0 args["token_gen_interval"] = timedelta() # Copy the parameters to their new names. load(conf, args, "token_total", "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_initial", "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_time", "token_gen_interval", conv=make_timedelta) load(conf, args, "token_max", "token_gen_max") # Remove some corner cases. if args["token_gen_initial"] is None: args["token_gen_initial"] = 0 if args["token_gen_interval"].total_seconds() == 0: args["token_gen_interval"] = timedelta(minutes=1) load(conf, args, "max_submission_number") load(conf, args, "max_user_test_number") load(conf, args, "min_submission_interval", conv=make_timedelta) load(conf, args, "min_user_test_interval", conv=make_timedelta) # Attachments args["attachments"] = [] if os.path.exists(os.path.join(task_path, "att")): for filename in os.listdir(os.path.join(task_path, "att")): digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "att", filename), "Attachment %s for task %s" % (filename, name)) args["attachments"] += [Attachment(filename, digest)] task = Task(**args) args = {} args["task"] = task args["description"] = conf.get("version", "Default") args["autojudge"] = False load(conf, args, ["time_limit", "timeout"], conv=float) load(conf, args, ["memory_limit", "memlimit"]) # Builds the parameters that depend on the task type args["managers"] = [] infile_param = conf.get("infile", "input.txt") outfile_param = conf.get("outfile", "output.txt") # If there is sol/grader.%l for some language %l, then, # presuming that the task type is Batch, we retrieve graders # in the form sol/grader.%l graders = False for lang in LANGUAGES: if os.path.exists(os.path.join( task_path, "sol", "grader.%s" % lang)): graders = True break if graders: # Read grader for each language for lang in LANGUAGES: grader_filename = os.path.join( task_path, "sol", "grader.%s" % lang) if os.path.exists(grader_filename): digest = self.file_cacher.put_file_from_path( grader_filename, "Grader for task %s and language %s" % (name, lang)) args["managers"] += [ Manager("grader.%s" % lang, digest)] else: logger.warning("Grader for language %s not found " % lang) # Read managers with other known file extensions for other_filename in os.listdir(os.path.join(task_path, "sol")): if other_filename.endswith('.h') or \ other_filename.endswith('lib.pas'): digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "sol", other_filename), "Manager %s for task %s" % (other_filename, name)) args["managers"] += [ Manager(other_filename, digest)] compilation_param = "grader" else: compilation_param = "alone" # If there is check/checker (or equivalent), then, presuming # that the task type is Batch or OutputOnly, we retrieve the # comparator paths = [os.path.join(task_path, "check", "checker"), os.path.join(task_path, "cor", "correttore")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % name) args["managers"] += [ Manager("checker", digest)] evaluation_param = "comparator" break else: evaluation_param = "diff" # Detect subtasks by checking GEN gen_filename = os.path.join(task_path, 'gen', 'GEN') try: with io.open(gen_filename, "rt", encoding="utf-8") as gen_file: subtasks = [] testcases = 0 points = None for line in gen_file: line = line.strip() splitted = line.split('#', 1) if len(splitted) == 1: # This line represents a testcase, otherwise it's # just a blank if splitted[0] != '': testcases += 1 else: testcase, comment = splitted testcase_detected = False subtask_detected = False if testcase.strip() != '': testcase_detected = True comment = comment.strip() if comment.startswith('ST:'): subtask_detected = True if testcase_detected and subtask_detected: raise Exception("No testcase and subtask in the" " same line allowed") # This line represents a testcase and contains a # comment, but the comment doesn't start a new # subtask if testcase_detected: testcases += 1 # This line starts a new subtask if subtask_detected: # Close the previous subtask if points is None: assert(testcases == 0) else: subtasks.append([points, testcases]) # Open the new one testcases = 0 points = int(comment[3:].strip()) # Close last subtask (if no subtasks were defined, just # fallback to Sum) if points is None: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = testcases if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = "%s" % input_value else: subtasks.append([points, testcases]) assert(100 == sum([int(st[0]) for st in subtasks])) n_input = sum([int(st[1]) for st in subtasks]) args["score_type"] = "GroupMin" args["score_type_parameters"] = "%s" % subtasks if "n_input" in conf: assert int(conf['n_input']) == n_input # If gen/GEN doesn't exist, just fallback to Sum except IOError: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = int(conf['n_input']) if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = "%s" % input_value # If output_only is set, then the task type is OutputOnly if conf.get('output_only', False): args["task_type"] = "OutputOnly" args["time_limit"] = None args["memory_limit"] = None args["task_type_parameters"] = '["%s"]' % evaluation_param task.submission_format = [ SubmissionFormatElement("output_%03d.txt" % i) for i in xrange(n_input)] # If there is check/manager (or equivalent), then the task # type is Communication else: paths = [os.path.join(task_path, "check", "manager"), os.path.join(task_path, "cor", "manager")] for path in paths: if os.path.exists(path): args["task_type"] = "Communication" args["task_type_parameters"] = '[]' digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % name) args["managers"] += [ Manager("manager", digest)] for lang in LANGUAGES: stub_name = os.path.join( task_path, "sol", "stub.%s" % lang) if os.path.exists(stub_name): digest = self.file_cacher.put_file_from_path( stub_name, "Stub for task %s and language %s" % (name, lang)) args["managers"] += [ Manager("stub.%s" % lang, digest)] else: logger.warning("Stub for language %s not " "found." % lang) break # Otherwise, the task type is Batch else: args["task_type"] = "Batch" args["task_type_parameters"] = \ '["%s", ["%s", "%s"], "%s"]' % \ (compilation_param, infile_param, outfile_param, evaluation_param) args["testcases"] = [] for i in xrange(n_input): input_digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "input", "input%d.txt" % i), "Input %d for task %s" % (i, name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "output", "output%d.txt" % i), "Output %d for task %s" % (i, name)) args["testcases"] += [ Testcase("%03d" % i, False, input_digest, output_digest)] if args["task_type"] == "OutputOnly": task.attachments += [ Attachment("input_%03d.txt" % i, input_digest)] public_testcases = load(conf, None, ["public_testcases", "risultati"], conv=lambda x: "" if x is None else x) if public_testcases != "": for x in public_testcases.split(","): args["testcases"][int(x.strip())].public = True dataset = Dataset(**args) task.active_dataset = dataset # Import was successful os.remove(os.path.join(task_path, ".import_error")) logger.info("Task parameters loaded.") return task
def get_task(self, get_statement=True): """See docstring in class TaskLoader.""" name = os.path.split(self.path)[1] if (not os.path.exists(os.path.join(self.path, "task.yaml"))) and \ (not os.path.exists(os.path.join(self.path, "..", name + ".yaml"))): logger.critical("File missing: \"task.yaml\"") return None # We first look for the yaml file inside the task folder, # and eventually fallback to a yaml file in its parent folder. try: conf = load_yaml_from_path(os.path.join(self.path, "task.yaml")) except OSError as err: try: deprecated_path = os.path.join(self.path, "..", name + ".yaml") conf = load_yaml_from_path(deprecated_path) logger.warning("You're using a deprecated location for the " "task.yaml file. You're advised to move %s to " "%s.", deprecated_path, os.path.join(self.path, "task.yaml")) except OSError: # Since both task.yaml and the (deprecated) "../taskname.yaml" # are missing, we will only warn the user that task.yaml is # missing (to avoid encouraging the use of the deprecated one) raise err # Here we update the time of the last import touch(os.path.join(self.path, ".itime")) # If this file is not deleted, then the import failed touch(os.path.join(self.path, ".import_error")) args = {} load(conf, args, ["name", "nome_breve"]) load(conf, args, ["title", "nome"]) if name != args["name"]: logger.info("The task name (%s) and the directory name (%s) are " "different. The former will be used.", args["name"], name) if args["name"] == args["title"]: logger.warning("Short name equals long name (title). " "Please check.") name = args["name"] logger.info("Loading parameters for task %s.", name) if get_statement: primary_language = load(conf, None, "primary_language") if primary_language is None: primary_language = 'it' paths = [os.path.join(self.path, "statement", "statement.pdf"), os.path.join(self.path, "testo", "testo.pdf")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Statement for task %s (lang: %s)" % (name, primary_language)) break else: logger.critical("Couldn't find any task statement, aborting.") sys.exit(1) args["statements"] = { primary_language: Statement(primary_language, digest) } args["primary_statements"] = [primary_language] args["submission_format"] = ["%s.%%l" % name] # Import the feedback level when explicitly set to full # (default behaviour is restricted) if conf.get("feedback_level", None) == FEEDBACK_LEVEL_FULL: args["feedback_level"] = FEEDBACK_LEVEL_FULL elif conf.get("feedback_level", None) == FEEDBACK_LEVEL_RESTRICTED: args["feedback_level"] = FEEDBACK_LEVEL_RESTRICTED if conf.get("score_mode", None) == SCORE_MODE_MAX: args["score_mode"] = SCORE_MODE_MAX elif conf.get("score_mode", None) == SCORE_MODE_MAX_SUBTASK: args["score_mode"] = SCORE_MODE_MAX_SUBTASK elif conf.get("score_mode", None) == SCORE_MODE_MAX_TOKENED_LAST: args["score_mode"] = SCORE_MODE_MAX_TOKENED_LAST # Use the new token settings format if detected. if "token_mode" in conf: load(conf, args, "token_mode") load(conf, args, "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_interval", conv=make_timedelta) load(conf, args, "token_gen_max") # Otherwise fall back on the old one. else: logger.warning( "task.yaml uses a deprecated format for token settings which " "will soon stop being supported, you're advised to update it.") # Determine the mode. if conf.get("token_initial", None) is None: args["token_mode"] = TOKEN_MODE_DISABLED elif conf.get("token_gen_number", 0) > 0 and \ conf.get("token_gen_time", 0) == 0: args["token_mode"] = TOKEN_MODE_INFINITE else: args["token_mode"] = TOKEN_MODE_FINITE # Set the old default values. args["token_gen_initial"] = 0 args["token_gen_number"] = 0 args["token_gen_interval"] = timedelta() # Copy the parameters to their new names. load(conf, args, "token_total", "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_initial", "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_time", "token_gen_interval", conv=make_timedelta) load(conf, args, "token_max", "token_gen_max") # Remove some corner cases. if args["token_gen_initial"] is None: args["token_gen_initial"] = 0 if args["token_gen_interval"].total_seconds() == 0: args["token_gen_interval"] = timedelta(minutes=1) load(conf, args, "max_submission_number") load(conf, args, "max_user_test_number") load(conf, args, "min_submission_interval", conv=make_timedelta) load(conf, args, "min_user_test_interval", conv=make_timedelta) # Attachments args["attachments"] = dict() if os.path.exists(os.path.join(self.path, "att")): for filename in os.listdir(os.path.join(self.path, "att")): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "att", filename), "Attachment %s for task %s" % (filename, name)) args["attachments"][filename] = Attachment(filename, digest) task = Task(**args) args = {} args["task"] = task args["description"] = conf.get("version", "Default") args["autojudge"] = False load(conf, args, ["time_limit", "timeout"], conv=float) # The Italian YAML format specifies memory limits in MiB. load(conf, args, ["memory_limit", "memlimit"], conv=lambda mb: mb * 1024 * 1024) # Builds the parameters that depend on the task type args["managers"] = [] infile_param = conf.get("infile", "input.txt") outfile_param = conf.get("outfile", "output.txt") # If there is sol/grader.%l for some language %l, then, # presuming that the task type is Batch, we retrieve graders # in the form sol/grader.%l graders = False for lang in LANGUAGES: if os.path.exists(os.path.join( self.path, "sol", "grader%s" % lang.source_extension)): graders = True break if graders: # Read grader for each language for lang in LANGUAGES: extension = lang.source_extension grader_filename = os.path.join( self.path, "sol", "grader%s" % extension) if os.path.exists(grader_filename): digest = self.file_cacher.put_file_from_path( grader_filename, "Grader for task %s and language %s" % (task.name, lang)) args["managers"] += [ Manager("grader%s" % extension, digest)] else: logger.warning("Grader for language %s not found ", lang) # Read managers with other known file extensions for other_filename in os.listdir(os.path.join(self.path, "sol")): if any(other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Manager %s for task %s" % (other_filename, task.name)) args["managers"] += [ Manager(other_filename, digest)] compilation_param = "grader" else: compilation_param = "alone" # If there is check/checker (or equivalent), then, presuming # that the task type is Batch or OutputOnly, we retrieve the # comparator paths = [os.path.join(self.path, "check", "checker"), os.path.join(self.path, "cor", "correttore")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [ Manager("checker", digest)] evaluation_param = "comparator" break else: evaluation_param = "diff" # Detect subtasks by checking GEN gen_filename = os.path.join(self.path, 'gen', 'GEN') try: with open(gen_filename, "rt", encoding="utf-8") as gen_file: subtasks = [] testcases = 0 points = None for line in gen_file: line = line.strip() splitted = line.split('#', 1) if len(splitted) == 1: # This line represents a testcase, otherwise # it's just a blank if splitted[0] != '': testcases += 1 else: testcase, comment = splitted testcase = testcase.strip() comment = comment.strip() testcase_detected = len(testcase) > 0 copy_testcase_detected = comment.startswith("COPY:") subtask_detected = comment.startswith('ST:') flags = [testcase_detected, copy_testcase_detected, subtask_detected] if len([x for x in flags if x]) > 1: raise Exception("No testcase and command in" " the same line allowed") # This line represents a testcase and contains a # comment, but the comment doesn't start a new # subtask if testcase_detected or copy_testcase_detected: testcases += 1 # This line starts a new subtask if subtask_detected: # Close the previous subtask if points is None: assert(testcases == 0) else: subtasks.append([points, testcases]) # Open the new one testcases = 0 points = int(comment[3:].strip()) # Close last subtask (if no subtasks were defined, just # fallback to Sum) if points is None: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = testcases if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = input_value else: subtasks.append([points, testcases]) assert(100 == sum([int(st[0]) for st in subtasks])) n_input = sum([int(st[1]) for st in subtasks]) args["score_type"] = "GroupMin" args["score_type_parameters"] = subtasks if "n_input" in conf: assert int(conf['n_input']) == n_input # If gen/GEN doesn't exist, just fallback to Sum except OSError: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = int(conf['n_input']) if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = input_value # Override score_type if explicitly specified if "score_type" in conf and "score_type_parameters" in conf: logger.info("Overriding 'score_type' and 'score_type_parameters' " "as per task.yaml") load(conf, args, "score_type") load(conf, args, "score_type_parameters") elif "score_type" in conf or "score_type_parameters" in conf: logger.warning("To override score type data, task.yaml must " "specify both 'score_type' and " "'score_type_parameters'.") # If output_only is set, then the task type is OutputOnly if conf.get('output_only', False): args["task_type"] = "OutputOnly" args["time_limit"] = None args["memory_limit"] = None args["task_type_parameters"] = [evaluation_param] task.submission_format = \ ["output_%03d.txt" % i for i in range(n_input)] # If there is check/manager (or equivalent), then the task # type is Communication else: paths = [os.path.join(self.path, "check", "manager"), os.path.join(self.path, "cor", "manager")] for path in paths: if os.path.exists(path): num_processes = load(conf, None, "num_processes") if num_processes is None: num_processes = 1 logger.info("Task type Communication") args["task_type"] = "Communication" args["task_type_parameters"] = \ [num_processes, "stub", "fifo_io"] digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [ Manager("manager", digest)] for lang in LANGUAGES: stub_name = os.path.join( self.path, "sol", "stub%s" % lang.source_extension) if os.path.exists(stub_name): digest = self.file_cacher.put_file_from_path( stub_name, "Stub for task %s and language %s" % ( task.name, lang.name)) args["managers"] += [ Manager( "stub%s" % lang.source_extension, digest)] else: logger.warning("Stub for language %s not " "found.", lang.name) for other_filename in os.listdir(os.path.join(self.path, "sol")): if any(other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Stub %s for task %s" % (other_filename, task.name)) args["managers"] += [ Manager(other_filename, digest)] break # Otherwise, the task type is Batch else: args["task_type"] = "Batch" args["task_type_parameters"] = \ [compilation_param, [infile_param, outfile_param], evaluation_param] args["testcases"] = [] for i in range(n_input): input_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "input", "input%d.txt" % i), "Input %d for task %s" % (i, task.name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "output", "output%d.txt" % i), "Output %d for task %s" % (i, task.name)) args["testcases"] += [ Testcase("%03d" % i, False, input_digest, output_digest)] if args["task_type"] == "OutputOnly": task.attachments.set( Attachment("input_%03d.txt" % i, input_digest)) public_testcases = load(conf, None, ["public_testcases", "risultati"], conv=lambda x: "" if x is None else x) if public_testcases == "all": for t in args["testcases"]: t.public = True elif len(public_testcases) > 0: for x in public_testcases.split(","): args["testcases"][int(x.strip())].public = True args["testcases"] = dict((tc.codename, tc) for tc in args["testcases"]) args["managers"] = dict((mg.filename, mg) for mg in args["managers"]) dataset = Dataset(**args) task.active_dataset = dataset # Import was successful os.remove(os.path.join(self.path, ".import_error")) logger.info("Task parameters loaded.") return task
def get_task(self, get_statement=True): """See docstring in class TaskLoader.""" name = os.path.split(self.path)[1] if (not os.path.exists(os.path.join(self.path, "task.yaml"))) and \ (not os.path.exists(os.path.join(self.path, "..", name + ".yaml"))): logger.critical("File missing: \"task.yaml\"") return None # We first look for the yaml file inside the task folder, # and eventually fallback to a yaml file in its parent folder. try: conf = yaml.safe_load( io.open(os.path.join(self.path, "task.yaml"), "rt", encoding="utf-8")) except IOError as err: try: deprecated_path = os.path.join(self.path, "..", name + ".yaml") conf = yaml.safe_load(io.open(deprecated_path, "rt", encoding="utf-8")) logger.warning("You're using a deprecated location for the " "task.yaml file. You're advised to move %s to " "%s.", deprecated_path, os.path.join(self.path, "task.yaml")) except IOError: # Since both task.yaml and the (deprecated) "../taskname.yaml" # are missing, we will only warn the user that task.yaml is # missing (to avoid encouraging the use of the deprecated one) raise err # Here we update the time of the last import touch(os.path.join(self.path, ".itime")) # If this file is not deleted, then the import failed touch(os.path.join(self.path, ".import_error")) args = {} load(conf, args, ["name", "nome_breve"]) load(conf, args, ["title", "nome"]) if name != args["name"]: logger.info("The task name (%s) and the directory name (%s) are " "different. The former will be used.", args["name"], name) if args["name"] == args["title"]: logger.warning("Short name equals long name (title). " "Please check.") name = args["name"] logger.info("Loading parameters for task %s.", name) if get_statement: primary_language = load(conf, None, "primary_language") if primary_language is None: primary_language = 'it' paths = [os.path.join(self.path, "statement", "statement.pdf"), os.path.join(self.path, "testo", "testo.pdf")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Statement for task %s (lang: %s)" % (name, primary_language)) break else: logger.critical("Couldn't find any task statement, aborting.") sys.exit(1) args["statements"] = { primary_language: Statement(primary_language, digest) } args["primary_statements"] = [primary_language] args["submission_format"] = ["%s.%%l" % name] if conf.get("score_mode", None) == SCORE_MODE_MAX: args["score_mode"] = SCORE_MODE_MAX elif conf.get("score_mode", None) == SCORE_MODE_MAX_TOKENED_LAST: args["score_mode"] = SCORE_MODE_MAX_TOKENED_LAST # Use the new token settings format if detected. if "token_mode" in conf: load(conf, args, "token_mode") load(conf, args, "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_interval", conv=make_timedelta) load(conf, args, "token_gen_max") # Otherwise fall back on the old one. else: logger.warning( "task.yaml uses a deprecated format for token settings which " "will soon stop being supported, you're advised to update it.") # Determine the mode. if conf.get("token_initial", None) is None: args["token_mode"] = TOKEN_MODE_DISABLED elif conf.get("token_gen_number", 0) > 0 and \ conf.get("token_gen_time", 0) == 0: args["token_mode"] = TOKEN_MODE_INFINITE else: args["token_mode"] = TOKEN_MODE_FINITE # Set the old default values. args["token_gen_initial"] = 0 args["token_gen_number"] = 0 args["token_gen_interval"] = timedelta() # Copy the parameters to their new names. load(conf, args, "token_total", "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_initial", "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_time", "token_gen_interval", conv=make_timedelta) load(conf, args, "token_max", "token_gen_max") # Remove some corner cases. if args["token_gen_initial"] is None: args["token_gen_initial"] = 0 if args["token_gen_interval"].total_seconds() == 0: args["token_gen_interval"] = timedelta(minutes=1) load(conf, args, "max_submission_number") load(conf, args, "max_user_test_number") load(conf, args, "min_submission_interval", conv=make_timedelta) load(conf, args, "min_user_test_interval", conv=make_timedelta) # Attachments args["attachments"] = dict() if os.path.exists(os.path.join(self.path, "att")): for filename in os.listdir(os.path.join(self.path, "att")): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "att", filename), "Attachment %s for task %s" % (filename, name)) args["attachments"][filename] = Attachment(filename, digest) task = Task(**args) args = {} args["task"] = task args["description"] = conf.get("version", "Default") args["autojudge"] = False load(conf, args, ["time_limit", "timeout"], conv=float) load(conf, args, ["memory_limit", "memlimit"]) # Builds the parameters that depend on the task type args["managers"] = [] infile_param = conf.get("infile", "input.txt") outfile_param = conf.get("outfile", "output.txt") # If there is sol/grader.%l for some language %l, then, # presuming that the task type is Batch, we retrieve graders # in the form sol/grader.%l graders = False for lang in LANGUAGES: if os.path.exists(os.path.join( self.path, "sol", "grader%s" % lang.source_extension)): graders = True break if graders: # Read grader for each language for lang in LANGUAGES: extension = lang.source_extension grader_filename = os.path.join( self.path, "sol", "grader%s" % extension) if os.path.exists(grader_filename): digest = self.file_cacher.put_file_from_path( grader_filename, "Grader for task %s and language %s" % (task.name, lang)) args["managers"] += [ Manager("grader%s" % extension, digest)] else: logger.warning("Grader for language %s not found ", lang) # Read managers with other known file extensions for other_filename in os.listdir(os.path.join(self.path, "sol")): if any(other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Manager %s for task %s" % (other_filename, task.name)) args["managers"] += [ Manager(other_filename, digest)] compilation_param = "grader" else: compilation_param = "alone" # If there is check/checker (or equivalent), then, presuming # that the task type is Batch or OutputOnly, we retrieve the # comparator paths = [os.path.join(self.path, "check", "checker"), os.path.join(self.path, "cor", "correttore")] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [ Manager("checker", digest)] evaluation_param = "comparator" break else: evaluation_param = "diff" # Detect subtasks by checking GEN gen_filename = os.path.join(self.path, 'gen', 'GEN') try: with io.open(gen_filename, "rt", encoding="utf-8") as gen_file: subtasks = [] testcases = 0 points = None for line in gen_file: line = line.strip() splitted = line.split('#', 1) if len(splitted) == 1: # This line represents a testcase, otherwise # it's just a blank if splitted[0] != '': testcases += 1 else: testcase, comment = splitted testcase = testcase.strip() comment = comment.strip() testcase_detected = len(testcase) > 0 copy_testcase_detected = comment.startswith("COPY:") subtask_detected = comment.startswith('ST:') flags = [testcase_detected, copy_testcase_detected, subtask_detected] if len([x for x in flags if x]) > 1: raise Exception("No testcase and command in" " the same line allowed") # This line represents a testcase and contains a # comment, but the comment doesn't start a new # subtask if testcase_detected or copy_testcase_detected: testcases += 1 # This line starts a new subtask if subtask_detected: # Close the previous subtask if points is None: assert(testcases == 0) else: subtasks.append([points, testcases]) # Open the new one testcases = 0 points = int(comment[3:].strip()) # Close last subtask (if no subtasks were defined, just # fallback to Sum) if points is None: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = testcases if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = input_value else: subtasks.append([points, testcases]) assert(100 == sum([int(st[0]) for st in subtasks])) n_input = sum([int(st[1]) for st in subtasks]) args["score_type"] = "GroupMin" args["score_type_parameters"] = subtasks if "n_input" in conf: assert int(conf['n_input']) == n_input # If gen/GEN doesn't exist, just fallback to Sum except IOError: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = int(conf['n_input']) if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = input_value # Override score_type if explicitly specified if "score_type" in conf and "score_type_parameters" in conf: logger.info("Overriding 'score_type' and 'score_type_parameters' " "as per task.yaml") load(conf, args, "score_type") load(conf, args, "score_type_parameters") elif "score_type" in conf or "score_type_parameters" in conf: logger.warning("To override score type data, task.yaml must " "specify both 'score_type' and " "'score_type_parameters'.") # If output_only is set, then the task type is OutputOnly if conf.get('output_only', False): args["task_type"] = "OutputOnly" args["time_limit"] = None args["memory_limit"] = None args["task_type_parameters"] = [evaluation_param] task.submission_format = \ ["output_%03d.txt" % i for i in range(n_input)] # If there is check/manager (or equivalent), then the task # type is Communication else: paths = [os.path.join(self.path, "check", "manager"), os.path.join(self.path, "cor", "manager")] for path in paths: if os.path.exists(path): num_processes = load(conf, None, "num_processes") if num_processes is None: num_processes = 1 logger.info("Task type Communication") args["task_type"] = "Communication" args["task_type_parameters"] = [num_processes] digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [ Manager("manager", digest)] for lang in LANGUAGES: stub_name = os.path.join( self.path, "sol", "stub%s" % lang.source_extension) if os.path.exists(stub_name): digest = self.file_cacher.put_file_from_path( stub_name, "Stub for task %s and language %s" % ( task.name, lang.name)) args["managers"] += [ Manager( "stub%s" % lang.source_extension, digest)] else: logger.warning("Stub for language %s not " "found.", lang.name) for other_filename in os.listdir(os.path.join(self.path, "sol")): if any(other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Stub %s for task %s" % (other_filename, task.name)) args["managers"] += [ Manager(other_filename, digest)] break # Otherwise, the task type is Batch else: args["task_type"] = "Batch" args["task_type_parameters"] = \ [compilation_param, [infile_param, outfile_param], evaluation_param] args["testcases"] = [] for i in range(n_input): input_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "input", "input%d.txt" % i), "Input %d for task %s" % (i, task.name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "output", "output%d.txt" % i), "Output %d for task %s" % (i, task.name)) args["testcases"] += [ Testcase("%03d" % i, False, input_digest, output_digest)] if args["task_type"] == "OutputOnly": task.attachments.set( Attachment("input_%03d.txt" % i, input_digest)) public_testcases = load(conf, None, ["public_testcases", "risultati"], conv=lambda x: "" if x is None else x) if public_testcases == "all": for t in args["testcases"]: t.public = True elif len(public_testcases) > 0: for x in public_testcases.split(","): args["testcases"][int(x.strip())].public = True args["testcases"] = dict((tc.codename, tc) for tc in args["testcases"]) args["managers"] = dict((mg.filename, mg) for mg in args["managers"]) dataset = Dataset(**args) task.active_dataset = dataset # Import was successful os.remove(os.path.join(self.path, ".import_error")) logger.info("Task parameters loaded.") return task
def get_task(self, get_statement=True): """See docstring in class TaskLoader.""" name = os.path.split(self.path)[1] if (not os.path.exists(os.path.join(self.path, "task.yaml"))) and \ (not os.path.exists(os.path.join(self.path, "problema.yaml"))) and \ (not os.path.exists(os.path.join(self.path, "..", name + ".yaml"))): logger.critical("File missing: \"task.yaml\"") return None # We first look for the yaml file inside the task folder, # and eventually fallback to a yaml file in its parent folder. try: conf = yaml.safe_load( io.open(os.path.join(self.path, "task.yaml"), "rt", encoding="utf-8")) except IOError as err: try: conf = yaml.safe_load( io.open(os.path.join(self.path, "problema.yaml"), "rt", encoding="utf-8")) except: try: deprecated_path = os.path.join(self.path, "..", name + ".yaml") conf = yaml.safe_load( io.open(deprecated_path, "rt", encoding="utf-8")) logger.warning( "You're using a deprecated location for the " "task.yaml file. You're advised to move %s to " "%s.", deprecated_path, os.path.join(self.path, "task.yaml")) except IOError: # Since both task.yaml and the (deprecated) "../taskname.yaml" # are missing, we will only warn the user that task.yaml is # missing (to avoid encouraging the use of the deprecated one) raise err # Here we update the time of the last import touch(os.path.join(self.path, ".itime")) # If this file is not deleted, then the import failed touch(os.path.join(self.path, ".import_error")) args = {} load(conf, args, ["name", "nome_breve"]) load(conf, args, ["title", "nome"]) load(conf, args, "hide_task_prefix") load(conf, args, "category") load(conf, args, "level") if "level" in args: args["level"] = unicode(args["level"]) if name != args["name"]: logger.info( "The task name (%s) and the directory name (%s) are " "different. The former will be used.", args["name"], name) if args["name"] == args["title"]: logger.warning("Short name equals long name (title). " "Please check.") name = args["name"] logger.info("Loading parameters for task %s.", name) if get_statement: primary_language = load(conf, None, "primary_language") if primary_language is None: primary_language = 'it' paths = [ os.path.join(self.path, "statement", "statement.pdf"), os.path.join(self.path, "statement.pdf"), os.path.join(self.path, "enunciado.pdf"), os.path.join(self.path, args["name"] + ".pdf"), os.path.join(self.path, "testo", "testo.pdf") ] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Statement for task %s (lang: %s)" % (name, primary_language)) break else: logger.critical("Couldn't find any task statement, aborting.") sys.exit(1) args["statements"] = [Statement(primary_language, digest)] args["primary_statements"] = '["%s"]' % (primary_language) args["attachments"] = [] # FIXME Use auxiliary args["submission_format"] = [SubmissionFormatElement("%s.%%l" % name)] if conf.get("score_mode", None) == SCORE_MODE_MAX: args["score_mode"] = SCORE_MODE_MAX elif conf.get("score_mode", None) == SCORE_MODE_MAX_TOKENED_LAST: args["score_mode"] = SCORE_MODE_MAX_TOKENED_LAST # Use the new token settings format if detected. if "token_mode" in conf: load(conf, args, "token_mode") load(conf, args, "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_interval", conv=make_timedelta) load(conf, args, "token_gen_max") # Otherwise fall back on the old one. else: logger.warning( "task.yaml uses a deprecated format for token settings which " "will soon stop being supported, you're advised to update it.") # Determine the mode. if conf.get("token_initial", None) is None: args["token_mode"] = "disabled" elif conf.get("token_gen_number", 0) > 0 and \ conf.get("token_gen_time", 0) == 0: args["token_mode"] = "infinite" else: args["token_mode"] = "finite" # Set the old default values. args["token_gen_initial"] = 0 args["token_gen_number"] = 0 args["token_gen_interval"] = timedelta() # Copy the parameters to their new names. load(conf, args, "token_total", "token_max_number") load(conf, args, "token_min_interval", conv=make_timedelta) load(conf, args, "token_initial", "token_gen_initial") load(conf, args, "token_gen_number") load(conf, args, "token_gen_time", "token_gen_interval", conv=make_timedelta) load(conf, args, "token_max", "token_gen_max") # Remove some corner cases. if args["token_gen_initial"] is None: args["token_gen_initial"] = 0 if args["token_gen_interval"].total_seconds() == 0: args["token_gen_interval"] = timedelta(minutes=1) load(conf, args, "max_submission_number") load(conf, args, "max_user_test_number") load(conf, args, "min_submission_interval", conv=make_timedelta) load(conf, args, "min_user_test_interval", conv=make_timedelta) # Attachments args["attachments"] = [] if os.path.exists(os.path.join(self.path, "att")): for filename in os.listdir(os.path.join(self.path, "att")): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "att", filename), "Attachment %s for task %s" % (filename, name)) args["attachments"] += [Attachment(filename, digest)] task = Task(**args) args = {} args["task"] = task args["description"] = conf.get("version", "Default") args["autojudge"] = False load(conf, args, ["time_limit", "timeout"], conv=float) load(conf, args, ["memory_limit", "memlimit"]) # Builds the parameters that depend on the task type args["managers"] = [] infile_param = conf.get("infile", "input.txt") outfile_param = conf.get("outfile", "output.txt") # If there is sol/grader.%l for some language %l, then, # presuming that the task type is Batch, we retrieve graders # in the form sol/grader.%l graders = False for lang in LANGUAGES: if os.path.exists( os.path.join(self.path, "sol", "grader%s" % lang.source_extension)): graders = True break if graders: # Read grader for each language for lang in LANGUAGES: extension = lang.source_extension grader_filename = os.path.join(self.path, "sol", "grader%s" % extension) if os.path.exists(grader_filename): digest = self.file_cacher.put_file_from_path( grader_filename, "Grader for task %s and language %s" % (task.name, lang)) args["managers"] += [ Manager("grader%s" % extension, digest) ] else: logger.warning("Grader for language %s not found ", lang) # Read managers with other known file extensions for other_filename in os.listdir(os.path.join(self.path, "sol")): if any( other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Manager %s for task %s" % (other_filename, task.name)) args["managers"] += [Manager(other_filename, digest)] compilation_param = "grader" else: compilation_param = "alone" # If there is check/checker (or equivalent), then, presuming # that the task type is Batch or OutputOnly, we retrieve the # comparator paths = [ os.path.join(self.path, "check", "checker"), os.path.join(self.path, "corrector.exe"), os.path.join(self.path, "cor", "correttore") ] for path in paths: if os.path.exists(path): digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [Manager("checker", digest)] evaluation_param = "comparator" break else: evaluation_param = "diff" # Detect subtasks by checking GEN gen_filename = os.path.join(self.path, 'gen', 'GEN') try: with io.open(gen_filename, "rt", encoding="utf-8") as gen_file: subtasks = [] testcases = 0 points = None for line in gen_file: line = line.strip() splitted = line.split('#', 1) if len(splitted) == 1: # This line represents a testcase, otherwise # it's just a blank if splitted[0] != '': testcases += 1 else: testcase, comment = splitted testcase = testcase.strip() comment = comment.strip() testcase_detected = testcase != '' copy_testcase_detected = comment.startswith("COPY:") subtask_detected = comment.startswith('ST:') flags = [ testcase_detected, copy_testcase_detected, subtask_detected ] if len([x for x in flags if x]) > 1: raise Exception("No testcase and command in" " the same line allowed") # This line represents a testcase and contains a # comment, but the comment doesn't start a new # subtask if testcase_detected or copy_testcase_detected: testcases += 1 # This line starts a new subtask if subtask_detected: # Close the previous subtask if points is None: assert (testcases == 0) else: subtasks.append([points, testcases]) # Open the new one testcases = 0 points = int(comment[3:].strip()) # Close last subtask (if no subtasks were defined, just # fallback to Sum) if points is None: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 n_input = testcases if n_input != 0: input_value = total_value / n_input args["score_type_parameters"] = "%s" % input_value else: subtasks.append([points, testcases]) assert (100 == sum([int(st[0]) for st in subtasks])) n_input = sum([int(st[1]) for st in subtasks]) args["score_type"] = "GroupMin" args["score_type_parameters"] = "%s" % subtasks if "n_input" in conf: assert int(conf['n_input']) == n_input # If gen/GEN doesn't exist, just fallback to Sum except IOError: if 'n_input' not in conf: conf['n_input'] = 0 n_input = int(conf['n_input']) if "score_type" in conf: args["score_type"] = conf["score_type"] if "score_type_parameters" in conf: args["score_type_parameters"] = ( "%s" % conf["score_type_parameters"]) args["score_type_parameters"] = re.sub( r'u\'([^\']+)\'', '\"\g<1>\"', args["score_type_parameters"]) else: args["score_type"] = "Sum" total_value = float(conf.get("total_value", 100.0)) input_value = 0.0 def count_testcases(folder): c = 0 if os.path.isdir(folder): for filename in sorted(os.listdir(folder)): nombre, ext = os.path.splitext(filename) if ext == ".in": c += 1 return c casos = n_input + count_testcases( os.path.join(self.path, "casos")) + count_testcases( os.path.join(self.path, "casos", "generados")) if casos != 0: input_value = total_value / casos args["score_type_parameters"] = "%s" % input_value # If output_only is set, then the task type is OutputOnly if conf.get('output_only', False): args["task_type"] = "OutputOnly" args["time_limit"] = None args["memory_limit"] = None args["task_type_parameters"] = '["%s"]' % evaluation_param task.submission_format = [ SubmissionFormatElement("output_%03d.txt" % i) for i in xrange(n_input) ] # If there is check/manager (or equivalent), then the task # type is Communication else: paths = [ os.path.join(self.path, "check", "manager"), os.path.join(self.path, "cor", "manager") ] for path in paths: if os.path.exists(path): num_processes = load(conf, None, "num_processes") if num_processes is None: num_processes = 1 logger.info("Task type Communication") args["task_type"] = "Communication" args["task_type_parameters"] = '[%d]' % num_processes digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % task.name) args["managers"] += [Manager("manager", digest)] for lang in LANGUAGES: stub_name = os.path.join( self.path, "sol", "stub%s" % lang.source_extension) if os.path.exists(stub_name): digest = self.file_cacher.put_file_from_path( stub_name, "Stub for task %s and language %s" % (task.name, lang.name)) args["managers"] += [ Manager("stub%s" % lang.source_extension, digest) ] else: logger.warning( "Stub for language %s not " "found.", lang.name) for other_filename in os.listdir( os.path.join(self.path, "sol")): if any( other_filename.endswith(header) for header in HEADER_EXTS): digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "sol", other_filename), "Stub %s for task %s" % (other_filename, task.name)) args["managers"] += [ Manager(other_filename, digest) ] break # Otherwise, the task type is Batch else: args["task_type"] = "Batch" args["task_type_parameters"] = \ '["%s", ["%s", "%s"], "%s"]' % \ (compilation_param, infile_param, outfile_param, evaluation_param) args["testcases"] = [] for i in xrange(n_input): input_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "input", "input%d.txt" % i), "Input %d for task %s" % (i, task.name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(self.path, "output", "output%d.txt" % i), "Output %d for task %s" % (i, task.name)) args["testcases"] += [ Testcase("%03d" % i, False, input_digest, output_digest) ] if args["task_type"] == "OutputOnly": task.attachments += [ Attachment("input_%03d.txt" % i, input_digest) ] def add_testcases_dir(folder): if os.path.isdir(folder): for filename in sorted(os.listdir(folder)): nombre, ext = os.path.splitext(filename) if ext == ".in": input_digest = self.file_cacher.put_file_from_path( os.path.join(folder, filename), "Input %s for task %s" % (nombre, task.name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(folder, nombre + ".dat"), "Output %s for task %s" % (nombre, task.name)) args["testcases"] += [ Testcase(nombre, False, input_digest, output_digest) ] if args["task_type"] == "OutputOnly": task.attachments += [ Attachment(filename, input_digest) ] add_testcases_dir(os.path.join(self.path, "casos")) add_testcases_dir(os.path.join(self.path, "casos", "generados")) public_testcases = load(conf, None, ["public_testcases", "risultati"], conv=lambda x: "" if x is None else x) if public_testcases == "all": for t in args["testcases"]: t.public = True elif public_testcases != "": for x in public_testcases.split(","): args["testcases"][int(x.strip())].public = True dataset = Dataset(**args) task.active_dataset = dataset # Import was successful os.remove(os.path.join(self.path, ".import_error")) logger.info("Task parameters loaded.") return task
def get_task(self, name): """See docstring in class Loader. """ try: num = self.tasks_order[name] # Here we expose an undocumented behavior, so that cmsMake can # import a task even without the whole contest; this is not to # be relied upon in general except AttributeError: num = 1 logger.info("Load task %s" % name) task_path = os.path.join(self.path, name) conf = {} try: conf = yaml.safe_load( io.open(os.path.join(task_path, "task.yaml"), "rt", encoding="utf-8")) except IOError: if os.path.exists(os.path.join(task_path, name + ".yaml")): conf = yaml.safe_load( io.open(os.path.join(task_path, name + ".yaml"), "rt", encoding="utf-8")) args = {} args["num"] = num args["name"] = name args["title"] = name.title() primary_language = conf.get("task", {}).get("primary_language", "en") for path in os.listdir(os.path.join(task_path, "statement")): digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "statement", path), "Statement for task %s (lang: %s)" % (name, primary_language)) break else: logger.critical("Couldn't find any task statement, aborting...") sys.exit(1) args["statements"] = [Statement(primary_language, digest)] args["primary_statements"] = '["%s"]' % (primary_language) args["submission_format"] = [ SubmissionFormatElement("%s.%%l" % name)] args["token_mode"] = "disabled" args.update(self.token_mode) # Load attachments args["attachments"] = [] if os.path.exists(os.path.join(task_path, "attachments")): for filename in os.listdir(os.path.join(task_path, "attachments")): digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "attachments", filename), "Attachment %s for task %s" % (filename, name)) args["attachments"] += [Attachment(filename, digest)] args.update(conf.get("task", {})) task = Task(**args) args = {} args["task"] = task args["description"] = "Default" args["autojudge"] = False args["time_limit"] = 2.0 args["memory_limit"] = 256 args["task_type"] = "Batch" args["score_type"] = "Sum" input_file = "" output_file = "" args["managers"] = [] # Overwrite parameters for key, param in conf.iteritems(): if key == "input": input_file = param elif key == "output": output_file = param elif key == "time_limit": args[key] = float(param) elif key in self.timedelta_params: args[key] = timedelta(seconds=param) elif key != "subtasks_parameters" and key != "subtasks" and key != "task": args[key] = param # Intelligent tests format detector # Load all tests recursively def load_tests(tests_path, name): if os.path.isfile(os.path.join(tests_path, name)): return [name] elif os.path.isdir(os.path.join(tests_path, name)): l = [] for path in os.listdir(os.path.join(tests_path, name)): l += load_tests(tests_path, os.path.join(name, path)) return l else: return [] full_names = load_tests(os.path.join(task_path, "tests"), "") tests_dict = dict((os.path.split(test)[-1], test) for test in full_names) tests = [] detected = False if not detected: # * / *.a format detected = True for test in tests_dict.keys(): if test.endswith(".a"): if test[:-2] not in tests_dict.keys(): detected = False else: if test + ".a" not in tests_dict.keys(): detected = False if detected: logger.info("Tests format * / *.a detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if not short_name.endswith(".a"): tests.append({"idx": idx, "input": test, "output": tests_dict[short_name + ".a"], "public": False }) idx += 1 if not detected: # *.in* / *.out* format detected = True for test in tests_dict.keys(): if test.find(".in") != -1: if test.replace(".in", ".out") not in tests_dict.keys(): detected = False elif test.find(".out") != -1: if test.replace(".out", ".in") not in tests_dict.keys(): detected = False else: detected = False if detected: logger.info("Tests format *.in* / *.out* detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if short_name.find(".in") != -1: tests.append({"idx": idx, "input": test, "output": tests_dict[short_name.replace(".in", ".out")], "public": False }) idx += 1 if not detected: # *.in* / *.sol* format detected = True for test in tests_dict.keys(): if test.find(".in") != -1: if test.replace(".in", ".sol") not in tests_dict.keys(): detected = False elif test.find(".sol") != -1: if test.replace(".sol", ".in") not in tests_dict.keys(): detected = False else: detected = False if detected: logger.info("Tests format *.in* / *.sol* detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if short_name.find(".in") != -1: tests.append({"idx": idx, "input": test, "output": tests_dict[short_name.replace(".in", ".sol")], "public": False }) idx += 1 if not detected: # *.in* / *.res* format detected = True for test in tests_dict.keys(): if test.find(".in") != -1: if test.replace(".in", ".res") not in tests_dict.keys(): detected = False elif test.find(".res") != -1: if test.replace(".res", ".in") not in tests_dict.keys(): detected = False else: detected = False if detected: logger.info("Tests format *.in* / *.res* detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if short_name.find(".in") != -1: tests.append({"idx": idx, "input": test, "output": tests_dict[short_name.replace(".in", ".res")], "public": False }) idx += 1 if not detected: # *.in* / *.ans* format detected = True for test in tests_dict.keys(): if test.find(".in") != -1: if test.replace(".in", ".ans") not in tests_dict.keys(): detected = False elif test.find(".ans") != -1: if test.replace(".ans", ".in") not in tests_dict.keys(): detected = False else: detected = False if detected: logger.info("Tests format *.in* / *.ans* detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if short_name.find(".in") != -1: tests.append({"idx": idx, "input": test, "output": tests_dict[short_name.replace(".in", ".ans")], "public": False }) idx += 1 if not detected: # *input* / *output* format detected = True for test in tests_dict.keys(): if test.find("input") != -1: if test.replace("input", "output") not in tests_dict.keys(): detected = False elif test.find("output") != -1: if test.replace("output", "input") not in tests_dict.keys(): detected = False else: detected = False if detected: logger.info("Tests format *input* / *output* detected") idx = 0 for (short_name, test) in sorted(tests_dict.items()): if short_name.find("input") != -1: tests.append({"idx": idx, "input": test, "output": tests_dict[short_name.replace("input", "output")], "public": False }) idx += 1 if not detected: # in* out* format using full paths detected = True for test in full_names: if test.startswith("in"): if "out" + test[2:] not in full_names: detected = False elif test.startswith("out"): if "in" + test[3:] not in full_names: detected = False else: detected = False if detected: logger.info("Tests format in* / out* with full paths detected") idx = 0 for test in sorted(full_names): if test.startswith("in"): tests.append({"idx": idx, "input": test, "output": "out" + test[2:], "public": False }) idx += 1 if not detected: # Need more intelligence logger.critical("Sorry, I can't recognize tests format") sys.exit(1) # Detect subtasks if "subtasks_parameters" in conf: logger.info("Detected simple subtask description") args["score_type"] = "NamedGroup" subtasks = conf["subtasks_parameters"] total_value = float(subtasks.get("total_value", 100)) is_public = subtasks.get("public_tests", False) if is_public: for test in tests: test["public"] = True samples = list(int(test.strip()) - 1 for test in subtasks.get("sample_tests", "").strip().split(",")) for i in samples: tests[i]["public"] = True samples_group = { "score": 0, "type": "sum", "public": rebuild_list(samples), "private": [], "hidden": [] } tests_group = { "score": total_value, "type": "sum", "public": [], "private": [], "hidden": [] } for i in xrange(len(tests)): if not i in samples: if is_public: tests_group["public"].append(i) else: tests_group["private"].append(i) tests_group["public"] = rebuild_list(tests_group["public"]) tests_group["private"] = rebuild_list(tests_group["private"]) if len(samples) == 0: args["score_type_parameters"] = json.dumps([tests_group]) else: args["score_type_parameters"] = json.dumps([samples_group, tests_group]) elif "subtasks" in conf: logger.info("Detected full subtask description") args["score_type"] = "NamedGroup" subtasks = conf.get("subtasks") for subtask in subtasks: if not "score" in subtask: subtask["score"] = 0 if not "type" in subtask: subtask["type"] = "sum" if subtask["type"] != "sum" and subtask["type"] != "min": # Custom evaluator parameter with open(os.path.join(task_path, subtask["type"]), "r") as prog_file: prog = prog_file.read() subtask["type"] = prog subtask["public"] = rebuild_list(subtask.get("public", []), test_list = tests, delta = 1) subtask["private"] = rebuild_list(subtask.get("private", []), test_list = tests, delta = 1) subtask["hidden"] = rebuild_list(subtask.get("hidden", []), test_list = tests, delta = 1) for i in subtask["public"]: tests[int(i)]["public"] = True args["score_type_parameters"] = json.dumps(conf.get("subtasks")) else: logger.info("Subtask description was not detected") args["score_type"] = "NamedGroup" # Autodetect samples samples = [] for test in tests: if test["input"].find("dummy") != -1 or test["input"].find("sample") != -1: samples.append(test["idx"]) for i in samples: tests[i]["public"] = True samples_group = { "score": 0, "type": "sum", "public": rebuild_list(samples), "private": [], "hidden": [] } tests_group = { "score": 100, "type": "sum", "public": [], "private": [], "hidden": [] } for i in xrange(len(tests)): if not i in samples: tests_group["private"].append(i) tests_group["public"] = rebuild_list(tests_group["public"]) tests_group["private"] = rebuild_list(tests_group["private"]) if len(samples) == 0: args["score_type_parameters"] = json.dumps([tests_group]) else: args["score_type_parameters"] = json.dumps([samples_group, tests_group]) # Load testcases args["testcases"] = [] for test in tests: i = test["idx"] input_digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "tests", test["input"]), "Input %d for task %s" % (i, name)) output_digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "tests", test["output"]), "Output %d for task %s" % (i, name)) args["testcases"] += [ Testcase("%03d" % i, test["public"], input_digest, output_digest)] # Load graders (and stubs if any) if os.path.isdir(os.path.join(task_path, "graders")): for filename in os.listdir(os.path.join(task_path, "graders")): digest = self.file_cacher.put_file_from_path( os.path.join(task_path, "graders", filename), "Grader %s for task %s" % (filename, name)) args["managers"] += [ Manager(filename, digest)] compilation_param = "grader" else: compilation_param = "alone" # Load checker paths = [os.path.join(task_path, "checker"), os.path.join(task_path, "check"), os.path.join(task_path, "check.exe")] for path in paths: if os.path.isfile(path): digest = self.file_cacher.put_file_from_path( path, "Checker for task %s" % name) args["managers"] += [ Manager("checker", digest)] evaluation_param = "comparator" break else: evaluation_param = "diff" # If the task type is Communication, try to load manager path = os.path.join(task_path, "manager") if os.path.isfile(path): args["task_type"] = "Communication" args["task_type_parameters"] = '[]' digest = self.file_cacher.put_file_from_path( path, "Manager for task %s" % name) args["managers"] += [ Manager("manager", digest)] # Set task type parameters if args["task_type"] == "OutputOnly": args["time_limit"] = None args["memory_limit"] = None args["task_type_parameters"] = '["%s"]' % evaluation_param task.submission_format = [ SubmissionFormatElement("%03d.out" % (i + 1)) for i in xrange(len(tests))] elif args["task_type"] == "Batch": args["task_type_parameters"] = \ '["%s", ["%s", "%s"], "%s"]' % \ (compilation_param, input_file, output_file, evaluation_param) logger.info("Task type is %s" % args["task_type"]) dataset = Dataset(**args) task.active_dataset = dataset logger.info("Task parameters loaded.") return task