Exemplo n.º 1
0
    def post_processing_function(examples, features, predictions,
                                 training_args):
        # Post-processing: we match the start logits and end logits to answers in the original context.
        predictions = postprocess_qa_predictions(
            examples=examples,
            features=features,
            predictions=predictions,
            max_answer_length=data_args.max_answer_length,
            output_dir=training_args.output_dir,
        )
        # Format the result to the format the metric expects.
        formatted_predictions = [{
            "id": k,
            "prediction_text": v
        } for k, v in predictions.items()]
        if training_args.do_predict:
            return formatted_predictions

        elif training_args.do_eval:
            references = [{
                "id": ex["id"],
                "answers": ex[answer_column_name]
            } for ex in datasets["validation"]]
            return EvalPrediction(predictions=formatted_predictions,
                                  label_ids=references)
Exemplo n.º 2
0
    def post_processing_function(examples,
                                 features,
                                 predictions,
                                 stage="eval"):
        # Post-processing: we match the start logits and end logits to answers in the original context.
        predictions = postprocess_qa_predictions(
            examples=examples,
            features=features,
            predictions=predictions,
            version_2_with_negative=args.version_2_with_negative,
            n_best_size=args.n_best_size,
            max_answer_length=args.max_answer_length,
            null_score_diff_threshold=args.null_score_diff_threshold,
            output_dir=args.output_dir,
            prefix=stage,
        )
        # Format the result to the format the metric expects.
        if args.version_2_with_negative:
            formatted_predictions = [{
                "id": k,
                "prediction_text": v,
                "no_answer_probability": 0.0
            } for k, v in predictions.items()]
        else:
            formatted_predictions = [{
                "id": k,
                "prediction_text": v
            } for k, v in predictions.items()]

        references = [{
            "id": ex["id"],
            "answers": ex[answer_column_name]
        } for ex in examples]
        return EvalPrediction(predictions=formatted_predictions,
                              label_ids=references)
Exemplo n.º 3
0
 def post_processing_function(examples, features, predictions):
     predictions = postprocess_qa_predictions(
         examples=examples,
         features=features,
         predictions=predictions,
         version_2_with_negative=data_args.version_2_with_negative,
         n_best_size=data_args.n_best_size,
         max_answer_length=data_args.max_answer_length,
         null_score_diff_threshold=data_args.null_score_diff_threshold,
         output_dir=training_args.output_dir,
         is_world_process_zero=trainer.is_world_process_zero(),
     )
     if data_args.version_2_with_negative:
         formatted_predictions = [{
             "id": k,
             "prediction_text": v,
             "no_answer_probability": 0.0
         } for k, v in predictions.items()]
     else:
         formatted_predictions = [{
             "id": k,
             "prediction_text": v
         } for k, v in predictions.items()]
     references = [{
         "id": ex["id"],
         "answers": ex[answer_column_name]
     } for ex in datasets["validation"]]
     return EvalPrediction(predictions=formatted_predictions,
                           label_ids=references)
Exemplo n.º 4
0
    def post_processing_function(examples,
                                 features,
                                 predictions,
                                 stage="eval"):
        # Post-processing: we match the start logits and end logits to answers in the original context.
        predictions = postprocess_qa_predictions(
            examples=examples,
            features=features,
            predictions=predictions,
            n_best_size=FLAGS.n_best_size,
            max_answer_length=FLAGS.max_answer_length,
            output_dir=args.output_dir,
            is_world_process_zero=trainer.is_world_process_zero(),
            prefix=stage,
            null_score_diff_threshold=float(FLAGS.diff_threshold),
            version_2_with_negative=True,
        )
        # Format the result to the format the metric expects.
        formatted_predictions = [{
            "id": k,
            "prediction_text": v
        } for k, v in predictions.items()]

        references = [{
            "id":
            ex['id'],
            'start_token':
            ex['annotations'][0]['long_answer']['start_token'],
            'end_token':
            ex['annotations'][0]['long_answer']['end_token']
        } for ex in examples]

        return EvalPrediction(predictions=formatted_predictions,
                              label_ids=references)
Exemplo n.º 5
0
def post_processing_function(features, predictions, text_data, data_args,
                             training_args):
    """
    post processing

    Args:
        features, predictions, text_data, data_args, training_args

    Returns:
        inference or evaluation results
    """
    predictions = postprocess_qa_predictions(
        examples=text_data["validation"],
        features=features,
        predictions=predictions,
        max_answer_length=data_args.max_answer_length,
        output_dir=training_args.output_dir,
    )

    formatted_predictions = [{
        "id": k,
        "prediction_text": v
    } for k, v in predictions.items()]
    if training_args.do_predict:
        return formatted_predictions

    elif training_args.do_eval:
        references = [{
            "id": ex["id"],
            "answers": ex["answers"].strip()
        } for ex in text_data["validation"]]
        return EvalPrediction(predictions=formatted_predictions,
                              label_ids=references)
Exemplo n.º 6
0
def post_processing_function(examples, features, predictions, text_data,
                             data_args, training_args):
    '''Model의 Prediction을 Text 형태로 변환하는 함수'''
    predictions = postprocess_qa_predictions(
        examples=examples,
        features=features,
        predictions=predictions,
        max_answer_length=data_args.max_answer_length,
        output_dir=training_args.output_dir,
    )

    formatted_predictions = [{
        "id": k,
        "prediction_text": last_processing(v)
    } for k, v in predictions.items()]
    if training_args.do_predict:
        return formatted_predictions

    references = [{
        "id": ex["id"],
        "answers": ex["answers"]
    } for ex in text_data["validation"]]
    return EvalPrediction(predictions=formatted_predictions,
                          label_ids=references)
Exemplo n.º 7
0
    def post_processing_function(examples,
                                 features,
                                 predictions,
                                 stage="eval"):
        # Post-processing: we match the start logits and end logits to
        # answers in the original context.

        if data_args.beam_search:
            predictions, scores_diff_json = \
                postprocess_qa_predictions_with_beam_search(
                    examples=examples,
                    features=features,
                    predictions=predictions,
                    version_2_with_negative=data_args.version_2_with_negative,
                    n_best_size=data_args.n_best_size,
                    max_answer_length=data_args.max_answer_length,
                    start_n_top=model.config.start_n_top,
                    end_n_top=model.config.end_n_top,
                    output_dir=training_args.output_dir,
                    # log_level=log_level,
                    prefix=stage,
                )

        else:
            predictions = postprocess_qa_predictions(
                examples=examples,
                features=features,
                predictions=predictions,
                version_2_with_negative=data_args.version_2_with_negative,
                n_best_size=data_args.n_best_size,
                max_answer_length=data_args.max_answer_length,
                output_dir=training_args.output_dir,
                prefix=stage,
            )

        if data_args.version_2_with_negative:
            if data_args.beam_search:
                formatted_predictions = [
                    {
                        "id": k,
                        "prediction_text": v,
                        "no_answer_probability": scores_diff_json[k]
                    }  # noqa E501
                    for k, v in predictions.items()
                ]
            else:
                formatted_predictions = [
                    {
                        "id": k,
                        "prediction_text": v,
                        "no_answer_probability": 0.0
                    } for k, v in predictions.items()  # noqa E501
                ]
        else:
            formatted_predictions = [{
                "id": k,
                "prediction_text": v
            } for k, v in predictions.items()]  # noqa E501

        references = [{
            "id": ex["id"],
            "answers": ex[answer_column_name]
        } for ex in examples]  # noqa E501
        return EvalPrediction(predictions=formatted_predictions,
                              label_ids=references)