""" vqaEval.evaluate(quesIds=quest_ids) # print accuracies print "\n" print "Overall Accuracy is: %.02f\n" % (vqaEval.accuracy['overall']) print "Per Question Type Accuracy is the following:" for quesType in vqaEval.accuracy['perQuestionType']: print "%s : %.02f" % (quesType, vqaEval.accuracy['perQuestionType'][quesType]) print "\n" print "Per Answer Type Accuracy is the following:" for ansType in vqaEval.accuracy['perAnswerType']: print "%s : %.02f" % (ansType, vqaEval.accuracy['perAnswerType'][ansType]) print "\n" if ans_type is None: return vqaEval.accuracy['overall'], vqaEval.accuracy['perAnswerType'] else: return vqaEval.accuracy['overall'], vqaEval.accuracy['perAnswerType'][ ans_type] if __name__ == '__main__': from config import TestConfig config = TestConfig() res_file = config.get_result_file('model-75001') evaluate_model('OpenEnded_mscoco_lstm_results.json')