예제 #1
0
def FUNSDBoxDetect_eval(config,
                        instance,
                        trainer,
                        metrics,
                        outDir=None,
                        startIndex=None,
                        lossFunc=None,
                        toEval=None):
    def __eval_metrics(data, target):
        acc_metrics = np.zeros((output.shape[0], len(metrics)))
        for ind in range(output.shape[0]):
            for i, metric in enumerate(metrics):
                acc_metrics[ind, i] += metric(output[ind:ind + 1],
                                              target[ind:ind + 1])
        return acc_metrics

    if toEval is None:
        toEval = ['bbs']
    THRESH = config['THRESH'] if 'THRESH' in config else 0.92
    axis_aligned_prediction = config['arch'] == 'OverSegBoxDetector'
    #print(type(instance['pixel_gt']))
    #if type(instance['pixel_gt']) == list:
    #    print(instance)
    #    print(startIndex)
    #data, targetBB, targetBBSizes = instance
    data = instance['img']
    batchSize = data.shape[0]
    targetBBs = instance['bb_gt']
    targetBBsSizes = instance['bb_sizes']
    targetPoints = instance['point_gt']
    targetPixels = instance['pixel_gt']
    imageName = instance['imgName']
    scale = instance['scale']
    target_num_neighbors = instance['num_neighbors']

    pretty = config['pretty'] if 'pretty' in config else False

    resultsDirName = 'results'
    #if outDir is not None and resultsDirName is not None:
    #rPath = os.path.join(outDir,resultsDirName)
    #if not os.path.exists(rPath):
    #    os.mkdir(rPath)
    #for name in targetBBs:
    #    nPath = os.path.join(rPath,name)
    #    if not os.path.exists(nPath):
    #        os.mkdir(nPath)

    #dataT = __to_tensor(data,gpu)
    #print('{}: {} x {}'.format(imageName,data.shape[2],data.shape[3]))
    #outputBBs, outputOffsets, outputLines, outputOffsetLines, outputPoints, outputPixels = model(dataT)
    losses, log, out = trainer.run(instance, get=toEval, val=True)
    #if outputPixels is not None:
    #    outputPixels = torch.sigmoid(outputPixels)
    index = 0
    loss = 0
    index = 0
    ttt_hit = True

    data = data.cpu().data.numpy()

    if 'bbs' in out:
        outputBBs = out['bbs']
        #threshConf = maxConf*THRESH
        if axis_aligned_prediction:  #overseg
            threshed_outputBBs = []
            for b in range(batchSize):
                maxConf = outputBBs[b, :, 0].max().item()
                threshConf = maxConf * THRESH  #max(maxConf*THRESH,0.5)
                print('b:{}, maxConf: {}, meanConf: {}, thresh: {}'.format(
                    b, maxConf, outputBBs[b, :, 0].mean().item(), threshConf))

                threshed_outputBBs.append(
                    outputBBs[b, outputBBs[b, :, 0] > threshConf].cpu())
                if 'SUP' in config:
                    threshed_outputBBs[0] = non_max_sup_overseg(
                        threshed_outputBBs[0])
            outputBBs = threshed_outputBBs

            #Uhh, I don't know if there is an effiecient method to prune out reduntant predictions, as you have to recompute things every time you prune one. For now, I won't worry about it
            #outputBBs = non_max_sup_keep_overlap_iou(outputBBs.cpu(),threshConf,0.4)

            #outputBBs = non_max_sup_iou(outputBBs.cpu(),threshConf,2,hard_limit=999000)
            #iou threshed, but also need to optimize to maintain overlap...
            #precompute all iou, and thus intersections
            #order by conf,#
            #   examine offending (over iou thresh) boxes, reverse conf (lowest up)
            #       does this the only bridge between me and another box?
        elif trainer.model.rotation:
            outputBBs = non_max_sup_dist(outputBBs.cpu(), threshConf, 3)
        else:
            maxConf = outputBBs[:, :, 0].max().item()
            threshConf = maxConf * THRESH
            outputBBs = non_max_sup_iou(outputBBs.cpu(), threshConf, 0.4)

    numClasses = trainer.model.numBBTypes

    if 'points' in out:
        outputPointsOld = outputPoints
        targetPointsOld = targetPoints
        outputPoints = {}
        targetPoints = {}
        i = 0
        for name, targ in targetPointsOld.items():
            if targ is not None:
                targetPoints[name] = targ.data.numpy()
            else:
                targetPoints[name] = None
            outputPoints[name] = outputPointsOld[i].cpu().data.numpy()
            i += 1
    if 'pixels' in out and outputPixels is not None:
        outputPixels = outputPixels.cpu().data.numpy()
    #metricsOut = __eval_metrics(outputBBs,targetBBs)
    #metricsOut = 0
    #if outDir is None:
    #    return metricsOut

    dists = defaultdict(list)
    dists_x = defaultdict(list)
    dists_y = defaultdict(list)
    scaleDiffs = defaultdict(list)
    rotDiffs = defaultdict(list)

    if 'bbs' in out:
        allPredNNs = []
        for b in range(batchSize):
            #outputBBs[b] = outputBBs[b].data.numpy()
            #print('image {} has {} {}'.format(startIndex+b,targetBBsSizes[name][b],name))
            #bbImage = np.ones_like(image)
            bbs = outputBBs[b].data.numpy()
            if bbs.shape[0] > 0:
                if trainer.model.predNumNeighbors:
                    predNN = bbs[:, 6]
                    allPredNNs += predNN.tolist()
                    predClass = bbs[:, 7:]
                else:
                    predClass = bbs[:, 6:]
            else:
                predNN = bbs  #i.e. a zero size tensor
                predClass = bbs

            if 'save_json' in config:
                assert (batchSize == 1)
                scale = scale[0]
                if targetBBs is not None:
                    if axis_aligned_prediction:
                        pass  #?
                    elif trainer.model.rotation:
                        targIndex, predWithNoIntersection = getTargIndexForPreds_dist(
                            targetBBs[b], torch.from_numpy(bbs), 1.1,
                            numClasses, extraPreds)
                    else:
                        targIndex, predWithNoIntersection = getTargIndexForPreds_iou(
                            targetBBs[b], torch.from_numpy(bbs), 0.4,
                            numClasses, extraPreds)
                    newId = targetBBs[b].size(0)
                else:
                    targIndex = -1 * torch.ones(bbs.shape[0])
                    newId = 1
                bbsData = []
                for j in range(bbs.shape[0]):
                    tl, tr, br, bl = getCorners(bbs[j, 1:])
                    id = targIndex[j].item()
                    if id < 0:
                        id = 'u{}'.format(newId)
                        newId += 1
                    else:
                        id = 'm{}'.format(id)
                    bb = {
                        'id':
                        id,
                        'poly_points':
                        [[float(tl[0] / scale),
                          float(tl[1] / scale)],
                         [float(tr[0] / scale),
                          float(tr[1] / scale)],
                         [float(br[0] / scale),
                          float(br[1] / scale)],
                         [float(bl[0] / scale),
                          float(bl[1] / scale)]],
                        'type':
                        'detectorPrediction',
                        'textPred':
                        float(predClass[j, 0]),
                        'fieldPred':
                        float(predClass[j, 1])
                    }
                    if numClasses == 2 and trainer.model.numBBTypes == 3:
                        bb['blankPred'] = float(predClass[j, 2])
                    if trainer.model.predNumNeighbors:
                        bb['nnPred'] = float(predNN[j])
                    bbsData.append(bb)

                if instance['pairs'] is None:
                    #import pdb; pdb.set_trace()
                    instance['pairs'] = []
                pairsData = [('m{}'.format(i1), 'm{}'.format(i2))
                             for i1, i2 in instance['pairs']]

                saveJSON = os.path.join(config['save_json'],
                                        imageName[b] + '.json')
                allData = {
                    'textBBs': bbsData,
                    'fieldBBs': [],
                    'pairs': pairsData
                }
                with open(saveJSON, 'w') as f:
                    json.dump(allData, f)
                    print('wrote {}'.format(saveJSON))

            if outDir is not None:
                #Write the results so we can train LF with them
                #saveFile = os.path.join(outDir,resultsDirName,name,'{}'.format(imageName[b]))
                #we must rescale the output to be according to the original image
                #rescaled_outputBBs_xyrs = outputBBs_xyrs[name][b]
                #rescaled_outputBBs_xyrs[:,1] /= scale[b]
                #rescaled_outputBBs_xyrs[:,2] /= scale[b]
                #rescaled_outputBBs_xyrs[:,4] /= scale[b]

                #np.save(saveFile,rescaled_outputBBs_xyrs)
                image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                                (1, 2, 0))) / 2.0)).copy()

                if image.shape[2] == 1:
                    image = img_f.gray2rgb(image)
                #if name=='text_start_gt':

                if not pretty:
                    for j in range(targetBBsSizes[b]):
                        plotRect(image, (1, 0.5, 0), targetBBs[b, j, 0:5])
                        #if trainer.model.predNumNeighbors:
                        #    x=int(targetBBs[b,j,0])
                        #    y=int(targetBBs[b,j,1]+targetBBs[b,j,3])
                        #    img_f.putText(image,'{:.2f}'.format(gtNumNeighbors[b,j]),(x,y), cv2.FONT_HERSHEY_SIMPLEX, 0.5,(0.6,0.3,0),2,cv2.LINE_AA)
                    #if alignmentBBs[b] is not None:
                    #    aj=alignmentBBs[b][j]
                    #    xc_gt = targetBBs[b,j,0]
                    #    yc_gt = targetBBs[b,j,1]
                    #    xc=outputBBs[b,aj,1]
                    #    yc=outputBBs[b,aj,2]
                    #    img_f.line(image,(xc,yc),(xc_gt,yc_gt),(0,1,0),1)
                    #    shade = 0.0+(outputBBs[b,aj,0]-threshConf)/(maxConf-threshConf)
                    #    shade = max(0,shade)
                    #    if outputBBs[b,aj,6] > outputBBs[b,aj,7]:
                    #        color=(0,shade,shade) #text
                    #    else:
                    #        color=(shade,shade,0) #field
                    #    plotRect(image,color,outputBBs[b,aj,1:6])

                #bbs=[]
                #pred_points=[]
                #maxConf = outputBBs[b,:,0].max()
                #threshConf = 0.5
                #threshConf = max(maxConf*0.9,0.5)
                #print("threshConf:{}".format(threshConf))
                #for j in range(outputBBs.shape[1]):
                #    conf = outputBBs[b,j,0]
                #    if conf>threshConf:
                #        bbs.append((conf,j))
                #    #pred_points.append(
                #bbs.sort(key=lambda a: a[0]) #so most confident bbs are draw last (on top)
                #import pdb; pdb.set_trace()
                for j in range(bbs.shape[0]):
                    #circle aligned predictions
                    conf = bbs[j, 0]
                    if outDir is not None:
                        shade = conf  #0.0+(conf-threshConf)/(maxConf-threshConf)
                        assert (shade >= 0 and shade <= 1)
                        #print(shade)
                        #if name=='text_start_gt' or name=='field_end_gt':
                        #    img_f.bb(bbImage[:,:,1],p1,p2,shade,2)
                        #if name=='text_end_gt':
                        #    img_f.bb(bbImage[:,:,2],p1,p2,shade,2)
                        #elif name=='field_end_gt' or name=='field_start_gt':
                        #    img_f.bb(bbImage[:,:,0],p1,p2,shade,2)
                        maxClassIndex = np.argmax(predClass[j])
                        if maxClassIndex == 0:  #header
                            color = (0, 0, shade)
                        elif maxClassIndex == 1:  #question
                            color = (0, shade, shade)
                        elif maxClassIndex == 2:  #answer
                            color = (shade, shade, 0)
                        elif maxClassIndex == 3:  #other
                            color = (shade, 0, shade)
                        else:
                            assert (False)

                        if pretty:
                            lineW = 2
                        else:
                            lineW = 1
                        if axis_aligned_prediction:
                            plotRectAndAngle(image, color, bbs[j, 1:6], lineW)
                        else:
                            plotRect(image, color, bbs[j, 1:6], lineW)
                        if trainer.model.predNumNeighbors and not pretty:
                            x = int(bbs[j, 1])
                            y = int(bbs[j, 2] - bbs[j, 4])
                            #color = int(min(abs(predNN[j]-target_num_neighbors[j]),2)*127)
                            #img_f.putText(image,'{}/{}'.format(predNN[j],target_num_neighbors[j]),(x,y), cv2.FONT_HERSHEY_SIMPLEX, 3,(color,0,0),2,cv2.LINE_AA)
                            img_f.putText(image, '{:.2f}'.format(predNN[j]),
                                          (x, y), cv2.FONT_HERSHEY_SIMPLEX,
                                          0.5, color, 2, cv2.LINE_AA)

                #for j in alignmentBBsTarg[name][b]:
                #    p1 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                #    p2 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                #    mid = ( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
                #    rad = round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
                #    #print(mid)
                #    #print(rad)
                #    img_f.circle(image,mid,rad,(1,0,1),1)

                #saveName = '{}_boxes_AP:{:.2f}'.format(imageName[b],aps_5all[b])
                saveName = '{}_boxes'.format(imageName[b])

                #for j in range(metricsOut.shape[1]):
                #    saveName+='_m:{0:.3f}'.format(metricsOut[i,j])
                saveName += '.png'
                img_f.imwrite(os.path.join(outDir, saveName), image)
                print('saved: ' + os.path.join(outDir, saveName))
            if 'points' in out:
                for name, out in outputPoints.items():
                    image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                                    (1, 2, 0))) / 2.0)).copy()
                    #if name=='text_start_gt':
                    for j in range(targetPointsSizes[name][b]):
                        p1 = (targetPoints[name][b, j,
                                                 0], targetPoints[name][b, j,
                                                                        1])
                        img_f.circle(image, p1, 2, (1, 0.5, 0), -1)
                    points = []
                    maxConf = max(out[b, :, 0].max(), 1.0)
                    threshConf = maxConf * 0.1
                    for j in range(out.shape[1]):
                        conf = out[b, j, 0]
                        if conf > threshConf:
                            p1 = (out[b, j, 1], out[b, j, 2])
                            points.append((conf, p1, j))
                    points.sort(key=lambda a: a[
                        0])  #so most confident bbs are draw last (on top)
                    for conf, p1, j in points:
                        shade = 0.0 + conf / maxConf
                        if name == 'table_points':
                            color = (0, 0, shade)
                        else:
                            color = (shade, 0, 0)
                        img_f.circle(image, p1, 2, color, -1)
                        if alignmentPointsPred[
                                name] is not None and j in alignmentPointsPred[
                                    name][b]:
                            mid = p1  #( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
                            rad = 4  #round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
                            #print(mid)
                            #print(rad)
                            #img_f.circle(image,mid,rad,(0,1,1),1)
                    #for j in alignmentBBsTarg[name][b]:
                    #    p1 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                    #    p2 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                    #    mid = ( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
                    #    rad = round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
                    #    #print(mid)
                    #    #print(rad)
                    #    img_f.circle(image,mid,rad,(1,0,1),1)

                    saveName = '{:06}_{}'.format(startIndex + b, name)
                    #for j in range(metricsOut.shape[1]):
                    #    saveName+='_m:{0:.3f}'.format(metricsOut[i,j])
                    saveName += '.png'
                    img_f.imwrite(os.path.join(outDir, saveName), image)

            if 'pixels' in out:
                image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                                (1, 2, 0))) / 2.0)).copy()
                if outputPixels is not None:
                    for ch in range(outputPixels.shape[1]):
                        image[:, :, ch] = 1 - outputPixels[b, ch, :, :]
                    saveName = '{:06}_pixels.png'.format(startIndex + b, name)
                    img_f.imwrite(os.path.join(outDir, saveName), image)
            #print('finished writing {}'.format(startIndex+b))

    #return metricsOut
    toRet = log
    #toRet=   { 'ap_5':aps_5,
    #             #'class_aps': class_aps,
    #             #'ap_3':aps_3,
    #             #'ap_7':aps_7,
    #           'recall':recalls_5,
    #           'prec':precs_5,
    #           'nn_loss': nn_loss,
    #         }
    #for i in range(numClasses):
    #    toRet['class{}_ap'.format(i)]=class_aps[i]

    return (toRet, None)
예제 #2
0
def FormsBoxDetect_printer(config,
                           instance,
                           model,
                           gpu,
                           metrics,
                           outDir=None,
                           startIndex=None,
                           lossFunc=None):
    def __eval_metrics(data, target):
        acc_metrics = np.zeros((output.shape[0], len(metrics)))
        for ind in range(output.shape[0]):
            for i, metric in enumerate(metrics):
                acc_metrics[ind, i] += metric(output[ind:ind + 1],
                                              target[ind:ind + 1])
        return acc_metrics

    def __to_tensor_old(data, gpu):
        if type(data) is np.ndarray:
            data = torch.FloatTensor(data.astype(np.float32))
        elif type(data) is torch.Tensor:
            data = data.type(torch.FloatTensor)
        if gpu is not None:
            data = data.to(gpu)
        return data

    def __to_tensor(instance, gpu):
        data = instance['img']
        if 'bb_gt' in instance:
            targetBBs = instance['bb_gt']
            targetBBs_sizes = instance['bb_sizes']
        else:
            targetBBs = {}
            targetBBs_sizes = {}
        target_num_neighbors = instance[
            'num_neighbors'] if 'num_neighbors' in instance else None
        if 'point_gt' in instance:
            targetPoints = instance['point_gt']
            targetPoints_sizes = instance['point_label_sizes']
        else:
            targetPoints = {}
            targetPoints_sizes = {}
        if 'pixel_gt' in instance:
            targetPixels = instance['pixel_gt']
        else:
            targetPixels = None
        if type(data) is np.ndarray:
            data = torch.FloatTensor(data.astype(np.float32))
        elif type(data) is torch.Tensor:
            data = data.type(torch.FloatTensor)

        def sendToGPU(targets):
            new_targets = {}
            for name, target in targets.items():
                if target is not None:
                    new_targets[name] = target.to(gpu)
                else:
                    new_targets[name] = None
            return new_targets

        if gpu is not None:
            data = data.to(gpu)
            if targetBBs is not None:
                targetBBs = targetBBs.to(gpu)
            targetPoints = sendToGPU(targetPoints)
            if targetPixels is not None:
                targetPixels = targetPixels.to(gpu)
            if target_num_neighbors is not None:
                target_num_neighbors = target_num_neighbors.to(gpu)
        return data, targetBBs, targetBBs_sizes, targetPoints, targetPoints_sizes, targetPixels, target_num_neighbors

    THRESH = config['THRESH'] if 'THRESH' in config else 0.92
    numClasses = 2
    #print(type(instance['pixel_gt']))
    #if type(instance['pixel_gt']) == list:
    #    print(instance)
    #    print(startIndex)
    #data, targetBB, targetBBSizes = instance
    if lossFunc is None:
        yolo_loss = YoloLoss(model.numBBTypes, model.rotation, model.scale,
                             model.anchors, **config['loss_params']['box'])
    else:
        yolo_loss = lossFunc
    data = instance['img']
    batchSize = data.shape[0]
    targetBBs = instance['bb_gt']
    targetPoints = instance['point_gt']
    targetPixels = instance['pixel_gt']
    imageName = instance['imgName']
    scale = instance['scale']
    target_num_neighbors = instance['num_neighbors']
    if not model.predNumNeighbors:
        del instance['num_neighbors']
    dataT, targetBBsT, targetBBsSizes, targetPointsT, targetPointsSizes, targetPixelsT, target_num_neighborsT = __to_tensor(
        instance, gpu)

    pretty = config['pretty'] if 'pretty' in config else False

    resultsDirName = 'results'
    #if outDir is not None and resultsDirName is not None:
    #rPath = os.path.join(outDir,resultsDirName)
    #if not os.path.exists(rPath):
    #    os.mkdir(rPath)
    #for name in targetBBs:
    #    nPath = os.path.join(rPath,name)
    #    if not os.path.exists(nPath):
    #        os.mkdir(nPath)

    #dataT = __to_tensor(data,gpu)
    #print('{}: {} x {}'.format(imageName,data.shape[2],data.shape[3]))
    outputBBs, outputOffsets, outputLines, outputOffsetLines, outputPoints, outputPixels = model(
        dataT)
    if outputPixels is not None:
        outputPixels = torch.sigmoid(outputPixels)
    index = 0
    loss = 0
    index = 0
    ttt_hit = True
    #if 22>=startIndex and 22<startIndex+batchSize:
    #    ttt_hit=22-startIndex
    #else:
    #    return 0
    lossThis, position_loss, conf_loss, class_loss, nn_loss, recall, precision = yolo_loss(
        outputOffsets, targetBBsT, targetBBsSizes, target_num_neighborsT)
    alignmentPointsPred = {}
    alignmentPointsTarg = {}
    index = 0
    for name, targ in targetPointsT.items():
        #print(outputPoints[0].shape)
        #print(targetPointsSizes)
        #print('{} {}'.format(index, name))
        lossThis, predIndexes, targetPointsIndexes = alignment_loss(
            outputPoints[index],
            targ,
            targetPointsSizes[name],
            **config['loss_params']['point'],
            return_alignment=True,
            debug=ttt_hit,
            points=True)
        alignmentPointsPred[name] = predIndexes
        alignmentPointsTarg[name] = targetPointsIndexes
        index += 1

    data = data.cpu().data.numpy()
    maxConf = outputBBs[:, :, 0].max().item()
    threshConf = max(maxConf * THRESH, 0.5)

    if model.rotation:
        outputBBs = non_max_sup_dist(outputBBs.cpu(), threshConf, 3)
    else:
        outputBBs = non_max_sup_iou(outputBBs.cpu(), threshConf, 0.4)

    numClasses = model.numBBTypes
    #aps_3=[]
    aps_5 = []
    class_aps = [[] for i in range(numClasses)]
    aps_5all = []
    #aps_7=[]
    recalls_5 = []
    precs_5 = []
    if 'no_blanks' in config[
            'data_loader'] and not config['data_loader']['no_blanks']:
        numClasses -= 1
    nn_acc = [-1] * batchSize

    for b in range(batchSize):
        if model.predNumNeighbors:
            extraPreds = 1
            #predNN=outputBBs[b][-(numClasses+1)]
            #diffs=torch.abs(predNN-target_num_neighborsT[:,bbAlignment].float())
            #nn_acc[b] = (diffs<0.5).sum().item()
            #nn_acc[b] /= predNN.size(0)
        else:
            extraPreds = 0
        #outputBBs[b]=torch.cat((outputBBs[b][:,0:6],outputBBs[b][:,7:]),dim=1)
        #extraPreds=0

        if targetBBs is not None:
            target_for_b = targetBBs[b, :targetBBsSizes[b], :]
        else:
            target_for_b = torch.empty(0)
        if model.rotation:
            ap_5, prec_5, recall_5, class_ap = AP_dist(target_for_b,
                                                       outputBBs[b],
                                                       0.9,
                                                       numClasses,
                                                       beforeCls=extraPreds,
                                                       getClassAP=True)
            #ap_3, prec_3, recall_3 =AP_dist(target_for_b,outputBBs[b],1.3,numClasses,beforeCls=extraPreds)
            #ap_7, prec_7, recall_7 =AP_dist(target_for_b,outputBBs[b],0.5,numClasses,beforeCls=extraPreds)
        else:
            ap_5, prec_5, recall_5, class_ap = AP_iou(target_for_b,
                                                      outputBBs[b],
                                                      0.5,
                                                      numClasses,
                                                      beforeCls=extraPreds,
                                                      getClassAP=True)
            #ap_3, prec_3, recall_3 =AP_iou(target_for_b,outputBBs[b],0.3,numClasses,beforeCls=extraPreds)
            #ap_7, prec_7, recall_7 =AP_iou(target_for_b,outputBBs[b],0.7,numClasses,beforeCls=extraPreds)

        if ap_5 is not None:
            aps_5.append(ap_5)
            aps_5all.append(ap_5)
        else:
            aps_5all.append(-1)
        for i in range(numClasses):
            if class_ap[i] is not None:
                class_aps[i].append(class_ap[i])
        #aps_3.append(ap_3 )
        #aps_7.append(ap_7 )
        recalls_5.append(recall_5)
        precs_5.append(prec_5)
        #for b in range(len(outputBBs)):
        outputBBs[b] = outputBBs[b].data.numpy()

    outputPointsOld = outputPoints
    targetPointsOld = targetPoints
    outputPoints = {}
    targetPoints = {}
    i = 0
    for name, targ in targetPointsOld.items():
        if targ is not None:
            targetPoints[name] = targ.data.numpy()
        else:
            targetPoints[name] = None
        outputPoints[name] = outputPointsOld[i].cpu().data.numpy()
        i += 1
    if outputPixels is not None:
        outputPixels = outputPixels.cpu().data.numpy()
    #metricsOut = __eval_metrics(outputBBs,targetBBs)
    #metricsOut = 0
    #if outDir is None:
    #    return metricsOut

    dists = defaultdict(list)
    dists_x = defaultdict(list)
    dists_y = defaultdict(list)
    scaleDiffs = defaultdict(list)
    rotDiffs = defaultdict(list)

    allPredNNs = []
    for b in range(batchSize):
        #print('image {} has {} {}'.format(startIndex+b,targetBBsSizes[name][b],name))
        #bbImage = np.ones_like(image)
        bbs = outputBBs[b]
        if bbs.shape[0] > 0:
            if model.predNumNeighbors:
                predNN = bbs[:, 6]
                allPredNNs += predNN.tolist()
                predClass = bbs[:, 7:]
            else:
                predClass = bbs[:, 6:]
        else:
            predNN = bbs  #i.e. a zero size tensor
            predClass = bbs

        if 'save_json' in config:
            assert (batchSize == 1)
            scale = scale[0]
            if targetBBs is not None:
                if model.rotation:
                    targIndex, predWithNoIntersection = getTargIndexForPreds_dist(
                        targetBBs[b], torch.from_numpy(bbs), 1.1, numClasses,
                        extraPreds)
                else:
                    targIndex, predWithNoIntersection = getTargIndexForPreds_iou(
                        targetBBs[b], torch.from_numpy(bbs), 0.4, numClasses,
                        extraPreds)
                newId = targetBBs[b].size(0)
            else:
                targIndex = -1 * torch.ones(bbs.shape[0])
                newId = 1
            bbsData = []
            for j in range(bbs.shape[0]):
                tl, tr, br, bl = getCorners(bbs[j, 1:])
                id = targIndex[j].item()
                if id < 0:
                    id = 'u{}'.format(newId)
                    newId += 1
                else:
                    id = 'm{}'.format(id)
                bb = {
                    'id':
                    id,
                    'poly_points':
                    [[float(tl[0] / scale),
                      float(tl[1] / scale)],
                     [float(tr[0] / scale),
                      float(tr[1] / scale)],
                     [float(br[0] / scale),
                      float(br[1] / scale)],
                     [float(bl[0] / scale),
                      float(bl[1] / scale)]],
                    'type':
                    'detectorPrediction',
                    'textPred':
                    float(predClass[j, 0]),
                    'fieldPred':
                    float(predClass[j, 1])
                }
                if numClasses == 2 and model.numBBTypes == 3:
                    bb['blankPred'] = float(predClass[j, 2])
                if model.predNumNeighbors:
                    bb['nnPred'] = float(predNN[j])
                bbsData.append(bb)

            if instance['pairs'] is None:
                #import pdb; pdb.set_trace()
                instance['pairs'] = []
            pairsData = [('m{}'.format(i1), 'm{}'.format(i2))
                         for i1, i2 in instance['pairs']]

            ensure_dir(config['save_json'])
            saveJSON = os.path.join(config['save_json'],
                                    imageName[b] + '.json')
            allData = {'textBBs': bbsData, 'fieldBBs': [], 'pairs': pairsData}
            with open(saveJSON, 'w') as f:
                json.dump(allData, f)
                print('wrote {}'.format(saveJSON))

        if outDir is not None:
            #Write the results so we can train LF with them
            #saveFile = os.path.join(outDir,resultsDirName,name,'{}'.format(imageName[b]))
            #we must rescale the output to be according to the original image
            #rescaled_outputBBs_xyrs = outputBBs_xyrs[name][b]
            #rescaled_outputBBs_xyrs[:,1] /= scale[b]
            #rescaled_outputBBs_xyrs[:,2] /= scale[b]
            #rescaled_outputBBs_xyrs[:,4] /= scale[b]

            #np.save(saveFile,rescaled_outputBBs_xyrs)
            image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                            (1, 2, 0))) / 2.0)).copy()
            if image.shape[2] == 1:
                image = cv2.cvtColor(image, cv2.COLOR_GRAY2RGB)
            #if name=='text_start_gt':

            if not pretty:
                for j in range(targetBBsSizes[b]):
                    plotRect(image, (1, 0.5, 0), targetBBs[b, j, 0:5])
                    if model.predNumNeighbors:
                        x = int(targetBBs[b, j, 0])
                        y = int(targetBBs[b, j, 1] + targetBBs[b, j, 3])
                        cv2.putText(image, '{:.2f}'.format(gtNumNeighbors[b,
                                                                          j]),
                                    (x, y), cv2.FONT_HERSHEY_SIMPLEX, 0.5,
                                    (0.6, 0.3, 0), 2, cv2.LINE_AA)
                #if alignmentBBs[b] is not None:
                #    aj=alignmentBBs[b][j]
                #    xc_gt = targetBBs[b,j,0]
                #    yc_gt = targetBBs[b,j,1]
                #    xc=outputBBs[b,aj,1]
                #    yc=outputBBs[b,aj,2]
                #    cv2.line(image,(xc,yc),(xc_gt,yc_gt),(0,1,0),1)
                #    shade = 0.0+(outputBBs[b,aj,0]-threshConf)/(maxConf-threshConf)
                #    shade = max(0,shade)
                #    if outputBBs[b,aj,6] > outputBBs[b,aj,7]:
                #        color=(0,shade,shade) #text
                #    else:
                #        color=(shade,shade,0) #field
                #    plotRect(image,color,outputBBs[b,aj,1:6])

            #bbs=[]
            #pred_points=[]
            #maxConf = outputBBs[b,:,0].max()
            #threshConf = 0.5
            #threshConf = max(maxConf*0.9,0.5)
            #print("threshConf:{}".format(threshConf))
            #for j in range(outputBBs.shape[1]):
            #    conf = outputBBs[b,j,0]
            #    if conf>threshConf:
            #        bbs.append((conf,j))
            #    #pred_points.append(
            #bbs.sort(key=lambda a: a[0]) #so most confident bbs are draw last (on top)
            #import pdb; pdb.set_trace()
            for j in range(bbs.shape[0]):
                #circle aligned predictions
                conf = bbs[j, 0]
                if outDir is not None:
                    shade = 0.0 + (conf - threshConf) / (maxConf - threshConf)
                    #print(shade)
                    #if name=='text_start_gt' or name=='field_end_gt':
                    #    cv2.bb(bbImage[:,:,1],p1,p2,shade,2)
                    #if name=='text_end_gt':
                    #    cv2.bb(bbImage[:,:,2],p1,p2,shade,2)
                    #elif name=='field_end_gt' or name=='field_start_gt':
                    #    cv2.bb(bbImage[:,:,0],p1,p2,shade,2)
                    if predClass[j, 0] > predClass[j, 1]:
                        color = [0, 0, shade]  #text
                    else:
                        if pretty:
                            color = [0, shade, shade]
                        else:
                            color = [shade, 0, 0]  #field
                    if numClasses == 2 and model.numBBTypes == 3 and predClass[
                            j, 2] > 0.5:
                        color[1] = shade
                    if pretty:
                        lineW = 2
                    else:
                        lineW = 1
                    plotRect(image, color, bbs[j, 1:6], lineW)
                    if model.predNumNeighbors and not pretty:
                        x = int(bbs[j, 1])
                        y = int(bbs[j, 2] - bbs[j, 4])
                        #color = int(min(abs(predNN[j]-target_num_neighbors[j]),2)*127)
                        #cv2.putText(image,'{}/{}'.format(predNN[j],target_num_neighbors[j]),(x,y), cv2.FONT_HERSHEY_SIMPLEX, 3,(color,0,0),2,cv2.LINE_AA)
                        cv2.putText(image, '{:.2f}'.format(predNN[j]), (x, y),
                                    cv2.FONT_HERSHEY_SIMPLEX, 0.5, color, 2,
                                    cv2.LINE_AA)

            #for j in alignmentBBsTarg[name][b]:
            #    p1 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
            #    p2 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
            #    mid = ( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
            #    rad = round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
            #    #print(mid)
            #    #print(rad)
            #    cv2.circle(image,mid,rad,(1,0,1),1)

            saveName = '{}_boxes_AP:{:.2f}'.format(imageName[b], aps_5all[b])
            #for j in range(metricsOut.shape[1]):
            #    saveName+='_m:{0:.3f}'.format(metricsOut[i,j])
            saveName += '.png'
            io.imsave(os.path.join(outDir, saveName), image)
            #print('saved: '+os.path.join(outDir,saveName))

            for name, out in outputPoints.items():
                image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                                (1, 2, 0))) / 2.0)).copy()
                #if name=='text_start_gt':
                for j in range(targetPointsSizes[name][b]):
                    p1 = (targetPoints[name][b, j, 0], targetPoints[name][b, j,
                                                                          1])
                    cv2.circle(image, p1, 2, (1, 0.5, 0), -1)
                points = []
                maxConf = max(out[b, :, 0].max(), 1.0)
                threshConf = maxConf * 0.1
                for j in range(out.shape[1]):
                    conf = out[b, j, 0]
                    if conf > threshConf:
                        p1 = (out[b, j, 1], out[b, j, 2])
                        points.append((conf, p1, j))
                points.sort(key=lambda a: a[0]
                            )  #so most confident bbs are draw last (on top)
                for conf, p1, j in points:
                    shade = 0.0 + conf / maxConf
                    if name == 'table_points':
                        color = (0, 0, shade)
                    else:
                        color = (shade, 0, 0)
                    cv2.circle(image, p1, 2, color, -1)
                    if alignmentPointsPred[
                            name] is not None and j in alignmentPointsPred[
                                name][b]:
                        mid = p1  #( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
                        rad = 4  #round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
                        #print(mid)
                        #print(rad)
                        #cv2.circle(image,mid,rad,(0,1,1),1)
                #for j in alignmentBBsTarg[name][b]:
                #    p1 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                #    p2 = (targetBBs[name][b,j,0], targetBBs[name][b,j,1])
                #    mid = ( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
                #    rad = round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
                #    #print(mid)
                #    #print(rad)
                #    cv2.circle(image,mid,rad,(1,0,1),1)

                saveName = '{:06}_{}'.format(startIndex + b, name)
                #for j in range(metricsOut.shape[1]):
                #    saveName+='_m:{0:.3f}'.format(metricsOut[i,j])
                saveName += '.png'
                io.imsave(os.path.join(outDir, saveName), image)

            image = (1 - ((1 + np.transpose(data[b][:, :, :],
                                            (1, 2, 0))) / 2.0)).copy()
            if outputPixels is not None:
                for ch in range(outputPixels.shape[1]):
                    image[:, :, ch] = 1 - outputPixels[b, ch, :, :]
                saveName = '{:06}_pixels.png'.format(startIndex + b, name)
                io.imsave(os.path.join(outDir, saveName), image)
            #print('finished writing {}'.format(startIndex+b))

    #return metricsOut
    toRet = {
        'ap_5': aps_5,
        #'class_aps': class_aps,
        #'ap_3':aps_3,
        #'ap_7':aps_7,
        'recall': recalls_5,
        'prec': precs_5,
        'nn_loss': nn_loss,
    }
    for i in range(numClasses):
        toRet['class{}_ap'.format(i)] = class_aps[i]

    return (toRet, (lossThis, position_loss, conf_loss, class_loss, nn_loss,
                    recall, precision, allPredNNs))
def FormsGraphPair_printer(config,instance, model, gpu, metrics, outDir=None, startIndex=None, lossFunc=None):
    def __eval_metrics(data,target):
        acc_metrics = np.zeros((output.shape[0],len(metrics)))
        for ind in range(output.shape[0]):
            for i, metric in enumerate(metrics):
                acc_metrics[ind,i] += metric(output[ind:ind+1], target[ind:ind+1])
        return acc_metrics

    def __to_tensor(instance,gpu):
        image = instance['img']
        bbs = instance['bb_gt']
        adjaceny = instance['adj']
        num_neighbors = instance['num_neighbors']

        if gpu is not None:
            image = image.to(gpu)
            if bbs is not None:
                bbs = bbs.to(gpu)
            if num_neighbors is not None:
                num_neighbors = num_neighbors.to(gpu)
            #adjacenyMatrix = adjacenyMatrix.to(self.gpu)
        return image, bbs, adjaceny, num_neighbors

    rel_thresholds = [config['THRESH']] if 'THRESH' in config else [0.5]
    if ('sweep_threshold' in config and config['sweep_threshold']) or ('sweep_thresholds' in config and config['sweep_thresholds']):
        rel_thresholds = np.arange(0.1,1.0,0.05)
    if ('sweep_threshold_big' in config and config['sweep_threshold_big']) or ('sweep_thresholds_big' in config and config['sweep_thresholds_big']):
        rel_thresholds = np.arange(0,20.0,1)
    if ('sweep_threshold_small' in config and config['sweep_threshold_small']) or ('sweep_thresholds_small' in config and config['sweep_thresholds_small']):
        rel_thresholds = np.arange(0,0.1,0.01)
    draw_rel_thresh = config['draw_thresh'] if 'draw_thresh' in config else rel_thresholds[0]
    #print(type(instance['pixel_gt']))
    #if type(instance['pixel_gt']) == list:
    #    print(instance)
    #    print(startIndex)
    #data, targetBB, targetBBSizes = instance
    lossWeights = config['loss_weights'] if 'loss_weights' in config else {"box": 1, "rel":1}
    if lossFunc is None:
        yolo_loss = YoloLoss(model.numBBTypes,model.rotation,model.scale,model.anchors,**config['loss_params']['box'])
    else:
        yolo_loss = lossFunc
    data = instance['img']
    batchSize = data.shape[0]
    assert(batchSize==1)
    targetBoxes = instance['bb_gt']
    adjacency = instance['adj']
    adjacency = list(adjacency)
    imageName = instance['imgName']
    scale = instance['scale']
    target_num_neighbors = instance['num_neighbors']
    if not model.detector.predNumNeighbors:
        instance['num_neighbors']=None
    dataT, targetBoxesT, adjT, target_num_neighborsT = __to_tensor(instance,gpu)


    pretty = config['pretty'] if 'pretty' in config else False
    useDetections = config['useDetections'] if 'useDetections' in config else False
    if 'useDetect' in config:
        useDetections = config['useDetect']
    confThresh = config['conf_thresh'] if 'conf_thresh' in config else None


    numClasses=2 #TODO no hard code

    resultsDirName='results'
    #if outDir is not None and resultsDirName is not None:
        #rPath = os.path.join(outDir,resultsDirName)
        #if not os.path.exists(rPath):
        #    os.mkdir(rPath)
        #for name in targetBoxes:
        #    nPath = os.path.join(rPath,name)
        #    if not os.path.exists(nPath):
        #        os.mkdir(nPath)

    #dataT = __to_tensor(data,gpu)
    #print('{}: {} x {}'.format(imageName,data.shape[2],data.shape[3]))
    if useDetections=='gt':
        outputBoxes, outputOffsets, relPred, relIndexes, bbPred = model(dataT,targetBoxesT,target_num_neighborsT,True,
                otherThresh=confThresh,
                otherThreshIntur=1 if confThresh is not None else None,
                hard_detect_limit=600)
        outputBoxes=torch.cat((torch.ones(targetBoxes.size(1),1),targetBoxes[0,:,0:5],targetBoxes[0,:,-numClasses:]),dim=1) #add score
    elif type(useDetections) is str:
        dataset=config['DATASET']
        jsonPath = os.path.join(useDetections,imageName+'.json')
        with open(os.path.join(jsonPath)) as f:
            annotations = json.loads(f.read())
        fixAnnotations(dataset,annotations)
        savedBoxes = torch.FloatTensor(len(annotations['byId']),6+model.detector.predNumNeighbors+numClasses)
        for i,(id,bb) in enumerate(annotations['byId'].items()):
            qX, qY, qH, qW, qR, qIsText, qIsField, qIsBlank, qNN = getBBInfo(bb,dataset.rotate,useBlankClass=not dataset.no_blanks)
            savedBoxes[i,0]=1 #conf
            savedBoxes[i,1]=qX*scale #x-center, already scaled
            savedBoxes[i,2]=qY*scale #y-center
            savedBoxes[i,3]=qR #rotation
            savedBoxes[i,4]=qH*scale/2
            savedBoxes[i,5]=qW*scale/2
            if model.detector.predNumNeighbors:
                extra=1
                savedBoxes[i,6]=qNN
            else:
                extra=0
            savedBoxes[i,6+extra]=qIsText
            savedBoxes[i,7+extra]=qIsField
            
        if gpu is not None:
            savedBoxes=savedBoxes.to(gpu)
        outputBoxes, outputOffsets, relPred, relIndexes, bbPred = model(dataT,savedBoxes,None,"saved",
                otherThresh=confThresh,
                otherThreshIntur=1 if confThresh is not None else None,
                hard_detect_limit=600)
        outputBoxes=savedBoxes.cpu()
    elif useDetections:
        print('Unknown detection flag: '+useDetections)
        exit()
    else:
        outputBoxes, outputOffsets, relPred, relIndexes, bbPred = model(dataT,
                otherThresh=confThresh,
                otherThreshIntur=1 if confThresh is not None else None,
                hard_detect_limit=600)

    if model.predNN and bbPred is not None:
        predNN = bbPred[:,0]
    else:
        predNN=None
    if  model.detector.predNumNeighbors and not useDetections:
        #useOutputBBs=torch.cat((outputBoxes[:,0:6],outputBoxes[:,7:]),dim=1) #throw away NN pred
        extraPreds=1
        if not model.predNN:
            predNN = outputBoxes[:,6]
    else:
        extraPreds=0
        if not model.predNN:
            predNN = None
        #useOutputBBs=outputBoxes

    if targetBoxesT is not None:
        targetSize=targetBoxesT.size(1)
    else:
        targetSize=0
    lossThis, position_loss, conf_loss, class_loss, nn_loss, recall, precision = yolo_loss(outputOffsets,targetBoxesT,[targetSize], target_num_neighborsT)

    if 'rule' in config:
        if config['rule']=='closest':
            dists = torch.FloatTensor(relPred.size())
            differentClass = torch.FloatTensor(relPred.size())
            predClasses = torch.argmax(outputBoxes[:,extraPreds+6:extraPreds+6+numClasses],dim=1)
            for i,(bb1,bb2) in enumerate(relIndexes):
                dists[i] = math.sqrt((outputBoxes[bb1,1]-outputBoxes[bb2,1])**2 + (outputBoxes[bb1,2]-outputBoxes[bb2,2])**2)
                differentClass[i] = predClasses[bb1]!=predClasses[bb2]
            maxDist = torch.max(dists)
            minDist = torch.min(dists)
            relPred = 1-(dists-minDist)/(maxDist-minDist)
            relPred *= differentClass
        elif config['rule']=='icdar':
            height = torch.FloatTensor(relPred.size())
            dists = torch.FloatTensor(relPred.size())
            right = torch.FloatTensor(relPred.size())
            sameClass = torch.FloatTensor(relPred.size())
            predClasses = torch.argmax(outputBoxes[:,extraPreds+6:extraPreds+6+numClasses],dim=1)
            for i,(bb1,bb2) in enumerate(relIndexes):
                sameClass[i] = predClasses[bb1]==predClasses[bb2]
                
                #g4 of the paper
                height[i] = max(outputBoxes[bb1,4],outputBoxes[bb2,4])/min(outputBoxes[bb1,4],outputBoxes[bb2,4])

                #g5 of the paper
                if predClasses[bb1]==0:
                    widthLabel = outputBoxes[bb1,5]*2 #we predict half width
                    widthValue = outputBoxes[bb2,5]*2
                    dists[i] = math.sqrt(((outputBoxes[bb1,1]+widthLabel)-(outputBoxes[bb2,1]-widthValue))**2 + (outputBoxes[bb1,2]-outputBoxes[bb2,2])**2)
                else:
                    widthLabel = outputBoxes[bb2,5]*2 #we predict half width
                    widthValue = outputBoxes[bb1,5]*2
                    dists[i] = math.sqrt(((outputBoxes[bb1,1]-widthValue)-(outputBoxes[bb2,1]+widthLabel))**2 + (outputBoxes[bb1,2]-outputBoxes[bb2,2])**2)
                if dists[i]>2*widthLabel:
                    dists[i]/=widthLabel
                else: #undefined
                    dists[i] = min(1,dists[i]/widthLabel)
            
                #g6 of the paper
                if predClasses[bb1]==0:
                    widthValue = outputBoxes[bb2,5]*2
                    hDist = outputBoxes[bb1,1]-outputBoxes[bb2,1]
                else:
                    widthValue = outputBoxes[bb1,5]*2
                    hDist = outputBoxes[bb2,1]-outputBoxes[bb1,1]
                right[i] = hDist/widthValue

            relPred = 1-(height+dists+right + 10000*sameClass)
        else:
            print('ERROR, unknown rule {}'.format(config['rule']))
            exit()
    elif relPred is not None:
        relPred = torch.sigmoid(relPred)[:,0]




    relCand = relIndexes
    if relCand is None:
        relCand=[]

    if model.rotation:
        bbAlignment, bbFullHit = getTargIndexForPreds_dist(targetBoxes[0],outputBoxes,0.9,numClasses,extraPreds,hard_thresh=False)
    else:
        bbAlignment, bbFullHit = getTargIndexForPreds_iou(targetBoxes[0],outputBoxes,0.5,numClasses,extraPreds,hard_thresh=False)
    if targetBoxes is not None:
        target_for_b = targetBoxes[0,:,:]
    else:
        target_for_b = torch.empty(0)

    if outputBoxes.size(0)>0:
        maxConf = outputBoxes[:,0].max().item()
        minConf = outputBoxes[:,0].min().item()
        if useDetections:
            minConf=0
    #threshConf = max(maxConf*THRESH,0.5)
    #if model.rotation:
    #    outputBoxes = non_max_sup_dist(outputBoxes.cpu(),threshConf,3)
    #else:
    #    outputBoxes = non_max_sup_iou(outputBoxes.cpu(),threshConf,0.4)
    if model.rotation:
        ap_5, prec_5, recall_5 =AP_dist(target_for_b,outputBoxes,0.9,model.numBBTypes,beforeCls=extraPreds)
    else:
        ap_5, prec_5, recall_5 =AP_iou(target_for_b,outputBoxes,0.5,model.numBBTypes,beforeCls=extraPreds)

    #precisionHistory={}
    #precision=-1
    #minStepSize=0.025
    #targetPrecisions=[None]
    #for targetPrecision in targetPrecisions:
    #    if len(precisionHistory)>0:
    #        closestPrec=9999
    #        for prec in precisionHistory:
    #            if abs(targetPrecision-prec)<abs(closestPrec-targetPrecision):
    #                closestPrec=prec
    #        precision=prec
    #        stepSize=precisionHistory[prec][0]
    #    else:
    #        stepSize=0.1
    #
    #    while True: #abs(precision-targetPrecision)>0.001:
    toRet={}
    for rel_threshold in rel_thresholds:

            if 'optimize' in config and config['optimize']:
                if 'penalty' in config:
                    penalty = config['penalty']
                else:
                    penalty = 0.25
                print('optimizing with penalty {}'.format(penalty))
                thresh=0.15
                while thresh<0.45:
                    keep = relPred>thresh
                    newRelPred = relPred[keep]
                    if newRelPred.size(0)<700:
                        break
                if newRelPred.size(0)>0:
                    #newRelCand = [ cand for i,cand in enumerate(relCand) if keep[i] ]
                    usePredNN= predNN is not None and config['optimize']!='gt'
                    idMap={}
                    newId=0
                    newRelCand=[]
                    numNeighbors=[]
                    for index,(id1,id2) in enumerate(relCand):
                        if keep[index]:
                            if id1 not in idMap:
                                idMap[id1]=newId
                                if not usePredNN:
                                    numNeighbors.append(target_num_neighbors[0,bbAlignment[id1]])
                                else:
                                    numNeighbors.append(predNN[id1])
                                newId+=1
                            if id2 not in idMap:
                                idMap[id2]=newId
                                if not usePredNN:
                                    numNeighbors.append(target_num_neighbors[0,bbAlignment[id2]])
                                else:
                                    numNeighbors.append(predNN[id2])
                                newId+=1
                            newRelCand.append( [idMap[id1],idMap[id2]] )            


                    #if not usePredNN:
                        #    decision = optimizeRelationships(newRelPred,newRelCand,numNeighbors,penalty)
                    #else:
                    decision= optimizeRelationshipsSoft(newRelPred,newRelCand,numNeighbors,penalty, rel_threshold)
                    decision= torch.from_numpy( np.round_(decision).astype(int) )
                    decision=decision.to(relPred.device)
                    relPred[keep] = torch.where(0==decision,relPred[keep]-1,relPred[keep])
                    relPred[1-keep] -=1
                    rel_threshold_use=0#-0.5
                else:
                    rel_threshold_use=rel_threshold
            else:
                rel_threshold_use=rel_threshold

            #threshed in model
            #if len(precisionHistory)==0:
            if len(toRet)==0:
                #align bb predictions (final) with GT
                if bbPred is not None and bbPred.size(0)>0:
                    #create aligned GT
                    #this was wrong...
                        #first, remove unmatched predicitons that didn't overlap (weren't close) to any targets
                        #toKeep = 1-((bbNoIntersections==1) * (bbAlignment==-1))
                    #remove predictions that overlapped with GT, but not enough
                    if model.predNN:
                        start=1
                        toKeep = 1-((bbFullHit==0) * (bbAlignment!=-1)) #toKeep = not (incomplete_overlap and did_overlap)
                        if toKeep.any():
                            bbPredNN_use = bbPred[toKeep][:,0]
                            bbAlignment_use = bbAlignment[toKeep]
                            #becuase we used -1 to indicate no match (in bbAlignment), we add 0 as the last position in the GT, as unmatched 
                            if target_num_neighborsT is not None:
                                target_num_neighbors_use = torch.cat((target_num_neighborsT[0].float(),torch.zeros(1).to(target_num_neighborsT.device)),dim=0)
                            else:
                                target_num_neighbors_use = torch.zeros(1).to(bbPred.device)
                            alignedNN_use = target_num_neighbors_use[bbAlignment_use]
                        else:
                            bbPredNN_use=None
                            alignedNN_use=None
                    else:
                        start=0
                    if model.predClass:
                        #We really don't care about the class of non-overlapping instances
                        if targetBoxes is not None:
                            toKeep = bbFullHit==1
                            if toKeep.any():
                                bbPredClass_use = bbPred[toKeep][:,start:start+model.numBBTypes]
                                bbAlignment_use = bbAlignment[toKeep]
                                alignedClass_use =  targetBoxesT[0][bbAlignment_use][:,13:13+model.numBBTypes] #There should be no -1 indexes in hereS
                            else:
                                bbPredClass_use=None
                                alignedClass_use=None
                        else:
                            alignedClass_use = None
                else:
                    bbPredNN_use = None
                    bbPredClass_use = None
                if model.predNN and bbPredNN_use is not None and bbPredNN_use.size(0)>0:
                    nn_loss_final = F.mse_loss(bbPredNN_use,alignedNN_use)
                    #nn_loss_final *= self.lossWeights['nn']

                    #loss += nn_loss_final
                    nn_loss_final = nn_loss_final.item()
                else:
                    nn_loss_final=0
                if model.predNN and predNN is not None:
                    predNN_p=bbPred[:,0]
                    diffs=torch.abs(predNN_p-target_num_neighborsT[0][bbAlignment].float())
                    nn_acc = (diffs<0.5).sum().item()
                    nn_acc /= predNN.size(0)
                elif model.predNN:
                    nn_acc = 0 
                if model.detector.predNumNeighbors and not useDetections:
                    predNN_d = outputBoxes[:,6]
                    diffs=torch.abs(predNN_d-target_num_neighbors[0][bbAlignment].float())
                    nn_acc_d = (diffs<0.5).sum().item()
                    nn_acc_d /= predNN.size(0)

                if model.predClass and bbPredClass_use is not None and bbPredClass_use.size(0)>0:
                    class_loss_final = F.binary_cross_entropy_with_logits(bbPredClass_use,alignedClass_use)
                    #class_loss_final *= self.lossWeights['class']
                    #loss += class_loss_final
                    class_loss_final = class_loss_final.item()
                else:
                    class_loss_final = 0
            #class_acc=0
            useOutputBBs=None

            truePred=falsePred=badPred=0
            scores=[]
            matches=0
            i=0
            numMissedByHeur=0
            targGotHit=set()
            for i,(n0,n1) in enumerate(relCand):
                t0 = bbAlignment[n0].item()
                t1 = bbAlignment[n1].item()
                if t0>=0 and bbFullHit[n0]:
                    targGotHit.add(t0)
                if t1>=0 and bbFullHit[n1]:
                    targGotHit.add(t1)
                if t0>=0 and t1>=0 and bbFullHit[n0] and bbFullHit[n1]:
                    if (min(t0,t1),max(t0,t1)) in adjacency:
                        matches+=1
                        scores.append( (relPred[i],True) )
                        if relPred[i]>rel_threshold_use:
                            truePred+=1
                    else:
                        scores.append( (relPred[i],False) )
                        if relPred[i]>rel_threshold_use:
                            falsePred+=1
                else:
                    scores.append( (relPred[i],False) )
                    if relPred[i]>rel_threshold_use:
                        badPred+=1
            for i in range(len(adjacency)-matches):
                numMissedByHeur+=1
                scores.append( (float('nan'),True) )
            rel_ap=computeAP(scores)

            numMissedByDetect=0
            for t0,t1 in adjacency:
                if t0 not in targGotHit or t1 not in targGotHit:
                    numMissedByHeur-=1
                    numMissedByDetect+=1
            heurRecall = (len(adjacency)-numMissedByHeur)/len(adjacency)
            detectRecall = (len(adjacency)-numMissedByDetect)/len(adjacency)
            if len(adjacency)>0:
                relRecall = truePred/len(adjacency)
            else:
                relRecall = 1
            #if falsePred>0:
            #    relPrec = truePred/(truePred+falsePred)
            #else:
            #    relPrec = 1
            if falsePred+badPred>0:
                precision = truePred/(truePred+falsePred+badPred)
            else:
                precision = 1
    

            toRet['prec@{}'.format(rel_threshold)]=precision
            toRet['recall@{}'.format(rel_threshold)]=relRecall
            if relRecall+precision>0:
                toRet['F-M@{}'.format(rel_threshold)]=2*relRecall*precision/(relRecall+precision)
            else:
                toRet['F-M@{}'.format(rel_threshold)]=0
            toRet['rel_AP@{}'.format(rel_threshold)]=rel_ap
            #precisionHistory[precision]=(draw_rel_thresh,stepSize)
            #if targetPrecision is not None:
            #    if abs(precision-targetPrecision)<0.001:
            #        break
            #    elif stepSize<minStepSize:
            #        if precision<targetPrecision:
            #            draw_rel_thresh += stepSize*2
            #            continue
            #        else:
            #            break
            #    elif precision<targetPrecision:
            #        draw_rel_thresh += stepSize
            #        if not wasTooSmall:
            #            reverse=True
            #            wasTooSmall=True
            #        else:
            #            reverse=False
            #    else:
            #        draw_rel_thresh -= stepSize
            #        if wasTooSmall:
            #            reverse=True
            #            wasTooSmall=False
            #        else:
            #            reverse=False
            #    if reverse:
            #        stepSize *= 0.5
            #else:
            #    break


            #import pdb;pdb.set_trace()

            #for b in range(len(outputBoxes)):
            
            
            dists=defaultdict(list)
            dists_x=defaultdict(list)
            dists_y=defaultdict(list)
            scaleDiffs=defaultdict(list)
            rotDiffs=defaultdict(list)
            b=0
            #print('image {} has {} {}'.format(startIndex+b,targetBoxesSizes[name][b],name))
            #bbImage = np.ones_like(image):w

    if outDir is not None:
        outputBoxes = outputBoxes.data.numpy()
        data = data.numpy()

        image = (1-((1+np.transpose(data[b][:,:,:],(1,2,0)))/2.0)).copy()
        if image.shape[2]==1:
            image = cv2.cvtColor(image,cv2.COLOR_GRAY2RGB)
        #if name=='text_start_gt':

        #Draw GT bbs
        if not pretty:
            for j in range(targetSize):
                plotRect(image,(1,0.5,0),targetBoxes[0,j,0:5])
            #x=int(targetBoxes[b,j,0])
            #y=int(targetBoxes[b,j,1]+targetBoxes[b,j,3])
            #cv2.putText(image,'{:.2f}'.format(target_num_neighbors[b,j]),(x,y), cv2.FONT_HERSHEY_SIMPLEX, 0.5,(0.6,0.3,0),2,cv2.LINE_AA)
            #if alignmentBBs[b] is not None:
            #    aj=alignmentBBs[b][j]
            #    xc_gt = targetBoxes[b,j,0]
            #    yc_gt = targetBoxes[b,j,1]
            #    xc=outputBoxes[b,aj,1]
            #    yc=outputBoxes[b,aj,2]
            #    cv2.line(image,(xc,yc),(xc_gt,yc_gt),(0,1,0),1)
            #    shade = 0.0+(outputBoxes[b,aj,0]-threshConf)/(maxConf-threshConf)
            #    shade = max(0,shade)
            #    if outputBoxes[b,aj,6] > outputBoxes[b,aj,7]:
            #        color=(0,shade,shade) #text
            #    else:
            #        color=(shade,shade,0) #field
            #    plotRect(image,color,outputBoxes[b,aj,1:6])

        #bbs=[]
        #pred_points=[]
        #maxConf = outputBoxes[b,:,0].max()
        #threshConf = 0.5 
        #threshConf = max(maxConf*0.9,0.5)
        #print("threshConf:{}".format(threshConf))
        #for j in range(outputBoxes.shape[1]):
        #    conf = outputBoxes[b,j,0]
        #    if conf>threshConf:
        #        bbs.append((conf,j))
        #    #pred_points.append(
        #bbs.sort(key=lambda a: a[0]) #so most confident bbs are draw last (on top)
        #import pdb; pdb.set_trace()

        #Draw pred bbs
        bbs = outputBoxes
        for j in range(bbs.shape[0]):
            #circle aligned predictions
            conf = bbs[j,0]
            if outDir is not None:
                shade = 0.0+(conf-minConf)/(maxConf-minConf)
                #print(shade)
                #if name=='text_start_gt' or name=='field_end_gt':
                #    cv2.bb(bbImage[:,:,1],p1,p2,shade,2)
                #if name=='text_end_gt':
                #    cv2.bb(bbImage[:,:,2],p1,p2,shade,2)
                #elif name=='field_end_gt' or name=='field_start_gt':
                #    cv2.bb(bbImage[:,:,0],p1,p2,shade,2)
                if bbs[j,6+extraPreds] > bbs[j,7+extraPreds]:
                    color=(0,0,shade) #text
                else:
                    color=(0,shade,shade) #field
                if pretty=='light':
                    lineWidth=2
                else:
                    lineWidth=1
                plotRect(image,color,bbs[j,1:6],lineWidth)

                if predNN is not None and not pretty: #model.detector.predNumNeighbors:
                    x=int(bbs[j,1])
                    y=int(bbs[j,2])#-bbs[j,4])
                    targ_j = bbAlignment[j].item()
                    if targ_j>=0:
                        gtNN = target_num_neighbors[0,targ_j].item()
                    else:
                        gtNN = 0
                    pred_nn = predNN[j].item()
                    color = min(abs(pred_nn-gtNN),1)#*0.5
                    cv2.putText(image,'{:.2}/{}'.format(pred_nn,gtNN),(x,y), cv2.FONT_HERSHEY_SIMPLEX, 0.5,(color,0,0),2,cv2.LINE_AA)

        #for j in alignmentBBsTarg[name][b]:
        #    p1 = (targetBoxes[name][b,j,0], targetBoxes[name][b,j,1])
        #    p2 = (targetBoxes[name][b,j,0], targetBoxes[name][b,j,1])
        #    mid = ( int(round((p1[0]+p2[0])/2.0)), int(round((p1[1]+p2[1])/2.0)) )
        #    rad = round(math.sqrt((p1[0]-p2[0])**2 + (p1[1]-p2[1])**2)/2.0)
        #    #print(mid)
        #    #print(rad)
        #    cv2.circle(image,mid,rad,(1,0,1),1)

        draw_rel_thresh = relPred.max() * draw_rel_thresh


        #Draw pred pairings
        numrelpred=0
        hits = [False]*len(adjacency)
        for i in range(len(relCand)):
            #print('{},{} : {}'.format(relCand[i][0],relCand[i][1],relPred[i]))
            if pretty:
                if relPred[i]>0 or pretty=='light':
                    score = relPred[i]
                    pruned=False
                    lineWidth=2
                else:
                    score = relPred[i]+1
                    pruned=True
                    lineWidth=1
                #else:
                #    score = (relPred[i]+1)/2
                #    pruned=False
                #    lineWidth=2
                #if pretty=='light':
                #    lineWidth=3
            else:
                lineWidth=1
            if relPred[i]>draw_rel_thresh or (pretty and score>draw_rel_thresh):
                ind1 = relCand[i][0]
                ind2 = relCand[i][1]
                x1 = round(bbs[ind1,1])
                y1 = round(bbs[ind1,2])
                x2 = round(bbs[ind2,1])
                y2 = round(bbs[ind2,2])

                if pretty:
                    targ1 = bbAlignment[ind1].item()
                    targ2 = bbAlignment[ind2].item()
                    aId=None
                    if bbFullHit[ind1] and bbFullHit[ind2]:
                        if (targ1,targ2) in adjacency:
                            aId = adjacency.index((targ1,targ2))
                        elif (targ2,targ1) in adjacency:
                            aId = adjacency.index((targ2,targ1))
                    if aId is None:
                        if pretty=='clean' and pruned:
                            color=np.array([1,1,0])
                        else:
                            color=np.array([1,0,0])
                    else:
                        if pretty=='clean' and pruned:
                            color=np.array([1,0,1])
                        else:
                            color=np.array([0,1,0])
                        hits[aId]=True
                    #if pruned:
                    #    color = color*0.7
                    cv2.line(image,(x1,y1),(x2,y2),color.tolist(),lineWidth)
                    #color=color/3
                    #x = int((x1+x2)/2)
                    #y = int((y1+y2)/2)
                    #if pruned:
                    #    cv2.putText(image,'[{:.2}]'.format(score),(x,y), cv2.FONT_HERSHEY_PLAIN, 0.6,color.tolist(),1)
                    #else:
                    #    cv2.putText(image,'{:.2}'.format(score),(x,y), cv2.FONT_HERSHEY_PLAIN,1.1,color.tolist(),1)
                else:
                    shade = (relPred[i].item()-draw_rel_thresh)/(1-draw_rel_thresh)

                    #print('draw {} {} {} {} '.format(x1,y1,x2,y2))
                    cv2.line(image,(x1,y1),(x2,y2),(0,shade,0),lineWidth)
                numrelpred+=1
        if pretty and pretty!="light" and pretty!="clean":
            for i in range(len(relCand)):
                #print('{},{} : {}'.format(relCand[i][0],relCand[i][1],relPred[i]))
                if relPred[i]>-1:
                    score = (relPred[i]+1)/2
                    pruned=False
                else:
                    score = (relPred[i]+2+1)/2
                    pruned=True
                if relPred[i]>draw_rel_thresh or (pretty and score>draw_rel_thresh):
                    ind1 = relCand[i][0]
                    ind2 = relCand[i][1]
                    x1 = round(bbs[ind1,1])
                    y1 = round(bbs[ind1,2])
                    x2 = round(bbs[ind2,1])
                    y2 = round(bbs[ind2,2])

                    targ1 = bbAlignment[ind1].item()
                    targ2 = bbAlignment[ind2].item()
                    aId=None
                    if bbFullHit[ind1] and bbFullHit[ind2]:
                        if (targ1,targ2) in adjacency:
                            aId = adjacency.index((targ1,targ2))
                        elif (targ2,targ1) in adjacency:
                            aId = adjacency.index((targ2,targ1))
                    if aId is None:
                        color=np.array([1,0,0])
                    else:
                        color=np.array([0,1,0])
                    color=color/2
                    x = int((x1+x2)/2)
                    y = int((y1+y2)/2)
                    if pruned:
                        cv2.putText(image,'[{:.2}]'.format(score),(x,y), cv2.FONT_HERSHEY_PLAIN, 0.6,color.tolist(),1)
                    else:
                        cv2.putText(image,'{:.2}'.format(score),(x,y), cv2.FONT_HERSHEY_PLAIN,1.1,color.tolist(),1)
        #print('number of pred rels: {}'.format(numrelpred))
        #Draw GT pairings
        if not pretty:
            gtcolor=(0.25,0,0.25)
            wth=3
        else:
            #gtcolor=(1,0,0.6)
            gtcolor=(1,0.6,0)
            wth=2
        for aId,(i,j) in enumerate(adjacency):
            if not pretty or not hits[aId]:
                x1 = round(targetBoxes[0,i,0].item())
                y1 = round(targetBoxes[0,i,1].item())
                x2 = round(targetBoxes[0,j,0].item())
                y2 = round(targetBoxes[0,j,1].item())
                cv2.line(image,(x1,y1),(x2,y2),gtcolor,wth)

        #Draw alginment between gt and pred bbs
        if not pretty:
            for predI in range(bbs.shape[0]):
                targI=bbAlignment[predI].item()
                x1 = int(round(bbs[predI,1]))
                y1 = int(round(bbs[predI,2]))
                if targI>0:

                    x2 = round(targetBoxes[0,targI,0].item())
                    y2 = round(targetBoxes[0,targI,1].item())
                    cv2.line(image,(x1,y1),(x2,y2),(1,0,1),1)
                else:
                    #draw 'x', indicating not match
                    cv2.line(image,(x1-5,y1-5),(x1+5,y1+5),(.1,0,.1),1)
                    cv2.line(image,(x1+5,y1-5),(x1-5,y1+5),(.1,0,.1),1)



        saveName = '{}_boxes_prec:{:.2f},{:.2f}_recall:{:.2f},{:.2f}_rels_AP:{:.3f}'.format(imageName,prec_5[0],prec_5[1],recall_5[0],recall_5[1],rel_ap)
        #for j in range(metricsOut.shape[1]):
        #    saveName+='_m:{0:.3f}'.format(metricsOut[i,j])
        saveName+='.png'
        io.imsave(os.path.join(outDir,saveName),image)
        #print('saved: '+os.path.join(outDir,saveName))

    print('\n{} ap:{}\tnumMissedByDetect:{}\tmissedByHuer:{}'.format(imageName,rel_ap,numMissedByDetect,numMissedByHeur))
    retData= { 'bb_ap':[ap_5],
               'bb_recall':[recall_5],
               'bb_prec':[prec_5],
               'bb_Fm': -1,#(recall_5[0]+recall_5[1]+prec_5[0]+prec_5[1])/4,
               'nn_loss': nn_loss,
               'rel_recall':relRecall,
               'rel_precision':precision,
               'rel_Fm':2*relRecall*precision/(relRecall+precision) if relRecall+precision>0 else 0,
               'relMissedByHeur':numMissedByHeur,
               'relMissedByDetect':numMissedByDetect,
               'heurRecall': heurRecall,
               'detectRecall': detectRecall,
               **toRet

             }
    if rel_ap is not None: #none ap if no relationships
        retData['rel_AP']=rel_ap
        retData['no_targs']=0
    else:
        retData['no_targs']=1
    if model.predNN:
        retData['nn_loss_final']=nn_loss_final
        retData['nn_loss_diff']=nn_loss_final-nn_loss
        retData['nn_acc_final'] = nn_acc
    if model.detector.predNumNeighbors and not useDetections:
        retData['nn_acc_detector'] = nn_acc_d
    if model.predClass:
        retData['class_loss_final']=class_loss_final
        retData['class_loss_diff']=class_loss_final-class_loss
    return (
             retData,
             (lossThis, position_loss, conf_loss, class_loss, recall, precision)
            )
예제 #4
0
    def alignEdgePred(self, targetBoxes, adj, outputBoxes, relPred,
                      relIndexes):
        if relPred is None or targetBoxes is None:
            if targetBoxes is None:
                if relPred is not None and (relPred > self.thresh_rel).any():
                    prec = 0
                    ap = 0
                else:
                    prec = 1
                    ap = 1
                recall = 1
                targIndex = -torch.ones(outputBoxes.size(0)).int()
            elif relPred is None:
                if targetBoxes is not None:
                    recall = 0
                    ap = 0
                else:
                    recall = 1
                    ap = 1
                prec = 1
                targIndex = None

            return torch.tensor([]), torch.tensor(
                []), recall, prec, prec, ap, targIndex, torch.ones(
                    outputBoxes.size(0))
        targetBoxes = targetBoxes.cpu()
        #decide which predicted boxes belong to which target boxes
        #should this be the same as AP_?
        numClasses = 2

        if self.model.rotation:
            targIndex, fullHit = getTargIndexForPreds_dist(targetBoxes[0],
                                                           outputBoxes,
                                                           1.1,
                                                           numClasses,
                                                           hard_thresh=False)
        else:
            targIndex, fullHit = getTargIndexForPreds_iou(
                targetBoxes[0],
                outputBoxes,
                0.4,
                numClasses,
                hard_thresh=False,
                fixed=self.fixedAlign)
        #else:
        #    if self.model.rotation:
        #        targIndex, predsWithNoIntersection = getTargIndexForPreds_dist(targetBoxes[0],outputBoxes,1.1,numClasses)
        #    else:
        #        targIndex, predsWithNoIntersection = getTargIndexForPreds_iou(targetBoxes[0],outputBoxes,0.4,numClasses)

        #Create gt vector to match relPred.values()

        rels = relIndexes  #relPred._indices().cpu()
        predsAll = relPred  #relPred._values()
        sigPredsAll = torch.sigmoid(predsAll[:, -1])
        predsPos = []
        predsNeg = []
        scores = []
        matches = 0
        truePred = falsePred = badPred = 0
        for i, (n0, n1) in enumerate(rels):
            t0 = targIndex[n0].item()
            t1 = targIndex[n1].item()
            if t0 >= 0 and t1 >= 0:
                if (min(t0, t1), max(t0, t1)) in adj:
                    #if self.useBadBBPredForRelLoss!='fixed' or (fullHit[n0] and fullHit[n1]):
                    if fullHit[n0] and fullHit[n1]:
                        matches += 1
                        predsPos.append(predsAll[i])
                        scores.append((sigPredsAll[i], True))
                        if sigPredsAll[i] > self.thresh_rel:
                            truePred += 1
                    else:
                        scores.append(
                            (sigPredsAll[i], False)
                        )  #for the sake of scoring, this is a bad relationship
                else:
                    predsNeg.append(predsAll[i])
                    scores.append((sigPredsAll[i], False))
                    if sigPredsAll[i] > self.thresh_rel:
                        falsePred += 1
            else:
                #if self.useBadBBPredForRelLoss=='fixed' or (self.useBadBBPredForRelLoss and (predsWithNoIntersection[n0] or predsWithNoIntersection[n1])):
                if self.useBadBBPredForRelLoss:
                    if self.useBadBBPredForRelLoss == 'full' or np.random.rand(
                    ) < self.useBadBBPredForRelLoss:
                        predsNeg.append(predsAll[i])
                scores.append((sigPredsAll[i], False))
                if sigPredsAll[i] > self.thresh_rel:
                    badPred += 1
        #Add score 0 for instances we didn't predict
        for i in range(len(adj) - matches):
            scores.append((float('nan'), True))

        if len(predsPos) > 0:
            predsPos = torch.stack(predsPos).to(relPred.device)
        else:
            predsPos = None
        if len(predsNeg) > 0:
            predsNeg = torch.stack(predsNeg).to(relPred.device)
        else:
            predsNeg = None

        if len(adj) > 0:
            recall = truePred / len(adj)
        else:
            recall = 1
        if falsePred > 0:
            prec = truePred / (truePred + falsePred)
        else:
            prec = 1
        if falsePred + badPred > 0:
            fullPrec = truePred / (truePred + falsePred + badPred)
        else:
            fullPrec = 1
        return predsPos, predsNeg, recall, prec, fullPrec, computeAP(
            scores), targIndex, fullHit