# Now if you call rank on a vector it will output a ranking score. In # particular, the ranking score for relevant vectors should be larger than the # score for non-relevant vectors. print "ranking score for a relevant vector: ", rank(data.relevant[0]) print "ranking score for a non-relevant vector: ", rank(data.nonrelevant[0]) # The output is the following: # ranking score for a relevant vector: 0.5 # ranking score for a non-relevant vector: -0.5 # If we want an overall measure of ranking accuracy we can compute the ordering # accuracy and mean average precision values by calling test_ranking_function(). # In this case, the ordering accuracy tells us how often a non-relevant vector # was ranked ahead of a relevant vector. In this case, it returns 1 for both # metrics, indicating that the rank function outputs a perfect ranking. print dlib.test_ranking_function(rank, data) # The ranking scores are computed by taking the dot product between a learned # weight vector and a data vector. If you want to see the learned weight vector # you can display it like so: print "weights: \n", rank.weights # In this case the weights are: # 0.5 # -0.5 # In the above example, our data contains just two sets of objects. The # relevant set and non-relevant set. The trainer is attempting to find a # ranking function that gives every relevant vector a higher score than every # non-relevant vector. Sometimes what you want to do is a little more complex # than this. #
# particular, the ranking score for relevant vectors should be larger than the # score for non-relevant vectors. print(("Ranking score for a relevant vector: {}".format( rank(data.relevant[0])))) print(("Ranking score for a non-relevant vector: {}".format( rank(data.nonrelevant[0])))) # The output is the following: # ranking score for a relevant vector: 0.5 # ranking score for a non-relevant vector: -0.5 # If we want an overall measure of ranking accuracy we can compute the ordering # accuracy and mean average precision values by calling test_ranking_function(). # In this case, the ordering accuracy tells us how often a non-relevant vector # was ranked ahead of a relevant vector. In this case, it returns 1 for both # metrics, indicating that the rank function outputs a perfect ranking. print((dlib.test_ranking_function(rank, data))) # The ranking scores are computed by taking the dot product between a learned # weight vector and a data vector. If you want to see the learned weight vector # you can display it like so: print(("Weights: {}".format(rank.weights))) # In this case the weights are: # 0.5 # -0.5 # In the above example, our data contains just two sets of objects. The # relevant set and non-relevant set. The trainer is attempting to find a # ranking function that gives every relevant vector a higher score than every # non-relevant vector. Sometimes what you want to do is a little more complex # than this. #
# score for non-relevant vectors. print("Ranking score for a relevant vector: {}".format( rank(data.relevant[0]))) print("Ranking score for a non-relevant vector: {}".format( rank(data.nonrelevant[0]))) # The output is the following: # ranking score for a relevant vector: 0.5 # ranking score for a non-relevant vector: -0.5 # If we want an overall measure of ranking accuracy we can compute the ordering # accuracy and mean average precision values by calling test_ranking_function(). # In this case, the ordering accuracy tells us how often a non-relevant vector # was ranked ahead of a relevant vector. In this case, it returns 1 for both # metrics, indicating that the rank function outputs a perfect ranking. print(dlib.test_ranking_function(rank, data)) # The ranking scores are computed by taking the dot product between a learned # weight vector and a data vector. If you want to see the learned weight vector # you can display it like so: print("Weights: {}".format(rank.weights)) # In this case the weights are: # 0.5 # -0.5 # In the above example, our data contains just two sets of objects. The # relevant set and non-relevant set. The trainer is attempting to find a # ranking function that gives every relevant vector a higher score than every # non-relevant vector. Sometimes what you want to do is a little more complex # than this. #
# Now if you call rank on a vector it will output a ranking score. In # particular, the ranking score for relevant vectors should be larger than the # score for non-relevant vectors. print ("ranking score for a relevant vector: ", rank(data.relevant[0])) print ("ranking score for a non-relevant vector: ", rank(data.nonrelevant[0])) # The output is the following: # ranking score for a relevant vector: 0.5 # ranking score for a non-relevant vector: -0.5 # If we want an overall measure of ranking accuracy we can compute the ordering # accuracy and mean average precision values by calling test_ranking_function(). # In this case, the ordering accuracy tells us how often a non-relevant vector # was ranked ahead of a relevant vector. In this case, it returns 1 for both # metrics, indicating that the rank function outputs a perfect ranking. print ("test: ", dlib.test_ranking_function(rank, data)) # The ranking scores are computed by taking the dot product between a learned # weight vector and a data vector. If you want to see the learned weight vector # you can display it like so: print ("weights: \n", rank.weights) # In this case the weights are: # 0.5 # -0.5 # In the above example, our data contains just two sets of objects. The # relevant set and non-relevant set. The trainer is attempting to find a # ranking function that gives every relevant vector a higher score than every