forked from ManchesterBioinference/EP_Bayes
/
classifiers_clean.py
289 lines (187 loc) · 14.9 KB
/
classifiers_clean.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
import config_variables
import itertools
from prepare_interactions_clean import un_string
import chrom_specific_negative_interactions
chroms_to_infer = config_variables.chroms_to_infer
dataset_names_option = config_variables.dataset_names_option
np = config_variables.np
dict_chrom_pro_survived = config_variables.dict_chrom_pro_survived
classificator_elements = config_variables.classificator_elements
mode = config_variables.mode
filter_value = config_variables.filter_value
alternative_classificator = config_variables.alternative_classificator
import time
def classifier(array_of_merged_probabilities_true, array_of_merged_probabilities_false, K):
prob = (1 + (K-1)*np.exp((np.log(array_of_merged_probabilities_false)-np.log(array_of_merged_probabilities_true)).sum(1)))**-1
return prob
classifiers_elements = {}
#-----------------------------------------------------------------
#initialises global dictionary
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
classifiers_elements[classification_of_interactions] = {}
for probability_of_being_positive_or_negative in ["probabilities_of_being_positive_interactions", "probabilities_of_being_negative_interactions"]:
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative] = {}
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative]["column_stacked_probabilities"] = {}
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative]["posterior_values"] = {}
for chrom_ in chroms_to_infer:
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative]["column_stacked_probabilities"][chrom_] = []
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative]["posterior_values"][chrom_] = []
#fill in the global dictionary:
#-----------------------------------------------------------------
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
for probability_of_being_positive_or_negative in ["probabilities_of_being_positive_interactions", "probabilities_of_being_negative_interactions"]:
for chrom_ in chroms_to_infer:
column_stacked_probabilities = classificator_elements[filter_value][mode][classification_of_interactions]["distance"][probability_of_being_positive_or_negative]["posterior_component_values"][chrom_]
for data_set_name in dataset_names_option:
column_stacked_probabilities = np.column_stack((column_stacked_probabilities, classificator_elements[filter_value][mode][classification_of_interactions]["correlation"][probability_of_being_positive_or_negative][data_set_name]["posterior_component_values"][chrom_]))
classifiers_elements[classification_of_interactions][probability_of_being_positive_or_negative]["column_stacked_probabilities"][chrom_] = column_stacked_probabilities
#calculates the posterior with different options
#-----------------------------------------------------------------
def interactions_extractor(chrom):
true_pro_enh_indexes = un_string(config_variables.chr_interactions_dict_pro_enh[chrom])
prom_enh_false_interactions = chrom_specific_negative_interactions.chrom_specific_negative_interactions(chrom, mode)
enh_coordinates, pro_coordinates, indexes_p, indexes_e, total_p, total_e = chrom_specific_negative_interactions.initialise_variables(chrom)
true_pro_enh_indexes[:,0] = true_pro_enh_indexes[:,0] - total_p
true_pro_enh_indexes[:,1] = true_pro_enh_indexes[:,1] - total_e
return true_pro_enh_indexes, prom_enh_false_interactions
def MOG_classifier(option_correl, total_posterior = False):
import os
number_of_samples = config_variables.number_of_samples
def loads_MoG_results(chrom):
comb = "_".join([config_variables.dict_option[el] for el in option_correl])
kappa_0, mu_0, alpha_0, Beta_0 = config_variables.kappa_0, config_variables.mu_0, config_variables.alpha_0, config_variables.Beta_0
name = 'cluster_trace_of_c_distance_{0}_{1}_{2}_{3}_{4}_{5}_{6}'.format(kappa_0, mu_0, alpha_0, Beta_0, chrom, comb, number_of_samples)
name = os.getcwd() + "/MOG_results_/" + name
import iter_loadtxt
_c_trace_raw = iter_loadtxt.iter_loadtxt(name, ",", dtype = int) # saves memory
num_of_promoters = len(config_variables.dict_chrom_pro_survived[chrom])
promoters_fixed_labels = np.zeros((len(_c_trace_raw), num_of_promoters),dtype = int)
promoters_fixed_labels[:] = np.arange(num_of_promoters, dtype = int)
_c_trace_distance = np.c_[promoters_fixed_labels, _c_trace_raw]
return _c_trace_distance
def cluster_estimator_similarity(_c_trace):
from multiprocessing import Pool
pool = Pool(processes = 4)
pack = 100
dim = _c_trace.shape
incr = int(dim[0]/pack)
a = [_c_trace[i*incr:(i+1)*incr] for i in np.arange(pack-1)] + [_c_trace[(pack-1)*incr:]]
import pararell_methods
start = time.time()
total_matrix_2 = sum(pool.imap_unordered(pararell_methods.pararell_calc_ne, a))
pool.close()
pool.join()
end = time.time()
print end-start
return total_matrix
def standard_size_converter(total_matrix, chrom):
indexes_p, indexes_e, total_p, total_e = chrom_specific_negative_interactions.initialise_variables(chrom)[2:]
length_chr = len(indexes_p) + len(indexes_e)
interaction_matrix = np.zeros((length_chr, length_chr), bool)
chrom_pro_not_survived = config_variables.dict_chrom_pro_not_survived[chrom]
chrom_enh_not_survived = config_variables.dict_chrom_enh_not_survived[chrom]
dict_chrom_proximal = config_variables.dict_chrom_proximal
if len(chrom_pro_not_survived): interaction_matrix[chrom_pro_not_survived - total_p, :] = True
if len(chrom_enh_not_survived): interaction_matrix[:, len(indexes_p) + chrom_enh_not_survived - total_e] = True # gets rid of filtered out enhancers which could be causing nans due to their correlations
if config_variables.distant_enh_only and len(dict_chrom_proximal[chrom]): interaction_matrix[:, len(indexes_p) + dict_chrom_proximal[chrom] - total_e] = True
interaction_matrix = np.invert(interaction_matrix + interaction_matrix.T)
temp_expanded_total_matrix = np.zeros(length_chr * length_chr, int)
temp_expanded_total_matrix[np.ravel(interaction_matrix)] = np.ravel(total_matrix)
expanded_total_matrix = temp_expanded_total_matrix.reshape(length_chr, length_chr)
return expanded_total_matrix
posterior_of_option = {}
chrom_posterior = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]: posterior_of_option[classification_of_interactions] = {}
for chrom_ in chroms_to_infer:
_c_trace_distance = loads_MoG_results(chrom_)
total_matrix = cluster_estimator_similarity(_c_trace_distance)
total_matrix = standard_size_converter(total_matrix, chrom_)
number_of_iterations = float(len(_c_trace_distance))
total_matrix = total_matrix / number_of_iterations
true_pro_enh_indexes, prom_enh_false_interactions = interactions_extractor(chrom_)
indexes_p, indexes_e, total_p, total_e = chrom_specific_negative_interactions.initialise_variables(chrom_)[2:]
chrom_posterior["positive_interactions"] = total_matrix[true_pro_enh_indexes[:,0], true_pro_enh_indexes[:,1] + len(indexes_p)]
chrom_posterior["negative_interactions"] = total_matrix[prom_enh_false_interactions[:,0], prom_enh_false_interactions[:,1] + len(indexes_p)]
for classification_of_interactions in ["positive_interactions", "negative_interactions"]: posterior_of_option[classification_of_interactions][chrom_] = chrom_posterior[classification_of_interactions]
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
if total_posterior: posterior_of_option[classification_of_interactions] = list(itertools.chain.from_iterable([posterior_of_option[classification_of_interactions][chrom__] for chrom__ in chroms_to_infer]))
return posterior_of_option["positive_interactions"], posterior_of_option["negative_interactions"]
def posterior_producer_non_domain(option_dist, option_correl, total_posterior = False):
option_2 = option_dist + ( np.array(option_correl) + 1).tolist()
print option_2
posterior_of_option = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
posterior_of_option[classification_of_interactions] = {}
for chrom_ in chroms_to_infer:
probs_true = classifiers_elements[classification_of_interactions]["probabilities_of_being_positive_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
probs_false = classifiers_elements[classification_of_interactions]["probabilities_of_being_negative_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
K = len(dict_chrom_pro_survived[chrom_])
chrom_posterior = classifier(probs_true, probs_false, K)
posterior_of_option[classification_of_interactions][chrom_] = chrom_posterior
if total_posterior: posterior_of_option[classification_of_interactions] = list(itertools.chain.from_iterable([posterior_of_option[classification_of_interactions][chrom_] for chrom_ in chroms_to_infer]))
return posterior_of_option["positive_interactions"], posterior_of_option["negative_interactions"]
def posterior_producer_domain(option_dist, option_correl, total_posterior = False):
import domain_allocator_clean
option_2 = option_dist + ( np.array(option_correl) + 1).tolist()
print option_2
posterior_of_option = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]: posterior_of_option[classification_of_interactions] = {}
one_totals = 0
total_of_non_empty = 0
for chrom_ in chroms_to_infer:
allocations = {}
allocations["positive_interactions"], allocations["negative_interactions"], filtered_promoters_in_sub_domains, size_domains = domain_allocator_clean.allocator(chrom_)
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
probs_true = classifiers_elements[classification_of_interactions]["probabilities_of_being_positive_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
probs_false = classifiers_elements[classification_of_interactions]["probabilities_of_being_negative_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
chrom_posterior = np.zeros(len(probs_true)) # will store classifiers result for each positive or negative interaction respecively
for sub_domain in np.unique(allocations[classification_of_interactions]):
K = filtered_promoters_in_sub_domains[sub_domain]
if K == 1: one_totals +=1
if K: total_of_non_empty += 1
interactions_of_subdomain = allocations[classification_of_interactions] == sub_domain
if sum(interactions_of_subdomain) == 0: continue
print K
'make sure that the oprobabilities of the interactions are in the right corresponding entries of the vector so that interaction <--> correct probability'
chrom_posterior[interactions_of_subdomain] = classifier(probs_true[interactions_of_subdomain], probs_false[interactions_of_subdomain], K)
posterior_of_option[classification_of_interactions][chrom_] = chrom_posterior
print "number of single promoter domains:", one_totals, total_of_non_empty, one_totals/float(total_of_non_empty)
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
if total_posterior: posterior_of_option[classification_of_interactions] = list(itertools.chain.from_iterable([posterior_of_option[classification_of_interactions][chrom_] for chrom_ in chroms_to_infer]))
return posterior_of_option["positive_interactions"], posterior_of_option["negative_interactions"]
def classifier_alternative(probs_true, probs_false, mask_of_existing_interactions_of_enhancer):
single_k_model_ratios = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
mask = mask_of_existing_interactions_of_enhancer[classification_of_interactions]
single_k_model_ratios[classification_of_interactions] = (probs_true[classification_of_interactions][mask] / probs_false[classification_of_interactions][mask]).prod(1)
denominator = np.sum(np.r_[single_k_model_ratios["positive_interactions"], single_k_model_ratios["negative_interactions"]])
return single_k_model_ratios["positive_interactions"]/denominator, single_k_model_ratios["negative_interactions"]/denominator
def posterior_producer_one_promoter_model(option_dist, option_correl, total_posterior = False):
option_2 = option_dist + ( np.array(option_correl) + 1).tolist()
print option_2
posterior_of_option = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]: posterior_of_option[classification_of_interactions] = {}
for chrom_ in chroms_to_infer:
interactions = {}
interactions["positive_interactions"], interactions["negative_interactions"] = interactions_extractor(chrom_)
probs_true = {}
probs_false = {}
chrom_posterior = {}
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
probs_true[classification_of_interactions] = classifiers_elements[classification_of_interactions]["probabilities_of_being_positive_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
probs_false[classification_of_interactions] = classifiers_elements[classification_of_interactions]["probabilities_of_being_negative_interactions"]["column_stacked_probabilities"][chrom_][:, option_2]
chrom_posterior[classification_of_interactions] = np.zeros(len(interactions[classification_of_interactions]))
mask_of_existing_interactions_of_enhancer = {}
enhancers_to_infer = np.unique(np.r_[interactions["positive_interactions"][:,1], interactions["negative_interactions"][:,1]])
for enhancer in enhancers_to_infer:
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
mask_of_existing_interactions_of_enhancer[classification_of_interactions] = interactions[classification_of_interactions][:,1] == enhancer
chrom_posterior["positive_interactions"][mask_of_existing_interactions_of_enhancer["positive_interactions"]], chrom_posterior["negative_interactions"][mask_of_existing_interactions_of_enhancer["negative_interactions"]] = classifier_alternative(probs_true, probs_false, mask_of_existing_interactions_of_enhancer)
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
posterior_of_option[classification_of_interactions][chrom_] = chrom_posterior[classification_of_interactions]
for classification_of_interactions in ["positive_interactions", "negative_interactions"]:
if total_posterior: posterior_of_option[classification_of_interactions] = list(itertools.chain.from_iterable([posterior_of_option[classification_of_interactions][chrom_] for chrom_ in chroms_to_infer]))
return posterior_of_option["positive_interactions"], posterior_of_option["negative_interactions"]
if alternative_classificator: posterior_producer = posterior_producer_one_promoter_model
elif config_variables.domain: posterior_producer = posterior_producer_domain
else: posterior_producer = posterior_producer_non_domain