-
Notifications
You must be signed in to change notification settings - Fork 0
/
ddi_dataset.py
352 lines (286 loc) · 13 KB
/
ddi_dataset.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
"""
Dataloader that is used is from the paper
"Drug-Drug Adverse Effect Prediction with Graph Co-Attention" https://arxiv.org/abs/1905.00534
Github repo with the code for the ddi dataloader https://github.com/andreeadeac22/graph_coattention
This is a slightly modified version of that ddi dataloader
"""
import pickle
import numpy as np
import torch.utils.data
def create_ddi_dataloaders(opt):
opt.fold_i, opt.n_fold = map(int, opt.fold.split('/'))
data_opt = np.load(opt.input_data_path + "input_data.npy", allow_pickle=True).item()
opt.n_atom_type = data_opt.n_atom_type
opt.n_bond_type = data_opt.n_bond_type
graph_dict = data_opt.graph_dict
if "decagon" in opt.dataset:
opt.n_side_effect = data_opt.n_side_effect
side_effects = data_opt.side_effects
side_effect_idx_dict = data_opt.side_effect_idx_dict
# 'pos'/'neg' will point to a dictionary where
# each se points to a list of drug-drug pairs.
train_dataset = {'pos': {}, 'neg': {}}
test_dataset = pickle.load(open(opt.input_data_path + "folds/" + str(opt.fold_i) + "fold.npy", "rb"))
if opt.fold_i == 1:
valid_fold = 2
else:
valid_fold = 1
valid_dataset = pickle.load(open(opt.input_data_path + "folds/" + str(valid_fold) + "fold.npy", "rb"))
for i in range(valid_fold + 1, opt.n_fold + 1):
if i != opt.fold_i:
dataset = pickle.load(open(opt.input_data_path + "folds/" + str(i) + "fold.npy", "rb"))
train_dataset['pos'] = combine(train_dataset['pos'], dataset['pos'])
train_dataset['neg'] = combine(train_dataset['neg'], dataset['neg'])
assert data_opt.n_side_effect == len(side_effects)
dataloaders = prepare_ddi_dataloaders(opt, train_dataset, valid_dataset, graph_dict, side_effect_idx_dict)
return dataloaders
def combine(d1, d2):
for (k, v) in d2.items():
if k not in d1:
d1[k] = v
else:
d1[k].extend(v)
return d1
def collate_fun(x):
# Has to be a separate function because pickle has a problem work with lambda functions
return ddi_collate_batch(x, return_label=True)
def prepare_ddi_dataloaders(opt, train_dataset, valid_dataset, graph_dict, side_effect_idx_dict):
train_loader = torch.utils.data.DataLoader(
PolypharmacyDataset(
drug_structure_dict=graph_dict,
se_idx_dict=side_effect_idx_dict,
se_pos_dps=train_dataset['pos'],
# TODO: inspect why I'm not just fetching opt.train_dataset['neg']
negative_sampling=True,
negative_sample_ratio=opt.train_neg_pos_ratio,
paired_input=True,
n_max_batch_se=10),
num_workers=4,
batch_size=opt.batch_size,
collate_fn=ddi_collate_paired_batch,
shuffle=True)
valid_loader = torch.utils.data.DataLoader(
PolypharmacyDataset(
drug_structure_dict=graph_dict,
se_idx_dict=side_effect_idx_dict,
se_pos_dps=valid_dataset['pos'],
se_neg_dps=valid_dataset['neg'],
n_max_batch_se=1),
num_workers=4,
batch_size=opt.batch_size,
collate_fn=collate_fun)
return train_loader, valid_loader
def prepare_ddi_testset_dataloader(positive_set, negative_set, train_opt, batch_size):
test_loader = torch.utils.data.DataLoader(
PolypharmacyDataset(
drug_structure_dict=train_opt.graph_dict,
se_idx_dict=train_opt.side_effect_idx_dict,
se_pos_dps=positive_set,
se_neg_dps=negative_set,
n_max_batch_se=1),
num_workers=2,
batch_size=batch_size,
collate_fn=collate_fun)
return test_loader
def ddi_collate_paired_batch(paired_batch):
pos_batch = []
neg_batch = []
seg_pos_neg = []
pos_se_i = 0
for ddi_pair in paired_batch:
pos_ddi, neg_ddis = ddi_pair
pos_batch += [pos_ddi] # flatten negative instances
neg_batch += neg_ddis
*_, pos_ses, _ = pos_ddi
for _ in range(len(pos_ses)):
seg_pos_neg += [pos_se_i] * len(neg_ddis)
pos_se_i += 1
seg_pos_neg = torch.LongTensor(np.array(seg_pos_neg))
pos_batch = ddi_collate_batch(pos_batch, return_label=False)
neg_batch = ddi_collate_batch(neg_batch, return_label=False)
return pos_batch, neg_batch, seg_pos_neg
def ddi_collate_batch(batch, return_label=True):
drug1, drug2, se_idx_lists, label = list(zip(*batch))
ddi_idxs1, ddi_idxs2 = collate_drug_pairs(drug1, drug2)
drug1 = (*collate_drugs(drug1), *ddi_idxs1)
drug2 = (*collate_drugs(drug2), *ddi_idxs2)
se_idx, se_seg = collate_side_effect(se_idx_lists)
if return_label:
label = np.hstack([
[label_i] * len(ses_i) for ses_i, label_i in zip(se_idx_lists, label)])
return (*drug1, *drug2, se_idx, se_seg, label)
else:
return (*drug1, *drug2, se_idx, se_seg)
def collate_drug_pairs(drugs1, drugs2):
n_atom1 = [d['n_atom'] for d in drugs1]
n_atom2 = [d['n_atom'] for d in drugs2]
c_atom1 = [sum(n_atom1[:k]) for k in range(len(n_atom1))]
c_atom2 = [sum(n_atom2[:k]) for k in range(len(n_atom2))]
ddi_seg_i1, ddi_seg_i2, ddi_idx_j1, ddi_idx_j2 = zip(*[
(i1 + c1, i2 + c2, i2, i1)
for l1, l2, c1, c2 in zip(n_atom1, n_atom2, c_atom1, c_atom2)
for i1 in range(l1) for i2 in range(l2)])
ddi_seg_i1 = torch.LongTensor(ddi_seg_i1)
ddi_idx_j1 = torch.LongTensor(ddi_idx_j1)
ddi_seg_i2 = torch.LongTensor(ddi_seg_i2)
ddi_idx_j2 = torch.LongTensor(ddi_idx_j2)
return (ddi_seg_i1, ddi_idx_j1), (ddi_seg_i2, ddi_idx_j2)
def collate_side_effect(se_idx_lists):
se_idx = torch.LongTensor(np.hstack(se_idx_lists).astype(np.int64))
se_seg = np.hstack([[i] * len(ses_i) for i, ses_i in enumerate(se_idx_lists)])
se_seg = torch.LongTensor(se_seg)
return se_idx, se_seg
def collate_drugs(drugs):
c_atoms = [sum(d['n_atom'] for d in drugs[:k]) for k in range(len(drugs))]
atom_feat = torch.FloatTensor(np.vstack([d['atom_feat'] for d in drugs]))
atom_type = torch.LongTensor(np.hstack([d['atom_type'] for d in drugs]))
bond_type = torch.LongTensor(np.hstack([d['bond_type'] for d in drugs]))
bond_seg_i = torch.LongTensor(np.hstack([
np.array(d['bond_seg_i']) + c for d, c in zip(drugs, c_atoms)]))
bond_idx_j = torch.LongTensor(np.hstack([
np.array(d['bond_idx_j']) + c for d, c in zip(drugs, c_atoms)]))
batch_seg_m = torch.LongTensor(np.hstack([
[k] * d['n_atom'] for k, d in enumerate(drugs)]))
return batch_seg_m, atom_type, atom_feat, bond_type, bond_seg_i, bond_idx_j
def collate_batch(batch):
'''
Creates a batch of same size graphs by zero-padding node features and adjacency matrices up to
the maximum number of nodes in the CURRENT batch rather than in the entire dataset.
Graphs in the batches are usually much smaller than the largest graph in the dataset, so this method is fast.
:param batch: batch in the PyTorch Geometric format or [node_features*batch_size, A*batch_size, label*batch_size]
:return: [node_features, A, graph_support, N_nodes, label]
'''
B = len(batch)
N_nodes = [len(batch[b][1]) for b in range(B)]
C = batch[0][0].shape[1]
N_nodes_max = int(np.max(N_nodes))
graph_support = torch.zeros(B, N_nodes_max)
A = torch.zeros(B, N_nodes_max, N_nodes_max)
x = torch.zeros(B, N_nodes_max, C)
for b in range(B):
x[b, :N_nodes[b]] = batch[b][0]
A[b, :N_nodes[b], :N_nodes[b]] = batch[b][1]
graph_support[b][:N_nodes[b]] = 1 # mask with values of 0 for dummy (zero padded) nodes, otherwise 1
N_nodes = torch.from_numpy(np.array(N_nodes)).long()
labels = torch.from_numpy([batch[b][2] for b in range(B)]).long()
return [x, A, graph_support, N_nodes, labels]
class PolypharmacyDataset(torch.utils.data.Dataset):
def __init__(
self,
drug_structure_dict,
se_idx_dict,
se_pos_dps=None,
se_neg_dps=None,
negative_sampling=False,
negative_sample_ratio=1,
n_max_batch_se=1,
paired_input=False):
assert se_pos_dps
assert se_neg_dps or negative_sampling
assert not (se_neg_dps and negative_sampling)
assert type(negative_sample_ratio) is int and negative_sample_ratio >= 1
self.negative_sampling = negative_sampling
self.paired_input = paired_input
self.se_idx_dict = se_idx_dict
"""
print("Se idx dict ")
with open("se_idx_dict.txt", "w") as filename:
for se in se_idx_dict:
print(se, se_idx_dict[se], file=filename)
"""
self.drug_structure_dict = drug_structure_dict
"""
print("Drug struct dict ")
with open("drug_struct_dict.txt", "w") as filename1:
for drug in drug_structure_dict:
print(drug, drug_structure_dict[se], file=filename1)
"""
self.drug_idx_list = list(drug_structure_dict.keys())
self.n_inst_batch_se = n_max_batch_se
self.n_corrupt = negative_sample_ratio
self.pos_ddis = self.collate_given_positive_set(se_pos_dps, se_idx_dict, negative_sampling)
self.neg_ddis = self.collate_given_negative_set(se_neg_dps, se_idx_dict)
self.feeding_insts = None
self.prepare_feeding_insts()
def collate_given_negative_set(self, se_neg_dps, se_idx_dict):
''' From se -> dps mapping to dp -> ses mapping '''
neg_ddis = {}
if se_neg_dps:
for se, dps in se_neg_dps.items():
for dp in dps:
if dp not in neg_ddis:
neg_ddis[dp] = []
neg_ddis[dp] += [se_idx_dict[se]]
return neg_ddis
def mapping_transpose(self, se_dps_dict):
''' From `se -> dps` mapping to `dp -> ses` mapping '''
dp_ses_dict = {}
for se, dps in se_dps_dict.items():
for dp in dps:
if dp not in dp_ses_dict:
dp_ses_dict[dp] = []
dp_ses_dict[dp] += [se]
return dp_ses_dict
def collate_given_positive_set(self, se_pos_dps, se_idx_dict, negative_sampling):
''' From se -> dps mapping to dp -> ses mapping '''
pos_ddis = {}
flip_drug_pair = lambda dp: tuple(reversed(dp))
for se, dps in se_pos_dps.items():
if negative_sampling:
dps = dps + list(map(flip_drug_pair, dps))
for dp in dps:
if dp not in pos_ddis:
pos_ddis[dp] = []
pos_ddis[dp] += [se_idx_dict[se]]
return pos_ddis
def prepare_feeding_insts(self):
def collect_with_proper_size_se(ddis, inst_label):
""" To reduce the duplicated computing on same graph pair for different labels. """
# split number of ses in k * batch(ses) to account
# for d-d with many vs few ses
ddis = dict(ddis)
inst_list = []
for dp, ses in ddis.items():
n_se_batch = int(np.ceil(len(ses) / self.n_inst_batch_se))
for i in range(n_se_batch):
start = i * self.n_inst_batch_se
end = (i + 1) * self.n_inst_batch_se
inst_list += [(*dp, ses[start: end], inst_label)]
return inst_list
pos_insts = collect_with_proper_size_se(self.pos_ddis, inst_label=True)
if self.negative_sampling:
feeding_insts = []
rand_drugs = list(np.random.choice(
self.drug_idx_list, size=self.n_corrupt * len(pos_insts)))
if not self.paired_input:
feeding_insts = pos_insts
for pos_inst in pos_insts:
d1, _, ses, _ = pos_inst
corr_insts = [(d1, rand_drugs.pop(), ses, False) for _ in range(self.n_corrupt)]
if self.paired_input:
paired_feed = (pos_inst, corr_insts)
feeding_insts += [paired_feed]
else:
feeding_insts += corr_insts
else:
neg_insts = collect_with_proper_size_se(self.neg_ddis, inst_label=False)
feeding_insts = pos_insts + neg_insts
self.feeding_insts = feeding_insts
def __len__(self):
return len(self.feeding_insts)
def __getitem__(self, idx):
instance = self.feeding_insts[idx]
# drug lookup
if self.paired_input:
pos_inst, neg_insts = instance
pos_inst = self.drug_structure_lookup(pos_inst)
neg_insts = list(map(self.drug_structure_lookup, neg_insts))
return pos_inst, neg_insts
else:
instance = self.drug_structure_lookup(instance)
return instance
def drug_structure_lookup(self, instance):
drug_idx1, drug_idx2, se_idx_lists, label = instance
drug1 = self.drug_structure_dict[drug_idx1]
drug2 = self.drug_structure_dict[drug_idx2]
return drug1, drug2, se_idx_lists, label