-
Notifications
You must be signed in to change notification settings - Fork 30
/
sampler.py
85 lines (71 loc) · 3.38 KB
/
sampler.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
import argparse
import yaml
import torch
import time
import numpy as np
import pandas as pd
from tqdm import tqdm
from sampler_core import ParallelSampler, TemporalGraphBlock
class NegLinkSampler:
def __init__(self, num_nodes):
self.num_nodes = num_nodes
def sample(self, n):
return np.random.randint(self.num_nodes, size=n)
class NegLinkInductiveSampler:
def __init__(self, nodes):
self.nodes = list(nodes)
def sample(self, n):
return np.random.choice(self.nodes, size=n)
if __name__ == '__main__':
parser=argparse.ArgumentParser()
parser.add_argument('--data', type=str, help='dataset name')
parser.add_argument('--config', type=str, help='path to config file')
parser.add_argument('--batch_size', type=int, default=600, help='path to config file')
parser.add_argument('--num_thread', type=int, default=64, help='number of thread')
args=parser.parse_args()
df = pd.read_csv('DATA/{}/edges.csv'.format(args.data))
g = np.load('DATA/{}/ext_full.npz'.format(args.data))
sample_config = yaml.safe_load(open(args.config, 'r'))['sampling'][0]
sampler = ParallelSampler(g['indptr'], g['indices'], g['eid'], g['ts'].astype(np.float32),
args.num_thread, 1, sample_config['layer'], sample_config['neighbor'],
sample_config['strategy']=='recent', sample_config['prop_time'],
sample_config['history'], float(sample_config['duration']))
num_nodes = max(int(df['src'].max()), int(df['dst'].max()))
neg_link_sampler = NegLinkSampler(num_nodes)
tot_time = 0
ptr_time = 0
coo_time = 0
sea_time = 0
sam_time = 0
uni_time = 0
total_nodes = 0
unique_nodes = 0
for _, rows in tqdm(df.groupby(df.index // args.batch_size), total=len(df) // args.batch_size):
root_nodes = np.concatenate([rows.src.values, rows.dst.values, neg_link_sampler.sample(len(rows))]).astype(np.int32)
ts = np.concatenate([rows.time.values, rows.time.values, rows.time.values]).astype(np.float32)
sampler.sample(root_nodes, ts)
ret = sampler.get_ret()
tot_time += ret[0].tot_time()
ptr_time += ret[0].ptr_time()
coo_time += ret[0].coo_time()
sea_time += ret[0].search_time()
sam_time += ret[0].sample_time()
# for i in range(sample_config['history']):
# total_nodes += ret[i].dim_in() - ret[i].dim_out()
# unique_nodes += ret[i].dim_in() - ret[i].dim_out()
# if ret[i].dim_in() > ret[i].dim_out():
# ts = torch.from_numpy(ret[i].ts()[ret[i].dim_out():])
# nid = torch.from_numpy(ret[i].nodes()[ret[i].dim_out():]).float()
# nts = torch.stack([ts,nid],dim=1).cuda()
# uni_t_s = time.time()
# unts, idx = torch.unique(nts, dim=0, return_inverse=True)
# uni_time += time.time() - uni_t_s
# total_nodes += idx.shape[0]
# unique_nodes += unts.shape[0]
print('total time : {:.4f}'.format(tot_time))
print('pointer time: {:.4f}'.format(ptr_time))
print('coo time : {:.4f}'.format(coo_time))
print('search time : {:.4f}'.format(sea_time))
print('sample time : {:.4f}'.format(sam_time))
# print('unique time : {:.4f}'.format(uni_time))
# print('unique per : {:.4f}'.format(unique_nodes / total_nodes))