-
Notifications
You must be signed in to change notification settings - Fork 2
/
measurement_dom_chunk_no_injection.py
91 lines (80 loc) · 3.06 KB
/
measurement_dom_chunk_no_injection.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
from automation import TaskManager, CommandSequence
from automation.Errors import CommandExecutionError
import crawl_utils
import time
import os
# The list of sites that we wish to crawl
NUM_BROWSERS = 15
NUM_BATCH = 5000
DATA_DIR = os.path.join(os.path.dirname(__file__), 'data')
manager_params, browser_params = TaskManager.load_default_params(NUM_BROWSERS)
date_prefix = 'XXX' # Updated by deployment script
prefix = date_prefix + '_dom_chunk_no_injection'
manager_params['database_name'] = prefix + '.sqlite'
manager_params['data_directory'] = '~/' + prefix
manager_params['log_directory'] = '~/' + prefix
# Read the site list
sites = list()
with open(os.path.join(os.path.dirname(__file__),
'data', 'sites_to_recrawl_DOM_chunk.txt')) as f:
for line in f:
if line.strip() == '':
continue
sites.append(tuple(line.strip().split(',', 2)))
TOTAL_NUM_SITES = len(sites)
for i in xrange(NUM_BROWSERS):
browser_params[i]['headless'] = True
browser_params[i]['js_instrument'] = True
browser_params[i]['cookie_instrument'] = True
browser_params[i]['http_instrument'] = True
browser_params[i]['save_javascript'] = True
browser_params[i]['spoof_identity']['enabled'] = False
browser_params[i]['spoof_identity']['dom_chunk'] = False
start_time = time.time()
# Manage control files
if not os.path.isdir(os.path.expanduser('~/.openwpm/')):
os.mkdir(os.path.expanduser('~/.openwpm/'))
if os.path.isfile(os.path.expanduser('~/.openwpm/reboot')):
os.remove(os.path.expanduser('~/.openwpm/reboot'))
if os.path.isfile(os.path.expanduser('~/.openwpm/current_site_index')):
with open(os.path.expanduser('~/.openwpm/current_site_index'), 'r') as f:
start_index = int(f.read()) + 1
end_index = start_index + NUM_BATCH
else:
start_index = 0
end_index = NUM_BATCH + 1
manager = TaskManager.TaskManager(manager_params, browser_params,
process_watchdog=True)
current_index = 0
for i in range(start_index, end_index):
current_index = i
if current_index >= TOTAL_NUM_SITES:
break
try:
try:
first_party, rank, url = sites[i]
except ValueError:
continue
cs = CommandSequence.CommandSequence(
url,
site_rank=rank,
first_party=first_party,
reset=True)
cs.get(sleep=10, timeout=120)
manager.execute_command_sequence(cs)
with open(os.path.expanduser('~/.openwpm/current_site_index'),
'w') as f:
f.write(str(i))
except CommandExecutionError:
with open(os.path.expanduser('~/.openwpm/reboot'), 'w') as f:
f.write(str(1))
break
print "CLOSING TaskManager after batch"
manager.close()
crawl_utils.clear_tmp_folder()
# Remove index file if we are done
if current_index >= TOTAL_NUM_SITES:
os.remove(os.path.expanduser('~/.openwpm/current_site_index'))
with open(os.path.expanduser('~/.openwpm/crawl_done'), 'w') as f:
f.write(str(1))
print "Total time: " + str(time.time() - start_time)