-
Notifications
You must be signed in to change notification settings - Fork 17
/
test_torchtest.py
50 lines (41 loc) · 1.25 KB
/
test_torchtest.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
from torchtest import torchtest as tt
import tc
import torch
import torch.nn as nn
import torch.nn.functional as F
torch.manual_seed(1)
if __name__ == '__main__':
# setup test suite
tt.setup()
# define model params
hparams={
'vocab_size' : 100,
'emb_dim' : 20,
'hidden_dim' : 30,
'output_size' : 2,
'loss_fn' : F.cross_entropy,
'batch_size' : 10
}
# create model
model = tc.LstmClassifier(hparams, weights={
# not really GloVe; random samples from uniform distribution
'glove' : torch.rand(hparams['vocab_size'], hparams['emb_dim'])
})
# create a random batch
# lets say seq_len = 15
batch = [
torch.randint(0, hparams['vocab_size'], (hparams['batch_size'], 15)).long(),
torch.randint(0, hparams['output_size'], (hparams['batch_size'],)).long()
]
# run all tests
tt.test_suite(
model,
hparams['loss_fn'], # loss function
torch.optim.Adam([p for p in model.parameters() if p.requires_grad]), # optimizer
batch, # random data
non_train_vars= [ # embedding is supposed to be fixed
('embedding.weight', model.embedding.weight) # variable(s) to check for change
],
test_gpu_available=True,
device='cuda:0'
)