From 461b57df7ac95c87a92d472d4f3a9a3d385b5252 Mon Sep 17 00:00:00 2001 From: Phil Wang Date: Tue, 31 Aug 2021 10:31:49 -0700 Subject: [PATCH] fix small bug --- setup.py | 2 +- tab_transformer_pytorch/tab_transformer_pytorch.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/setup.py b/setup.py index 0b5f0c9..b090565 100644 --- a/setup.py +++ b/setup.py @@ -3,7 +3,7 @@ setup( name = 'tab-transformer-pytorch', packages = find_packages(), - version = '0.1.2', + version = '0.1.4', license='MIT', description = 'Tab Transformer - Pytorch', author = 'Phil Wang', diff --git a/tab_transformer_pytorch/tab_transformer_pytorch.py b/tab_transformer_pytorch/tab_transformer_pytorch.py index 4a5e7e3..7cd6d22 100644 --- a/tab_transformer_pytorch/tab_transformer_pytorch.py +++ b/tab_transformer_pytorch/tab_transformer_pytorch.py @@ -112,7 +112,7 @@ def __init__(self, dims, act = None): dims_pairs = list(zip(dims[:-1], dims[1:])) layers = [] for ind, (dim_in, dim_out) in enumerate(dims_pairs): - is_last = ind >= (len(dims) - 1) + is_last = ind >= (len(dims_pairs) - 1) linear = nn.Linear(dim_in, dim_out) layers.append(linear)