Upped lr and dropout

This commit is contained in:
Dominik Moritz Roth 2021-09-21 09:17:01 +02:00
parent f4444f3a9e
commit 627bf370bc

View File

@ -13,7 +13,7 @@ class Model(nn.Module):
input_size=8, input_size=8,
hidden_size=16, hidden_size=16,
num_layers=3, num_layers=3,
dropout=0.05, dropout=0.1,
) )
self.fc = nn.Linear(16, 1) self.fc = nn.Linear(16, 1)
self.out = nn.Sigmoid() self.out = nn.Sigmoid()
@ -32,7 +32,7 @@ def train(model, seq_len=16*64):
model.train() model.train()
criterion = nn.BCELoss() criterion = nn.BCELoss()
optimizer = optim.Adam(model.parameters(), lr=0.001) optimizer = optim.Adam(model.parameters(), lr=0.01)
for epoch in range(1024): for epoch in range(1024):
state_h, state_c = model.init_state(seq_len) state_h, state_c = model.init_state(seq_len)