verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
kris1 has quit [Quit: kris1]
mikeling has quit [Quit: Connection closed for inactivity]
partobs-mdp has joined #mlpack
< partobs-mdp> zoq: Ran experiment on the AddTask representation you described. Didn't get any sensible performance boost. The weird thing (about both representation) is that the best validation score is steadily achieved only after *two* iterations (we're talking about maxLen = 2, as always), and then SGD just grinds to a standstill
< partobs-mdp> (although there were some slight boost around 20-30 iterations)
< partobs-mdp> Oh, and I still have the issue with running the experiment in the loop
< partobs-mdp> Error message in the beginning of the second iteration: error: Mat::rows(): indices out of bounds or incorrectly used
andrzejku has joined #mlpack
vivekp has quit [Ping timeout: 260 seconds]
vivekp has joined #mlpack
vivekp has quit [Ping timeout: 246 seconds]
vivekp has joined #mlpack
vivekp has quit [Ping timeout: 268 seconds]
shikhar has joined #mlpack
vivekp has joined #mlpack
kris1 has joined #mlpack
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
andrzejku has quit [Quit: My iMac has gone to sleep. ZZZzzz…]
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
< zoq> partobs-mdp: Hello, I'll go and rerun the experiments and post my results here. Also how can I produce the "error: Mat::rows()" issue, just run the script?
< partobs-mdp> Yes, run the script (or, rather, the bin/mlpack_lstm_baseline -t add -e 500 -b 2 -r 1 -s 128 -v --trials 5 - the script suppresses all output except the last line of cout)
< partobs-mdp> The second iteration should fail
< zoq> okay, thanks
< zoq> kris1: With model level forward you mean, a visitor which calls the Forward function not just for the layer but also for the layer inside the vector/Model()?
< kris1> Yes.
< kris1> right now the forward function for the model is private
< zoq> But what should the input be? The way it's implemented right now is that the layer that holds the vector/Model() is responsible for calling the Forward function of each layer. Maybe I misunderstood your question?
< zoq> Maybe you like to stick two FFN classes together?
< kris1> Yes, so right now the Forward function in the ffn class takes arma::mat as input. I want to directly expose this function. Right now the way it is implemented is that you call the evaluate function and then it in turn calls the forward function with predictors.col(i) as an argument
< kris1> I could actually change the Predictors matrix before calling evaluate and store the result. But that seems a but weird to me to change the predictors every time. Also i don’t need the output of outputLayer but just the output of the network.Back()
< zoq> If it helps, make them public, but maybe you could use sequential layer which is basically like the FFN class?
< kris1> okay thanks. I will take a look at the sequential class also.
< zoq> If it's easier to make the functions public, I don't mind to do that.
< zoq> ironstark: I'll take a look at the issue later today.
< zoq> partobs-mdp: "Can't see the logs - did you respond yet?" with logs you mean: http://www.mlpack.org/irc/?
partobs-mdp has quit [Remote host closed the connection]
govg has joined #mlpack
vivekp has quit [Read error: Connection reset by peer]
vivekp has joined #mlpack
andrzejku has joined #mlpack
shikhar has quit [Quit: WeeChat 1.7]
kris1 has quit [Quit: kris1]
< zoq> partobs-mdp: Since you changed the representation for the Add task, the network input layer should take a single element and output a single output right? So you should use RunTask<AddTask>(task, 1, 1, epochs, samples, trials); instead of RunTask<AddTask>(task, 3, 3, epochs, samples, trials);
< zoq> Also, I tested the minibatch SGD optimizer on the copy task and couldn't see any gradient explosion:
< zoq> bin/mlpack_lstm_baseline -t copy -e 800 -b 2 -l 10 -r 3 -s 1000 -v
< zoq> [INFO ] RNN::RNN(): final objective of trained model is 0.0994736.
< zoq> [INFO ] Running evaluation loop.
< zoq> [INFO ] Final score: 1
< zoq> so maybe we should switch to minibatch SGD for the other tasks as well
< zoq> I can reproduce your results
< zoq> Currently testing minibatch SGD on: bin/mlpack_lstm_baseline -t copy -e 800 -b 2 -l 10 -r 4 -s 1000 -v
< zoq> so far the object decreases in every iterations
< zoq> Also, currently we reset the update policy every time we call train; we should do: updatePolicy.Initialize(iterate.n_rows, iterate.n_cols); only once. We could introduce some parameter 'reset' that prevents this.
shikhar has joined #mlpack
kris1 has joined #mlpack
vivekp has quit [Ping timeout: 248 seconds]
shikhar has quit [Quit: WeeChat 1.7]
sumedhghaisas has joined #mlpack
sumedhghaisas has quit [Remote host closed the connection]
kris1 has quit [Quit: kris1]
andrzejku has quit [Quit: Textual IRC Client: www.textualapp.com]
sumedhghaisas has joined #mlpack
sumedhghaisas has quit [Ping timeout: 260 seconds]
sumedhghaisas has joined #mlpack